MOFA-Video_Traj / README.md
myniu
init
826d651
|
raw
history blame
1.26 kB
metadata
license: apache-2.0
sdk_version: 4.5.0

Updates πŸ”₯πŸ”₯πŸ”₯

We have released the Gradio demo for Hybrid (Trajectory + Landmark) Controls HERE!

Introduction

This repo provides the inference Gradio demo for Trajectory Control of MOFA-Video.

Environment Setup

pip install -r requirements.txt

Download checkpoints

  1. Download the pretrained checkpoints of SVD_xt from huggingface to ./ckpts.

  2. Download the checkpint of MOFA-Adapter from huggingface to ./ckpts.

The final structure of checkpoints should be:

./ckpts/
|-- controlnet
|   |-- config.json
|   `-- diffusion_pytorch_model.safetensors
|-- stable-video-diffusion-img2vid-xt-1-1
|   |-- feature_extractor
|       |-- ...
|   |-- image_encoder
|       |-- ...
|   |-- scheduler
|       |-- ...
|   |-- unet
|       |-- ...
|   |-- vae
|       |-- ...
|   |-- svd_xt_1_1.safetensors
|   `-- model_index.json

Run Gradio Demo

python run_gradio.py

Please refer to the instructions on the gradio interface during the inference process.

Paper

arxiv.org/abs/2405.20222