Input Image | Naive SVD | Forward LoRA |
---|---|---|
Backward Camera movement LoRA, trained with 512 X 512 resolution
Input Image | Naive SVD | Backward LoRA |
---|---|---|
- Simple codebase for finetuning StableVideoDiffusion
- Motion LoRA training codebase for StableVideoDiffusion
- You can train a LoRA for motion control!
- Compatible with
diffusers
[2024.05.28] The training code for Motion LoRA based on Stable Video Diffusion is uploaded!
You can also perfom Stable Video Diffusion fine tuning when you turn off the argument --train_lora
git clone https://github.com/tykim0507/Motion-LoRA.git
cd MotionLoRA
☀️ Start with StableVideoDiffusion
conda create -n motionlora python=3.10
conda activate motionlora
pip install -r requirements.txt
Model | Resolution | Checkpoint |
---|---|---|
Stable-Video-Diffusion (Text2Video) | 1024x576 | Hugging Face |
Store them as following structure:
cd MotionLoRA
.
└── checkpoints
└── stable-video-diffusion-img2vid-xt-1-1
We recommend git cloning the huggingface repository using git lfs.
mkdir checkpoints
cd checkpoints
git lfs install
git clone https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt-1-1
We have used the Mixkit dataset.
You can simply prepare any type of videos, but with similar motion encoded.
Actually using 1 video is enough for training the motion LoRA!
sh train.sh
@article{motionloratykim,
title = {MotionLoRA: Learn motion using Low-Rank Adaptation},
author = {Taeyoon Kim},
year = {2024},
}
Our codebase builds on Stable Video Diffusion. Thanks to the authors for sharing their codebases!
Additionally, GPU and NFS resources for training are supported by fal.ai🔥.
Feel free to refer to the fal Research Grants!