Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/wz0919/DreamRunner
Official implementation of DreamRunner: Fine-Grained Storytelling Video Generation with Retrieval-Augmented Motion Adaptation
https://github.com/wz0919/DreamRunner
Last synced: about 1 month ago
JSON representation
Official implementation of DreamRunner: Fine-Grained Storytelling Video Generation with Retrieval-Augmented Motion Adaptation
- Host: GitHub
- URL: https://github.com/wz0919/DreamRunner
- Owner: wz0919
- Created: 2024-11-25T23:38:12.000Z (about 2 months ago)
- Default Branch: main
- Last Pushed: 2024-11-26T01:16:32.000Z (about 2 months ago)
- Last Synced: 2024-11-26T01:30:26.277Z (about 2 months ago)
- Language: Python
- Size: 1.71 MB
- Stars: 2
- Watchers: 1
- Forks: 0
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
Awesome Lists containing this project
- awesome-diffusion-categorized - [Code
README
# DreamRunner: Fine-Grained Storytelling Video Generation with Retrieval-Augmented Motion Adaptation
[![Project Website](https://img.shields.io/badge/Project-Website-blue)](https://dreamrunner-story2video.github.io) [![arXiv](https://img.shields.io/badge/arXiv-2411.1665-b31b1b.svg)](https://arxiv.org/pdf/2411.16657)
#### [Zun Wang](https://zunwang1.github.io/), [Jialu Li](https://jialuli-luka.github.io/), [Han Lin](https://hl-hanlin.github.io/), [Jaehong Yoon](https://jaehong31.github.io), [Mohit Bansal](https://www.cs.unc.edu/~mbansal/)
#### Code coming soon! Expected before December 4th, 2024.
## ToDos
- [x] Release the inference code on T2V-ComBench.
- [ ] Release the code for retrieving videos and training character and motion loras.
- [ ] Release the inference code for storytelling video genetation.## Setup
### Environment Setup
```shell
conda create -n dreamrunner python==3.10
conda activate dreamrunner
pip install -r requirements.txt
```### Download Models
DreamRunner is implemented using CogVideoX-2B. You can download it [here](https://huggingface.co/THUDM/CogVideoX-2b) and put it to `pretrained_models/CogVideoX-2b`.## Running the Code
### T2V-Combench
#### Inference
We provide the plans we used for T2V-ComBench in `MotionDirector_SR3AI/t2v-combench/plan`.
You can specify the GPUs you want use in `MotionDirector_SR3AI/t2v-combench-2b.sh` for parallel inference.
Then directly Infer 600 videos on 6 dimensions of T2V-ComBnech with the following script
```
cd MotionDirector_SR3AI
bash run_bench_2b.sh
```
The generated videos will be saved at `MotionDirector_SR3AI/T2V-CompBench`.#### Evaluation
Please follow [T2V-ComBench](https://github.com/KaiyueSun98/T2V-CompBench) for evaluating the generated videos.### Storytell Video Generation
#### Coming soon!# Citation
If you find our project useful in your research, please cite the following paper:
```bibtex
@article{zun2024dreamrunner,
author = {Zun Wang and Jialu Li and Han Lin and Jaehong Yoon and Mohit Bansal},
title = {DreamRunner: Fine-Grained Storytelling Video Generation with Retrieval-Augmented Motion Adaptation},
journal = {arxiv},
year = {2024},
url = {https://arxiv.org/abs/2411.16657}
}
```