https://github.com/SkyworkAI/SkyReels-A1
SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers
https://github.com/SkyworkAI/SkyReels-A1
condition-render video-diffusion
Last synced: 3 months ago
JSON representation
SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers
- Host: GitHub
- URL: https://github.com/SkyworkAI/SkyReels-A1
- Owner: SkyworkAI
- License: other
- Created: 2025-02-13T02:37:51.000Z (3 months ago)
- Default Branch: main
- Last Pushed: 2025-02-18T02:17:19.000Z (3 months ago)
- Last Synced: 2025-02-18T03:24:11.231Z (3 months ago)
- Topics: condition-render, video-diffusion
- Language: Python
- Homepage: https://www.skyreels.ai
- Size: 47.9 MB
- Stars: 26
- Watchers: 1
- Forks: 1
- Open Issues: 0
-
Metadata Files:
- Readme: README.md
- License: LICENSE.txt
Awesome Lists containing this project
- ai-game-devtools - SkyReels-A1 - A1: Expressive Portrait Animation in Video Diffusion Transformers. |[arXiv](https://arxiv.org/abs/2502.10841) | | Video | (<span id="video">Video</span> / <span id="tool">Tool (AI LLM)</span>)
README
![]()
SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers
Skywork AI
![]()
🔥 For more results, visit our homepage 🔥This repo, named **SkyReels-A1**, contains the official PyTorch implementation of our paper [SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers](https://arxiv.org).
## 🔥🔥🔥 News!!
* Feb 18, 2025: 👋 We release the inference code and model weights of SkyReels-A1. [Download](https://huggingface.co/Skywork/SkyReels-A1)
* Feb 18, 2025: 🎉 We have made our technical report available as open source. [Read](https://skyworkai.github.io/skyreels-a1.github.io/report.pdf)## Getting Started 🏁
### 1. Clone the code and prepare the environment 🛠️
First git clone the repository with code:
```bash
git clone https://github.com/SkyworkAI/SkyReels-A1.git
cd SkyReels-A1# create env using conda
conda create -n skyreels-a1 python=3.10
conda activate skyreels-a1
```
Then, install the remaining dependencies:
```bash
pip install -r requirements.txt
```### 2. Download pretrained weights 📥
You can download the pretrained weights is from HuggingFace:
```bash
# !pip install -U "huggingface_hub[cli]"
huggingface-cli download SkyReels-A1 --local-dir local_path --exclude "*.git*" "README.md" "docs"
```The FLAME, mediapipe, and smirk models are located in the SkyReels-A1/extra_models folder.
The directory structure of our SkyReels-A1 code is formulated as:
```text
pretrained_weights
├── FLAME
├── SkyReels-A1-5B
│ ├── pose_guider
│ ├── scheduler
│ ├── tokenizer
│ ├── siglip-so400m-patch14-384
│ ├── transformer
│ ├── vae
│ └── text_encoder
├── mediapipe
└── smirk```
### 3. Inference 🚀
You can simply run the inference scripts as:
```bash
python inference.py
```If the script runs successfully, you will get an output mp4 file. This file includes the following results: driving video, input image or video, and generated result.
## Gradio Interface 🤗
We provide a [Gradio](https://huggingface.co/docs/hub/spaces-sdks-gradio) interface for a better experience, just run by:
```bash
python app.py
```The graphical interactive interface is shown as below:

## Metric Evaluation 👓
We also provide all scripts for automatically calculating the metrics, including SimFace, FID, and L1 distance between expression and motion, reported in the paper.
All codes can be found in the ```eval``` folder. After setting the video result path, run the following commands in sequence:
```bash
python arc_score.py
python expression_score.py
python pose_score.py
```## Acknowledgements 💐
We would like to thank the contributors of [CogvideoX](https://github.com/THUDM/CogVideo) and [finetrainers](https://github.com/a-r-r-o-w/finetrainers) repositories, for their open research and contributions.## Citation 💖
If you find SkyReels-A1 useful for your research, welcome to 🌟 this repo and cite our work using the following BibTeX:
```bibtex
@misc{qiu2025skyreelsa1expressiveportraitanimation,
title={SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers},
author={Di Qiu and Zhengcong Fei and Rui Wang and Jialin Bai and Changqian Yu and Mingyuan Fan and Guibin Chen and Xiang Wen},
year={2025},
eprint={2502.10841},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2502.10841},
}
```