Update README.md
Browse files
README.md
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
license: apache-2.0
|
| 3 |
---
|
| 4 |
|
| 5 |
-
# FastVideo FastWan2.2-TI2V-5B-Diffusers Model
|
| 6 |
<p align="center">
|
| 7 |
<img src="https://raw.githubusercontent.com/hao-ai-lab/FastVideo/main/assets/logo.png" width="200"/>
|
| 8 |
</p>
|
|
@@ -22,7 +22,7 @@ license: apache-2.0
|
|
| 22 |
## Introduction
|
| 23 |
We're excited to introduce the **FastWan2.2 series**—a new line of models finetuned with our novel **Sparse-distill** strategy. This approach jointly integrates DMD and VSA in a single training process, combining the benefits of both **distillation** to shorten diffusion steps and **sparse attention** to reduce attention computations, enabling even faster video generation.
|
| 24 |
|
| 25 |
-
FastWan2.2-TI2V-5B-Diffusers is built upon Wan-AI/Wan2.2-TI2V-5B-Diffusers. It supports efficient **3-step inference** and produces high-quality videos at 121×704×1280 resolution. For training, we used simulated forward for the generator model, making the process data-free. **The current FastWan2.2-TI2V-5B-Diffusers model is trained using only DMD**.
|
| 26 |
|
| 27 |
---
|
| 28 |
|
|
@@ -37,7 +37,7 @@ FastWan2.2-TI2V-5B-Diffusers is built upon Wan-AI/Wan2.2-TI2V-5B-Diffusers. It s
|
|
| 37 |
```python
|
| 38 |
num_gpus=1
|
| 39 |
export FASTVIDEO_ATTENTION_BACKEND=FLASH_ATTN
|
| 40 |
-
export MODEL_BASE=FastVideo/FastWan2.2-TI2V-5B-Diffusers
|
| 41 |
# export MODEL_BASE=hunyuanvideo-community/HunyuanVideo
|
| 42 |
# You can either use --prompt or --prompt-txt, but not both.
|
| 43 |
fastvideo generate \
|
|
@@ -62,7 +62,7 @@ fastvideo generate \
|
|
| 62 |
|
| 63 |
Training was conducted on **8 nodes with 64 H200 GPUs** in total, using a `global batch size = 64`, and training runs for **3000 steps (~12 hours)**
|
| 64 |
|
| 65 |
-
If you use the FastWan2.2-TI2V-5B-Diffusers model for your research, please cite our paper:
|
| 66 |
```
|
| 67 |
@article{zhang2025vsa,
|
| 68 |
title={VSA: Faster Video Diffusion with Trainable Sparse Attention},
|
|
|
|
| 2 |
license: apache-2.0
|
| 3 |
---
|
| 4 |
|
| 5 |
+
# FastVideo FastWan2.2-TI2V-5B-Full-Diffusers Model
|
| 6 |
<p align="center">
|
| 7 |
<img src="https://raw.githubusercontent.com/hao-ai-lab/FastVideo/main/assets/logo.png" width="200"/>
|
| 8 |
</p>
|
|
|
|
| 22 |
## Introduction
|
| 23 |
We're excited to introduce the **FastWan2.2 series**—a new line of models finetuned with our novel **Sparse-distill** strategy. This approach jointly integrates DMD and VSA in a single training process, combining the benefits of both **distillation** to shorten diffusion steps and **sparse attention** to reduce attention computations, enabling even faster video generation.
|
| 24 |
|
| 25 |
+
FastWan2.2-TI2V-5B-Full-Diffusers is built upon Wan-AI/Wan2.2-TI2V-5B-Diffusers. It supports efficient **3-step inference** and produces high-quality videos at 121×704×1280 resolution. For training, we used simulated forward for the generator model, making the process data-free. **The current FastWan2.2-TI2V-5B-Full-Diffusers model is trained using only DMD**.
|
| 26 |
|
| 27 |
---
|
| 28 |
|
|
|
|
| 37 |
```python
|
| 38 |
num_gpus=1
|
| 39 |
export FASTVIDEO_ATTENTION_BACKEND=FLASH_ATTN
|
| 40 |
+
export MODEL_BASE=FastVideo/FastWan2.2-TI2V-5B-Full-Diffusers
|
| 41 |
# export MODEL_BASE=hunyuanvideo-community/HunyuanVideo
|
| 42 |
# You can either use --prompt or --prompt-txt, but not both.
|
| 43 |
fastvideo generate \
|
|
|
|
| 62 |
|
| 63 |
Training was conducted on **8 nodes with 64 H200 GPUs** in total, using a `global batch size = 64`, and training runs for **3000 steps (~12 hours)**
|
| 64 |
|
| 65 |
+
If you use the FastWan2.2-TI2V-5B-Full-Diffusers model for your research, please cite our paper:
|
| 66 |
```
|
| 67 |
@article{zhang2025vsa,
|
| 68 |
title={VSA: Faster Video Diffusion with Trainable Sparse Attention},
|