Text-to-Video
Diffusers
Safetensors
WanDMDPipeline
BrianChen1129's picture
Create README.md
6901306 verified
|
raw
history blame
1.86 kB
metadata
license: apache-2.0

license: apache-2.0

FastVideo FastWan2.1-T2V-1.3B-Diffusers Model

Model Overview

  • This model is jointly finetuned with DMD and VSA, based on Wan-AI/Wan2.1-T2V-1.3B-Diffusers.
  • It supports 3-step inference and achieves up to 20 FPS on a single H100 GPU.
  • Both finetuning and inference scripts are available in the FastVideo repository.
  • Try it out on FastVideo — we support a wide range of GPUs from H100 to 4090, and even support Mac users!

If you use FastWan2.1-T2V-1.3B-Diffusers model for your research, please cite our paper:

@article{zhang2025vsa,
  title={VSA: Faster Video Diffusion with Trainable Sparse Attention},
  author={Zhang, Peiyuan and Huang, Haofeng and Chen, Yongqi and Lin, Will and Liu, Zhengzhong and Stoica, Ion and Xing, Eric and Zhang, Hao},
  journal={arXiv preprint arXiv:2505.13389},
  year={2025}
}
@article{zhang2025fast,
  title={Fast video generation with sliding tile attention},
  author={Zhang, Peiyuan and Chen, Yongqi and Su, Runlong and Ding, Hangliang and Stoica, Ion and Liu, Zhengzhong and Zhang, Hao},
  journal={arXiv preprint arXiv:2502.04507},
  year={2025}
}