Upload README.md
Browse files
README.md
CHANGED
|
@@ -29,23 +29,14 @@ This repository contains our TI2V-5B model, built with the advanced Wan2.2-VAE t
|
|
| 29 |
|
| 30 |
## Video Demos
|
| 31 |
|
| 32 |
-
|
| 33 |
<div align="center">
|
| 34 |
-
|
| 35 |
-
<source src="https://cloud.video.taobao.com/vod/NnCd0fC-1eckDUuVBMz43oD_U6mTsPpBwga3wdnAkXA.mp4" type="video/mp4">
|
| 36 |
-
Your browser does not support the video tag.
|
| 37 |
-
</video>
|
| 38 |
</div>
|
| 39 |
|
| 40 |
-
|
| 41 |
## 🔥 Latest News!!
|
| 42 |
|
| 43 |
-
* Jul 28, 2025: 👋 Wan2.1 has been integrated into ComfyUI ([CN](https://docs.comfy.org/zh-CN/tutorials/video/wan/wan2_2) | [EN](https://docs.comfy.org/tutorials/video/wan/wan2_2)). Enjoy!
|
| 44 |
-
* Jul 28, 2025: 👋 Wan2.2's T2V, I2V and TI2V have been integrated into Diffusers ([T2V-A14B](https://huggingface.co/Wan-AI/Wan2.2-T2V-A14B-Diffusers) | [I2V-A14B](https://huggingface.co/Wan-AI/Wan2.2-I2V-A14B-Diffusers) | [TI2V-5B](https://huggingface.co/Wan-AI/Wan2.2-TI2V-5B-Diffusers)). Feel free to give it a try!
|
| 45 |
* Jul 28, 2025: 👋 We've released the inference code and model weights of **Wan2.2**.
|
| 46 |
|
| 47 |
-
|
| 48 |
-
|
| 49 |
## Community Works
|
| 50 |
If your research or project builds upon [**Wan2.1**](https://github.com/Wan-Video/Wan2.1) or Wan2.2, we welcome you to share it with us so we can highlight it for the broader community.
|
| 51 |
|
|
@@ -79,7 +70,6 @@ cd Wan2.2
|
|
| 79 |
Install dependencies:
|
| 80 |
```sh
|
| 81 |
# Ensure torch >= 2.4.0
|
| 82 |
-
# If the installation of `flash_attn` fails, try installing the other packages first and install `flash_attn` last
|
| 83 |
pip install -r requirements.txt
|
| 84 |
```
|
| 85 |
|
|
@@ -87,6 +77,8 @@ pip install -r requirements.txt
|
|
| 87 |
#### Model Download
|
| 88 |
|
| 89 |
|
|
|
|
|
|
|
| 90 |
| Models | Download Links | Description |
|
| 91 |
|--------------------|---------------------------------------------------------------------------------------------------------------------------------------------|-------------|
|
| 92 |
| T2V-A14B | 🤗 [Huggingface](https://huggingface.co/Wan-AI/Wan2.2-T2V-A14B) 🤖 [ModelScope](https://modelscope.cn/models/Wan-AI/Wan2.2-T2V-A14B) | Text-to-Video MoE model, supports 480P & 720P |
|
|
@@ -136,6 +128,7 @@ python generate.py --task ti2v-5B --size 1280*704 --ckpt_dir ./Wan2.2-TI2V-5B --
|
|
| 136 |
|
| 137 |
> 💡Similar to Image-to-Video, the `size` parameter represents the area of the generated video, with the aspect ratio following that of the original input image.
|
| 138 |
|
|
|
|
| 139 |
- Multi-GPU inference using FSDP + DeepSpeed Ulysses
|
| 140 |
|
| 141 |
```sh
|
|
|
|
| 29 |
|
| 30 |
## Video Demos
|
| 31 |
|
|
|
|
| 32 |
<div align="center">
|
| 33 |
+
<video src="https://github.com/user-attachments/assets/4aca6063-60bf-4953-bfb7-e265053f49ef" width="70%" poster=""> </video>
|
|
|
|
|
|
|
|
|
|
| 34 |
</div>
|
| 35 |
|
|
|
|
| 36 |
## 🔥 Latest News!!
|
| 37 |
|
|
|
|
|
|
|
| 38 |
* Jul 28, 2025: 👋 We've released the inference code and model weights of **Wan2.2**.
|
| 39 |
|
|
|
|
|
|
|
| 40 |
## Community Works
|
| 41 |
If your research or project builds upon [**Wan2.1**](https://github.com/Wan-Video/Wan2.1) or Wan2.2, we welcome you to share it with us so we can highlight it for the broader community.
|
| 42 |
|
|
|
|
| 70 |
Install dependencies:
|
| 71 |
```sh
|
| 72 |
# Ensure torch >= 2.4.0
|
|
|
|
| 73 |
pip install -r requirements.txt
|
| 74 |
```
|
| 75 |
|
|
|
|
| 77 |
#### Model Download
|
| 78 |
|
| 79 |
|
| 80 |
+
|
| 81 |
+
|
| 82 |
| Models | Download Links | Description |
|
| 83 |
|--------------------|---------------------------------------------------------------------------------------------------------------------------------------------|-------------|
|
| 84 |
| T2V-A14B | 🤗 [Huggingface](https://huggingface.co/Wan-AI/Wan2.2-T2V-A14B) 🤖 [ModelScope](https://modelscope.cn/models/Wan-AI/Wan2.2-T2V-A14B) | Text-to-Video MoE model, supports 480P & 720P |
|
|
|
|
| 128 |
|
| 129 |
> 💡Similar to Image-to-Video, the `size` parameter represents the area of the generated video, with the aspect ratio following that of the original input image.
|
| 130 |
|
| 131 |
+
|
| 132 |
- Multi-GPU inference using FSDP + DeepSpeed Ulysses
|
| 133 |
|
| 134 |
```sh
|