kelseye commited on
Commit
5515990
·
verified ·
1 Parent(s): 22ffe7a

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -11
README.md CHANGED
@@ -29,23 +29,14 @@ This repository contains our TI2V-5B model, built with the advanced Wan2.2-VAE t
29
 
30
  ## Video Demos
31
 
32
-
33
  <div align="center">
34
- <video width="80%" controls>
35
- <source src="https://cloud.video.taobao.com/vod/NnCd0fC-1eckDUuVBMz43oD_U6mTsPpBwga3wdnAkXA.mp4" type="video/mp4">
36
- Your browser does not support the video tag.
37
- </video>
38
  </div>
39
 
40
-
41
  ## 🔥 Latest News!!
42
 
43
- * Jul 28, 2025: 👋 Wan2.1 has been integrated into ComfyUI ([CN](https://docs.comfy.org/zh-CN/tutorials/video/wan/wan2_2) | [EN](https://docs.comfy.org/tutorials/video/wan/wan2_2)). Enjoy!
44
- * Jul 28, 2025: 👋 Wan2.2's T2V, I2V and TI2V have been integrated into Diffusers ([T2V-A14B](https://huggingface.co/Wan-AI/Wan2.2-T2V-A14B-Diffusers) | [I2V-A14B](https://huggingface.co/Wan-AI/Wan2.2-I2V-A14B-Diffusers) | [TI2V-5B](https://huggingface.co/Wan-AI/Wan2.2-TI2V-5B-Diffusers)). Feel free to give it a try!
45
  * Jul 28, 2025: 👋 We've released the inference code and model weights of **Wan2.2**.
46
 
47
-
48
-
49
  ## Community Works
50
  If your research or project builds upon [**Wan2.1**](https://github.com/Wan-Video/Wan2.1) or Wan2.2, we welcome you to share it with us so we can highlight it for the broader community.
51
 
@@ -79,7 +70,6 @@ cd Wan2.2
79
  Install dependencies:
80
  ```sh
81
  # Ensure torch >= 2.4.0
82
- # If the installation of `flash_attn` fails, try installing the other packages first and install `flash_attn` last
83
  pip install -r requirements.txt
84
  ```
85
 
@@ -87,6 +77,8 @@ pip install -r requirements.txt
87
  #### Model Download
88
 
89
 
 
 
90
  | Models | Download Links | Description |
91
  |--------------------|---------------------------------------------------------------------------------------------------------------------------------------------|-------------|
92
  | T2V-A14B | 🤗 [Huggingface](https://huggingface.co/Wan-AI/Wan2.2-T2V-A14B) 🤖 [ModelScope](https://modelscope.cn/models/Wan-AI/Wan2.2-T2V-A14B) | Text-to-Video MoE model, supports 480P & 720P |
@@ -136,6 +128,7 @@ python generate.py --task ti2v-5B --size 1280*704 --ckpt_dir ./Wan2.2-TI2V-5B --
136
 
137
  > 💡Similar to Image-to-Video, the `size` parameter represents the area of the generated video, with the aspect ratio following that of the original input image.
138
 
 
139
  - Multi-GPU inference using FSDP + DeepSpeed Ulysses
140
 
141
  ```sh
 
29
 
30
  ## Video Demos
31
 
 
32
  <div align="center">
33
+ <video src="https://github.com/user-attachments/assets/4aca6063-60bf-4953-bfb7-e265053f49ef" width="70%" poster=""> </video>
 
 
 
34
  </div>
35
 
 
36
  ## 🔥 Latest News!!
37
 
 
 
38
  * Jul 28, 2025: 👋 We've released the inference code and model weights of **Wan2.2**.
39
 
 
 
40
  ## Community Works
41
  If your research or project builds upon [**Wan2.1**](https://github.com/Wan-Video/Wan2.1) or Wan2.2, we welcome you to share it with us so we can highlight it for the broader community.
42
 
 
70
  Install dependencies:
71
  ```sh
72
  # Ensure torch >= 2.4.0
 
73
  pip install -r requirements.txt
74
  ```
75
 
 
77
  #### Model Download
78
 
79
 
80
+
81
+
82
  | Models | Download Links | Description |
83
  |--------------------|---------------------------------------------------------------------------------------------------------------------------------------------|-------------|
84
  | T2V-A14B | 🤗 [Huggingface](https://huggingface.co/Wan-AI/Wan2.2-T2V-A14B) 🤖 [ModelScope](https://modelscope.cn/models/Wan-AI/Wan2.2-T2V-A14B) | Text-to-Video MoE model, supports 480P & 720P |
 
128
 
129
  > 💡Similar to Image-to-Video, the `size` parameter represents the area of the generated video, with the aspect ratio following that of the original input image.
130
 
131
+
132
  - Multi-GPU inference using FSDP + DeepSpeed Ulysses
133
 
134
  ```sh