tttoaster commited on
Commit
bbc9a99
·
verified ·
1 Parent(s): 15e4fb8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -9
README.md CHANGED
@@ -1,7 +1,6 @@
1
  # ARC-Hunyuan-Video-7B
2
 
3
- <!-- [![arXiv](https://img.shields.io/badge/arXiv-2404.14396-b31b1b.svg)](https://arxiv.org/abs/2404.14396)-->
4
-
5
  [![Demo](https://img.shields.io/badge/ARC-Demo-blue)](https://arc.tencent.com/en/ai-demos/multimodal)
6
  [![Code](https://img.shields.io/badge/Github-Code-orange)](https://github.com/TencentARC/ARC-Hunyuan-Video-7B)
7
  [![Static Badge](https://img.shields.io/badge/Model-Huggingface-yellow)](https://huggingface.co/TencentARC/ARC-Hunyuan-Video-7B)
@@ -126,18 +125,17 @@ Due to video file size limitations imposed by the deployment API, we compressed
126
 
127
  We observe that incorporating generic video datasets during training may inadvertently compromise the model's capacity for real-world video understanding, potentially due to domain shift or noise introduced by non-real-world samples. To address this limitation, we plan to develop a dedicated model trained exclusively on rigorously curated real-world video data.
128
 
129
- <!-- ## Citation
130
 
131
  If you find the work helpful, please consider citing:
132
 
133
  ```bash
134
- @article{ge2024seed,
135
- title={SEED-X: Multimodal Models with Unified Multi-granularity Comprehension and Generation},
136
- author={Ge, Yuying and Zhao, Sijie and Zhu, Jinguo and Ge, Yixiao and Yi, Kun and Song, Lin and Li, Chen and Ding, Xiaohan and Shan, Ying},
137
- journal={arXiv preprint arXiv:2404.14396},
138
- year={2024}
139
  }
140
  ```
141
- -->
142
 
143
 
 
1
  # ARC-Hunyuan-Video-7B
2
 
3
+ [![arXiv](https://img.shields.io/badge/arXiv-2507.20939-b31b1b.svg)](https://arxiv.org/abs/2507.20939)
 
4
  [![Demo](https://img.shields.io/badge/ARC-Demo-blue)](https://arc.tencent.com/en/ai-demos/multimodal)
5
  [![Code](https://img.shields.io/badge/Github-Code-orange)](https://github.com/TencentARC/ARC-Hunyuan-Video-7B)
6
  [![Static Badge](https://img.shields.io/badge/Model-Huggingface-yellow)](https://huggingface.co/TencentARC/ARC-Hunyuan-Video-7B)
 
125
 
126
  We observe that incorporating generic video datasets during training may inadvertently compromise the model's capacity for real-world video understanding, potentially due to domain shift or noise introduced by non-real-world samples. To address this limitation, we plan to develop a dedicated model trained exclusively on rigorously curated real-world video data.
127
 
128
+ ## Citation
129
 
130
  If you find the work helpful, please consider citing:
131
 
132
  ```bash
133
+ @article{ge2025seed,
134
+ title={ARC-Hunyuan-Video-7B: Structured Video Comprehension of Real-World Shorts},
135
+ author={Ge, Yuying and Ge, Yixiao and Li, Chen and Wang, Teng and Pu, Junfu and Li, Yizhuo and Qiu, Lu and Ma, Jin and Duan, Lisheng and Zuo, Xinyu and Luo, Jinwen and Gu, Weibo and Li, Zexuan and Zhang, Xiaojing and Tao, Yangyu and Hu, Han and Wang, Di and Shan Ying},
136
+ journal={arXiv preprint arXiv:2507.20939},
137
+ year={2025}
138
  }
139
  ```
 
140
 
141