|
---
|
|
license: other
|
|
license_name: license
|
|
license_link: LICENSE
|
|
library_name: DepthCrafter
|
|
arxiv: 2409.02095
|
|
tags:
|
|
- vision
|
|
pipeline_tag: depth-estimation
|
|
widget:
|
|
- inference: false
|
|
---
|
|
|
|
## ___***DepthCrafter: Generating Consistent Long Depth Sequences for Open-world Videos***___
|
|
<div align="center">
|
|
<img src='https://depthcrafter.github.io/img/logo.png' style="height:140px"></img>
|
|
|
|
|
|
|
|
<a href='https://arxiv.org/abs/2409.02095'><img src='https://img.shields.io/badge/arXiv-2409.02095-b31b1b.svg'></a>
|
|
<a href='https://depthcrafter.github.io'><img src='https://img.shields.io/badge/Project-Page-Green'></a>
|
|
|
|
|
|
_**[Wenbo Hu<sup>1* †</sup>](https://wbhu.github.io),
|
|
[Xiangjun Gao<sup>2*</sup>](https://scholar.google.com/citations?user=qgdesEcAAAAJ&hl=en),
|
|
[Xiaoyu Li<sup>1* †</sup>](https://xiaoyu258.github.io),
|
|
[Sijie Zhao<sup>1</sup>](https://scholar.google.com/citations?user=tZ3dS3MAAAAJ&hl=en),
|
|
[Xiaodong Cun<sup>1</sup>](https://vinthony.github.io/academic), <br>
|
|
[Yong Zhang<sup>1</sup>](https://yzhang2016.github.io),
|
|
[Long Quan<sup>2</sup>](https://home.cse.ust.hk/~quan),
|
|
[Ying Shan<sup>3, 1</sup>](https://scholar.google.com/citations?user=4oXBp9UAAAAJ&hl=en)**_
|
|
<br><br>
|
|
<sup>1</sup>Tencent AI Lab
|
|
<sup>2</sup>The Hong Kong University of Science and Technology
|
|
<sup>3</sup>ARC Lab, Tencent PCG
|
|
|
|
arXiv preprint, 2024
|
|
|
|
</div>
|
|
|
|
## 🔆 Introduction
|
|
🤗 DepthCrafter can generate temporally consistent long depth sequences with fine-grained details for open-world videos,
|
|
without requiring additional information such as camera poses or optical flow.
|
|
|
|
## 🎥 Visualization
|
|
We provide some demos of unprojected point cloud sequences, with reference RGB and estimated depth videos.
|
|
Please refer to our [project page](https://depthcrafter.github.io) for more details.
|
|
<img src="./assets/visualization.gif">
|
|
|
|
|