|
--- |
|
license: cc-by-4.0 |
|
--- |
|
|
|
# GEN3C: 3D-Informed World-Consistent Video Generation with Precise Camera Control |
|
CVPR 2025 (Highlight) |
|
|
|
[Xuanchi Ren*](https://xuanchiren.com/), |
|
[Tianchang Shen*](https://www.cs.toronto.edu/~shenti11/) |
|
[Jiahui Huang](https://huangjh-pub.github.io/), |
|
[Huan Ling](https://www.cs.toronto.edu/~linghuan/), |
|
[Yifan Lu](https://yifanlu0227.github.io/), |
|
[Merlin Nimier-David](https://merlin.nimierdavid.fr/), |
|
[Thomas Müller](https://research.nvidia.com/person/thomas-muller), |
|
[Alexander Keller](https://research.nvidia.com/person/alex-keller), |
|
[Sanja Fidler](https://www.cs.toronto.edu/~fidler/), |
|
[Jun Gao](https://www.cs.toronto.edu/~jungao/) <br> |
|
\* indicates equal contribution <br> |
|
**[Paper](https://arxiv.org/pdf/2503.03751), [Project Page](https://research.nvidia.com/labs/toronto-ai/GEN3C/)** |
|
|
|
Abstract: We present GEN3C, a generative video model with precise Camera Control and |
|
temporal 3D Consistency. Prior video models already generate realistic videos, |
|
but they tend to leverage little 3D information, leading to inconsistencies, |
|
such as objects popping in and out of existence. Camera control, if implemented |
|
at all, is imprecise, because camera parameters are mere inputs to the neural |
|
network which must then infer how the video depends on the camera. In contrast, |
|
GEN3C is guided by a 3D cache: point clouds obtained by predicting the |
|
pixel-wise depth of seed images or previously generated frames. When generating |
|
the next frames, GEN3C is conditioned on the 2D renderings of the 3D cache with |
|
the new camera trajectory provided by the user. Crucially, this means that |
|
GEN3C neither has to remember what it previously generated nor does it have to |
|
infer the image structure from the camera pose. The model, instead, can focus |
|
all its generative power on previously unobserved regions, as well as advancing |
|
the scene state to the next frame. Our results demonstrate more precise camera |
|
control than prior work, as well as state-of-the-art results in sparse-view |
|
novel view synthesis, even in challenging settings such as driving scenes and |
|
monocular dynamic video. Results are best viewed in videos. |
|
|
|
## Citation |
|
``` |
|
@inproceedings{ren2025gen3c, |
|
title={GEN3C: 3D-Informed World-Consistent Video Generation with Precise Camera Control}, |
|
author={Ren, Xuanchi and Shen, Tianchang and Huang, Jiahui and Ling, Huan and |
|
Lu, Yifan and Nimier-David, Merlin and Müller, Thomas and Keller, Alexander and |
|
Fidler, Sanja and Gao, Jun}, |
|
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition}, |
|
year={2025} |
|
} |