nielsr HF Staff commited on
Commit
2b40a4c
·
verified ·
1 Parent(s): e5be543

Add model card for Lyra 2.0

Browse files

This PR populates the model card for Lyra 2.0 with relevant information from the paper and official repository. It includes:
- Metadata for the `image-to-3d` pipeline and the Apache 2.0 license.
- Links to the paper ([2604.13036](https://huggingface.co/papers/2604.13036)), project page, and official GitHub repository.
- A brief summary of the framework's approach to long-horizon, 3D-consistent scene generation.
- Citation information.

Files changed (1) hide show
  1. README.md +33 -2
README.md CHANGED
@@ -1,5 +1,36 @@
1
  ---
2
- {}
 
3
  ---
4
 
5
- Model Release Today. Stay tuned!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: apache-2.0
3
+ pipeline_tag: image-to-3d
4
  ---
5
 
6
+ # Lyra 2.0: Explorable Generative 3D Worlds
7
+
8
+ [Paper](https://huggingface.co/papers/2604.13036) | [Project Page](https://research.nvidia.com/labs/sil/lyra2/) | [Code](https://github.com/nv-tlabs/lyra)
9
+
10
+ Lyra 2.0 is a framework for generating persistent, explorable 3D worlds at scale. It addresses the challenges of spatial forgetting and temporal drifting in long-horizon 3D scene creation. By maintaining per-frame 3D geometry for information routing and using self-augmented histories during training to correct synthesis errors, Lyra 2.0 enables substantially longer and 3D-consistent video trajectories that can be reconstructed into high-quality 3D scenes.
11
+
12
+ ## Overview
13
+
14
+ Recent advances in video generation enable a new paradigm for 3D scene creation: generating camera-controlled videos that simulate scene walkthroughs, then lifting them to 3D via feed-forward reconstruction techniques. Lyra 2.0 scales this approach to large, complex environments through two key mechanisms:
15
+ - **Geometry-aware Information Routing**: Maintaining per-frame 3D geometry to retrieve relevant past frames and establish correspondences.
16
+ - **Drift Correction**: Training with self-augmented histories to teach the model to correct accumulated synthesis errors rather than propagate them.
17
+
18
+ ## Citation
19
+
20
+ ```bibtex
21
+ @article{shen2026lyra2,
22
+ title={Lyra 2.0: Explorable Generative 3D Worlds},
23
+ author={Shen, Tianchang and Bahmani, Sherwin and He, Kai and Srinivasan, Sangeetha Grama and Cao, Tianshi and Ren, Jiawei and Li, Ruilong and Wang, Zian and Sharp, Nicholas and Gojcic, Zan and Fidler, Sanja and Huang, Jiahui and Ling, Huan and Gao, Jun and Ren, Xuanchi},
24
+ journal={arXiv preprint arXiv:2604.13036},
25
+ year={2026}
26
+ }
27
+
28
+ @inproceedings{bahmani2026lyra,
29
+ title={Lyra: Generative 3D Scene Reconstruction via Video Diffusion Model Self-Distillation},
30
+ author={Bahmani, Sherwin and Shen, Tianchang and Ren, Jiawei and Huang, Jiahui and Jiang, Yifeng and
31
+ Turki, Haithem and Tagliasacchi, Andrea and Lindell, David B. and Gojcic, Zan and Fidler, Sanja and
32
+ Ling, Huan and Gao, Jun and Ren, Xuanchi},
33
+ booktitle={International Conference on Learning Representations (ICLR)},
34
+ year={2026}
35
+ }
36
+ ```