haoningwu commited on
Commit
7995e55
·
verified ·
1 Parent(s): be4749b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -22
README.md CHANGED
@@ -5,11 +5,11 @@ language:
5
  - en
6
  ---
7
 
8
- # SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass
9
 
10
- This repository contains the official PyTorch implementation of SceneGen: https://arxiv.org/abs/2508.15769/. Feel free to reach out for discussions!
11
 
12
- **Now the Inference Code and Pretrained Models are released!**
13
 
14
  <div align="center">
15
  <img src="./assets/SceneGen.png">
@@ -19,6 +19,9 @@ This repository contains the official PyTorch implementation of SceneGen: https:
19
  [Project Page](https://mengmouxu.github.io/SceneGen/) · [Paper](https://arxiv.org/abs/2508.15769/) · [Checkpoints](https://huggingface.co/haoningwu/SceneGen/)
20
 
21
  ## ⏩ News
 
 
 
22
  - [2025.8] The inference code and checkpoints are released.
23
  - [2025.8] Our pre-print paper has been released on arXiv.
24
 
@@ -82,7 +85,7 @@ This script launches a Gradio web interface for interactive scene generation.
82
  >
83
  > ### 🗃️ Step 2: Manage Cache
84
  > 1. Click **"Add to Cache"** when satisfied with the segmentation.
85
- > 2. Repeat Step 1-2 for multiple images.
86
  > 3. Use **"Delete Selected"** or **"Clear All"** to manage cached images.
87
  >
88
  > ### 🎮 Step 3: Generate Scene
@@ -92,6 +95,11 @@ This script launches a Gradio web interface for interactive scene generation.
92
  >
93
  > **💡 Pro Tip:** Try the examples below to get started quickly!
94
 
 
 
 
 
 
95
  ### Pre-segmented Image Inference
96
  This script processes a directory of pre-segmented images.
97
  - **Input**: The input folder structure should be similar to `assets/masked_image_test`, containing segmented scene images.
@@ -102,33 +110,64 @@ This script processes a directory of pre-segmented images.
102
  ```
103
 
104
  ## 📚 Dataset
105
- To be updated soon...
106
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  ## 🏋️‍♂️ Training
108
- To be updated soon...
109
-
110
- ## Evaluation
111
- To be updated soon...
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
112
 
113
  ## 📜 Citation
114
  If you use this code and data for your research or project, please cite:
115
-
116
- @article{meng2025scenegen,
117
- author = {Meng, Yanxu and Wu, Haoning and Zhang, Ya and Xie, Weidi},
118
- title = {SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass},
119
- journal = {arXiv preprint arXiv:2508.15769},
120
- year = {2025},
121
- }
122
-
123
  ## TODO
124
  - [x] Release Paper
125
  - [x] Release Checkpoints & Inference Code
126
- - [ ] Release Training Code
127
- - [ ] Release Evaluation Code
128
- - [ ] Release Data Processing Code
129
 
130
  ## Acknowledgements
131
  Many thanks to the code bases from [TRELLIS](https://github.com/microsoft/TRELLIS), [DINOv2](https://github.com/facebookresearch/dinov2), and [VGGT](https://github.com/facebookresearch/vggt).
132
 
133
  ## Contact
134
- If you have any questions, please feel free to contact [meng-mou-xu@sjtu.edu.cn](mailto:meng-mou-xu@sjtu.edu.cn) and [haoningwu3639@gmail.com](mailto:haoningwu3639@gmail.com).
 
5
  - en
6
  ---
7
 
8
+ # SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass (3DV 2026)
9
 
10
+ This repository contains the official PyTorch implementation of SceneGen: https://arxiv.org/abs/2508.15769/.
11
 
12
+ **Now the Training, Inference Code, and Pretrained Models have all been released! Feel free to reach out for discussions!**
13
 
14
  <div align="center">
15
  <img src="./assets/SceneGen.png">
 
19
  [Project Page](https://mengmouxu.github.io/SceneGen/) · [Paper](https://arxiv.org/abs/2508.15769/) · [Checkpoints](https://huggingface.co/haoningwu/SceneGen/)
20
 
21
  ## ⏩ News
22
+ - [2025.11] Evaluation code has been released.
23
+ - [2025.11] Glad to share that SceneGen has been accepted to 3DV 2026.
24
+ - [2025.9] Our training code and data processing code are released.
25
  - [2025.8] The inference code and checkpoints are released.
26
  - [2025.8] Our pre-print paper has been released on arXiv.
27
 
 
85
  >
86
  > ### 🗃️ Step 2: Manage Cache
87
  > 1. Click **"Add to Cache"** when satisfied with the segmentation.
88
+ > 2. Repeat Steps 1-2 for multiple images.
89
  > 3. Use **"Delete Selected"** or **"Clear All"** to manage cached images.
90
  >
91
  > ### 🎮 Step 3: Generate Scene
 
95
  >
96
  > **💡 Pro Tip:** Try the examples below to get started quickly!
97
 
98
+ https://github.com/user-attachments/assets/d0d53506-70cd-4bd3-a6ab-2f9b5b16f4d8
99
+
100
+
101
+ *Click the image above to watch the demo video*
102
+
103
  ### Pre-segmented Image Inference
104
  This script processes a directory of pre-segmented images.
105
  - **Input**: The input folder structure should be similar to `assets/masked_image_test`, containing segmented scene images.
 
110
  ```
111
 
112
  ## 📚 Dataset
113
+ To train and evaluate SceneGen, we use the [3D-FUTURE](https://tianchi.aliyun.com/dataset/98063) dataset. Please download and preprocess the dataset as follows:
114
+ 1. Download the 3D-FUTURE dataset from [here](https://tianchi.aliyun.com/dataset/98063) which requires applying for access.
115
+ 2. Follow the [TRELLIS](https://github.com/microsoft/TRELLIS) data processing instructions to preprocess the dataset. Make sure to follow their directory structure for compatibility and fully generate the necessary files and ``metadata.csv``.
116
+ 3. Run the ``dataset_toolkits/build_metadata_scene.py`` script to create the scene-level metadata file:
117
+ ```sh
118
+ python dataset_toolkits/build_metadata_scene.py 3D-FUTURE
119
+ --output_dir <path_to_3D-FUTURE>
120
+ --set <train or test>
121
+ --vggt_ckpt checkpoints/VGGT-1B --save_mask
122
+ ```
123
+ This will generate a `metadata_scene.csv` file or a `metadata_scene_test.csv` file in the specified dataset directory.
124
+ 4. For evaluation, run the ``dataset_toolkits/build_scene.sh`` script to render scene image for each scene(with Blender installed and the configs in the script set correctly):
125
+ ```sh
126
+ bash dataset_toolkits/build_scene.sh
127
+ ```
128
+ This will create a `scene_test_render` folder in the dataset directory containing the rendered images of the test scenes with Blender, which will be further used for evaluation.
129
  ## 🏋️‍♂️ Training
130
+ With the processed 3D-FUTURE dataset and the pretrained `ss_flow_img_dit_L_16l8_fp16.safetensors` model checkpoint from [TRELLIS](https://huggingface.co/microsoft/TRELLIS-image-large) correctly placed in the `checkpoints/scenegen/ckpts` directory, you can train SceneGen using the following command:
131
+ ```
132
+ bash scripts/train.sh
133
+ ```
134
+ For detailed training configurations, please refer to `configs/generation/ss_scenegen_flow_img_train.json` and change the parameters as needed.
135
+
136
+ ## 🧪 Evaluation
137
+ To generate the 3D scenes on the 3D-FUTURE test set using the SceneGen model, use the following command:
138
+ ```
139
+ bash scenegen_eval.sh
140
+ ```
141
+ which will use the `scenegen_eval.py` script to generate the normalized scenes.
142
+
143
+ To evaluate the trained SceneGen model on the 3D-FUTURE test set, use the following command:
144
+ ```
145
+ cd evalscene
146
+ bash eval_scenegen.sh
147
+ ```
148
+ Make sure to have the processed 3D-FUTURE dataset and the rendered images in place as described in the Dataset section and the evaluation configs in `evalscene/configs/test/scene_evaluation_scenegen.yaml` set correctly. Then the evaluation script will compute metrics between the normalized generated scenes and the ground truth.
149
+
150
+ Some packages used in the evaluation require additional installation. Please install the packages: `torchmetrics`, `lpips`, `clip`, and `probreg` via pip.
151
 
152
  ## 📜 Citation
153
  If you use this code and data for your research or project, please cite:
154
+ ```
155
+ @inproceedings{meng2026scenegen,
156
+ author = {Meng, Yanxu and Wu, Haoning and Zhang, Ya and Xie, Weidi},
157
+ title = {SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass},
158
+ booktitle = {International Conference on 3D Vision 2026},
159
+ year = {2026},
160
+ }
161
+ ```
162
  ## TODO
163
  - [x] Release Paper
164
  - [x] Release Checkpoints & Inference Code
165
+ - [x] Release Training Code
166
+ - [x] Release Data Processing Code
167
+ - [x] Release Evaluation Code
168
 
169
  ## Acknowledgements
170
  Many thanks to the code bases from [TRELLIS](https://github.com/microsoft/TRELLIS), [DINOv2](https://github.com/facebookresearch/dinov2), and [VGGT](https://github.com/facebookresearch/vggt).
171
 
172
  ## Contact
173
+ If you have any questions, please feel free to contact [meng-mou-xu@sjtu.edu.cn](mailto:meng-mou-xu@sjtu.edu.cn) and [haoningwu3639@gmail.com](mailto:haoningwu3639@gmail.com).