Datasets:
Update dataset card with metadata and project links (#2)
Browse files- Update dataset card with metadata and project links (af2c79422f423c10b7f29d6ee1f52c7abb3b40bc)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
|
@@ -1,4 +1,7 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
dataset_info:
|
| 3 |
features:
|
| 4 |
- name: uid
|
|
@@ -123,8 +126,33 @@ configs:
|
|
| 123 |
path: data/val-*
|
| 124 |
- split: test
|
| 125 |
path: data/test-*
|
| 126 |
-
license: cc-by-sa-4.0
|
| 127 |
---
|
| 128 |
-
# Dataset Card
|
| 129 |
|
| 130 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
license: cc-by-sa-4.0
|
| 3 |
+
task_categories:
|
| 4 |
+
- image-to-3d
|
| 5 |
dataset_info:
|
| 6 |
features:
|
| 7 |
- name: uid
|
|
|
|
| 126 |
path: data/val-*
|
| 127 |
- split: test
|
| 128 |
path: data/test-*
|
|
|
|
| 129 |
---
|
|
|
|
| 130 |
|
| 131 |
+
# PixARMesh Training Dataset
|
| 132 |
+
|
| 133 |
+
[**Project Page**](https://mlpc-ucsd.github.io/PixARMesh/) | [**Paper**](https://arxiv.org/abs/2603.05888) | [**GitHub**](https://github.com/mlpc-ucsd/PixARMesh)
|
| 134 |
+
|
| 135 |
+
This repository contains the training dataset for **PixARMesh**, a method to autoregressively reconstruct complete 3D indoor scene meshes directly from a single RGB image. Unlike prior methods that rely on implicit signed distance fields, PixARMesh jointly predicts object layout and geometry within a unified model, producing coherent and artist-ready meshes in a single forward pass.
|
| 136 |
+
|
| 137 |
+
## Dataset Preparation
|
| 138 |
+
|
| 139 |
+
According to the official repository, you should flatten the dataset to ensure uniform instance sampling across scenes:
|
| 140 |
+
|
| 141 |
+
```bash
|
| 142 |
+
python -m scripts.flatten_dataset
|
| 143 |
+
```
|
| 144 |
+
|
| 145 |
+
This prevents instances from scenes with many objects from being under-sampled during training.
|
| 146 |
+
|
| 147 |
+
## Citation
|
| 148 |
+
|
| 149 |
+
If you find PixARMesh useful in your research, please consider citing:
|
| 150 |
+
|
| 151 |
+
```bibtex
|
| 152 |
+
@article{zhang2026pixarmesh,
|
| 153 |
+
title={PixARMesh: Autoregressive Mesh-Native Single-View Scene Reconstruction},
|
| 154 |
+
author={Zhang, Xiang and Yoo, Sohyun and Wu, Hongrui and Li, Chuan and Xie, Jianwen and Tu, Zhuowen},
|
| 155 |
+
journal={arXiv preprint arXiv:2603.05888},
|
| 156 |
+
year={2026}
|
| 157 |
+
}
|
| 158 |
+
```
|