ellisbrown commited on
Commit
76c0bf4
·
verified ·
1 Parent(s): 8492db4

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +48 -232
README.md CHANGED
@@ -14,278 +14,94 @@ tags:
14
 
15
  # Objaverse VIDA Dataset
16
 
17
- This dataset contains processed 3D assets from [Objaverse](https://objaverse.allenai.org/) prepared for use with [AI2-THOR](https://ai2thor.allenai.org/) and [ProcTHOR](https://procthor.allenai.org/) environments. It includes ~40,000 processed 3D objects with textures and metadata, plus ~160,000 procedurally generated house layouts.
18
 
19
- ## What's Included
20
 
21
  | Component | Description | Size |
22
  |-----------|-------------|------|
23
- | `processed_2023_07_28/` | 39,664 processed 3D objects from Objaverse with textures and AI2-THOR metadata | 27 GB |
24
- | `houses_2023_07_28/` | 160,000+ procedurally generated house layouts (train/test/val splits) | 3.7 GB |
25
- | `procthor_databases_2023_07_28/` | Asset databases, material definitions, placement annotations | 70 MB |
26
- | `0.json` | Sample house layout | 90 KB |
27
 
28
- ## Data Format
29
-
30
- ### Processed Objects (`processed_2023_07_28/`)
31
-
32
- Each object is stored in its own directory named by UUID. Each directory contains:
33
-
34
- ```
35
- {object_id}/
36
- ├── {object_id}.pkl.gz # Processed 3D mesh data (gzip-compressed pickle)
37
- ├── albedo.jpg # Albedo/diffuse texture map
38
- ├── normal.jpg # Normal map
39
- ├── emission.jpg # Emission map
40
- └── thor_metadata.json # AI2-THOR compatible metadata (bounding box, properties, etc.)
41
- ```
42
-
43
- ### House Layouts (`houses_2023_07_28/`)
44
-
45
- - `train.jsonl.gz` / `test.jsonl.gz` / `val.jsonl.gz` - Compressed JSONL files with full house definitions
46
- - `train/` / `test/` / `val/` directories - Individual house JSON files (packed in `.tar` archives)
47
-
48
- Each house JSON contains rooms, doors, windows, objects, walls, and procedural parameters for AI2-THOR.
49
-
50
- ### Asset Databases (`procthor_databases_2023_07_28/`)
51
-
52
- - `asset-database.json` - Full asset catalog with metadata
53
- - `material-database.json` - Material definitions
54
- - `placement-annotations.json` - Object placement rules
55
- - `receptacles.json` - Receptacle definitions for object placement
56
- - `asset_groups/` - Predefined furniture groupings
57
-
58
- ---
59
-
60
- ## Unpacking Instructions
61
-
62
- The data is archived in `.tar` files for efficient storage. **To restore the original directory structure**, follow these steps:
63
-
64
- ### Quick Start (Full Dataset)
65
 
66
  ```bash
67
- # 1. Install huggingface_hub with hf_transfer for fast downloads
68
  pip install huggingface_hub[hf_transfer]
69
- export HF_HUB_ENABLE_HF_TRANSFER=1
70
-
71
- # 2. Download and unpack everything
72
- python -c "
73
- from huggingface_hub import snapshot_download
74
- snapshot_download(
75
- repo_id='spatial-training/objaverse_vida',
76
- repo_type='dataset',
77
- local_dir='./objaverse_vida'
78
- )
79
- "
80
-
81
- # 3. Unpack the tar archives to restore original structure
82
- cd objaverse_vida
83
-
84
- # Unpack processed objects (13 shards → ~40K directories)
85
- mkdir -p processed_2023_07_28_unpacked
86
- for shard in processed_2023_07_28/shard_*.tar; do
87
- echo "Extracting $shard..."
88
- tar -xf "$shard" -C processed_2023_07_28_unpacked/
89
- done
90
- # Replace archived version with unpacked
91
- rm -rf processed_2023_07_28/*.tar
92
- mv processed_2023_07_28_unpacked/* processed_2023_07_28/
93
- rmdir processed_2023_07_28_unpacked
94
-
95
- # Unpack house individual files
96
- cd houses_2023_07_28
97
- for split in train test val; do
98
- if [ -f "${split}_individual.tar" ]; then
99
- echo "Extracting ${split}_individual.tar..."
100
- mkdir -p "$split"
101
- tar -xf "${split}_individual.tar" -C "$split/"
102
- rm "${split}_individual.tar"
103
- fi
104
- done
105
- cd ..
106
-
107
- echo "Done! Dataset restored to original structure."
108
- ```
109
-
110
- ### Python Script (Recommended)
111
-
112
- Save this as `unpack_dataset.py`:
113
-
114
- ```python
115
- #!/usr/bin/env python3
116
- """Download and unpack objaverse_vida to original directory structure."""
117
 
118
- import os
119
- import tarfile
120
- from pathlib import Path
121
- from huggingface_hub import snapshot_download
122
 
123
- def unpack_dataset(target_dir: str = "./objaverse_vida"):
124
- target = Path(target_dir)
125
-
126
- # Enable hf_transfer for fast downloads
127
- os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
128
-
129
- print("Downloading dataset...")
130
- snapshot_download(
131
- repo_id="spatial-training/objaverse_vida",
132
- repo_type="dataset",
133
- local_dir=str(target)
134
- )
135
-
136
- # Unpack processed shards
137
- processed_dir = target / "processed_2023_07_28"
138
- if processed_dir.exists():
139
- print("\nUnpacking processed objects...")
140
- shards = sorted(processed_dir.glob("shard_*.tar"))
141
- for shard in shards:
142
- print(f" Extracting {shard.name}...")
143
- with tarfile.open(shard) as tar:
144
- tar.extractall(processed_dir)
145
- shard.unlink() # Remove tar after extraction
146
- # Remove manifest (no longer needed after extraction)
147
- manifest = processed_dir / "manifest.json"
148
- if manifest.exists():
149
- manifest.unlink()
150
-
151
- # Unpack houses individual files
152
- houses_dir = target / "houses_2023_07_28"
153
- if houses_dir.exists():
154
- print("\nUnpacking house files...")
155
- for split in ["train", "test", "val"]:
156
- tar_file = houses_dir / f"{split}_individual.tar"
157
- if tar_file.exists():
158
- print(f" Extracting {tar_file.name}...")
159
- split_dir = houses_dir / split
160
- split_dir.mkdir(exist_ok=True)
161
- with tarfile.open(tar_file) as tar:
162
- tar.extractall(split_dir)
163
- tar_file.unlink() # Remove tar after extraction
164
-
165
- print(f"\nDone! Dataset unpacked to: {target.absolute()}")
166
- print("\nStructure:")
167
- print(" processed_2023_07_28/ - 39,664 object directories")
168
- print(" houses_2023_07_28/ - train/test/val splits")
169
- print(" procthor_databases_2023_07_28/")
170
- print(" 0.json")
171
-
172
- if __name__ == "__main__":
173
- import sys
174
- target = sys.argv[1] if len(sys.argv) > 1 else "./objaverse_vida"
175
- unpack_dataset(target)
176
  ```
177
 
178
- Run with:
179
  ```bash
180
- python unpack_dataset.py /path/to/destination
181
  ```
182
 
183
- ### Download Specific Components Only
184
-
185
- ```python
186
- from huggingface_hub import hf_hub_download
187
- import tarfile
188
 
189
- # Download just one processed shard
190
- shard_path = hf_hub_download(
191
- repo_id="spatial-training/objaverse_vida",
192
- filename="processed_2023_07_28/shard_00.tar",
193
- repo_type="dataset"
194
- )
195
 
196
- # Extract to local directory
197
- with tarfile.open(shard_path) as tar:
198
- tar.extractall("./processed_2023_07_28")
 
 
 
 
 
199
  ```
200
 
201
- ### Find Which Shard Contains a Specific Object
202
 
203
- ```python
204
- from huggingface_hub import hf_hub_download
205
- import json
206
 
207
- # Download manifest
208
- manifest_path = hf_hub_download(
209
- repo_id="spatial-training/objaverse_vida",
210
- filename="processed_2023_07_28/manifest.json",
211
- repo_type="dataset"
212
- )
213
 
214
- with open(manifest_path) as f:
215
- manifest = json.load(f)
 
 
216
 
217
- # Find object by ID
218
- object_id = "000074a334c541878360457c672b6c2e"
219
- for shard_name, shard_info in manifest["shards"].items():
220
- if object_id in shard_info["dirs"]:
221
- print(f"Object {object_id} is in {shard_name}")
222
- break
223
- ```
224
 
225
- ---
226
 
227
- ## Final Directory Structure (After Unpacking)
 
 
228
 
229
- ```
230
- objaverse_vida/
231
- ├── 0.json
232
- ├── processed_2023_07_28/
233
- │ ├── 000074a334c541878360457c672b6c2e/
234
- │ │ ├── 000074a334c541878360457c672b6c2e.pkl.gz
235
- │ │ ├── albedo.jpg
236
- │ │ ├── emission.jpg
237
- │ │ ├── normal.jpg
238
- │ │ └── thor_metadata.json
239
- │ ├── 0002e50309b44e409c96f440202d90b3/
240
- │ │ └── ...
241
- │ └── ... (39,664 object directories)
242
- ├── houses_2023_07_28/
243
- │ ├── train.jsonl.gz
244
- │ ├── test.jsonl.gz
245
- │ ├── val.jsonl.gz
246
- │ ├── train/
247
- │ │ ├── 0.json.gz
248
- │ │ ├── 1.json.gz
249
- │ │ └── ... (159,745 files)
250
- │ ├── test/
251
- │ │ └── ... (15,994 files)
252
- │ └── val/
253
- │ └── ... (15,829 files)
254
- └── procthor_databases_2023_07_28/
255
- ├── asset-database.json
256
- ├── material-database.json
257
- ├── placement-annotations.json
258
- ├── receptacles.json
259
- ├── refined_annotations.json
260
- ├── skyboxes.json
261
- ├── solid-wall-colors.json
262
- ├── wall-holes.json
263
- └── asset_groups/
264
- └── ... (furniture groupings)
265
- ```
266
 
267
- ---
 
 
 
 
 
268
 
269
  ## Citation
270
 
271
- If you use this dataset, please cite the original works:
272
-
273
  ```bibtex
274
  @inproceedings{deitke2023objaverse,
275
  title={Objaverse: A Universe of Annotated 3D Objects},
276
- author={Deitke, Matt and Schwenk, Dustin and Salvador, Jordi and Weihs, Luca and Michel, Oscar and VanderBilt, Eli and Schmidt, Ludwig and Ehsani, Kiana and Kembhavi, Aniruddha and Farhadi, Ali},
277
  booktitle={CVPR},
278
  year={2023}
279
  }
280
 
281
  @inproceedings{deitke2022procthor,
282
  title={ProcTHOR: Large-Scale Embodied AI Using Procedural Generation},
283
- author={Deitke, Matt and VanderBilt, Eli and Herrasti, Alvaro and Weihs, Luca and Ehsani, Kiana and Salvador, Jordi and Han, Winson and Kolve, Eric and Kembhavi, Aniruddha and Mottaghi, Roozbeh},
284
  booktitle={NeurIPS},
285
  year={2022}
286
  }
287
  ```
288
-
289
- ## License
290
-
291
- Apache 2.0
 
14
 
15
  # Objaverse VIDA Dataset
16
 
17
+ Processed 3D assets from [Objaverse](https://objaverse.allenai.org/) for use with [AI2-THOR](https://ai2thor.allenai.org/) and [ProcTHOR](https://procthor.allenai.org/).
18
 
19
+ ## Contents
20
 
21
  | Component | Description | Size |
22
  |-----------|-------------|------|
23
+ | `processed_2023_07_28/` | ~40K processed 3D objects with textures | 27 GB |
24
+ | `houses_2023_07_28/` | 160K+ procedurally generated house layouts | 3.7 GB |
25
+ | `procthor_databases_2023_07_28/` | Asset databases, materials, placement rules | 70 MB |
 
26
 
27
+ ## Quick Start
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
 
29
  ```bash
30
+ # Install dependencies
31
  pip install huggingface_hub[hf_transfer]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
+ # Download the unpack script
34
+ wget https://huggingface.co/datasets/spatial-training/objaverse_vida/raw/main/unpack.py
 
 
35
 
36
+ # Run it (downloads ~30GB and extracts to ./objaverse_vida)
37
+ python unpack.py ./objaverse_vida
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  ```
39
 
40
+ Or specify a custom path:
41
  ```bash
42
+ python unpack.py /path/to/destination
43
  ```
44
 
45
+ ## Data Format
 
 
 
 
46
 
47
+ ### Processed Objects
 
 
 
 
 
48
 
49
+ Each object directory contains:
50
+ ```
51
+ {object_id}/
52
+ ├── {object_id}.pkl.gz # 3D mesh data (gzip pickle)
53
+ ├── albedo.jpg # Diffuse texture
54
+ ├── normal.jpg # Normal map
55
+ ├── emission.jpg # Emission map
56
+ └── thor_metadata.json # AI2-THOR metadata
57
  ```
58
 
59
+ ### House Layouts
60
 
61
+ - `train.jsonl.gz` / `test.jsonl.gz` / `val.jsonl.gz` - Full house definitions
62
+ - `train/` / `test/` / `val/` - Individual house JSON files
 
63
 
64
+ ### Asset Databases
 
 
 
 
 
65
 
66
+ - `asset-database.json` - Asset catalog
67
+ - `material-database.json` - Materials
68
+ - `placement-annotations.json` - Placement rules
69
+ - `receptacles.json` - Receptacle definitions
70
 
71
+ ## Manual Download
 
 
 
 
 
 
72
 
73
+ If you prefer not to use the script:
74
 
75
+ ```bash
76
+ # Clone the dataset
77
+ huggingface-cli download spatial-training/objaverse_vida --repo-type dataset --local-dir ./objaverse_vida
78
 
79
+ # Extract processed shards
80
+ cd objaverse_vida/processed_2023_07_28
81
+ for f in shard_*.tar; do tar -xf "$f" && rm "$f"; done
82
+ rm manifest.json
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
 
84
+ # Extract house files
85
+ cd ../houses_2023_07_28
86
+ for split in train test val; do
87
+ tar -xf "${split}_individual.tar" -C "$split/" && rm "${split}_individual.tar"
88
+ done
89
+ ```
90
 
91
  ## Citation
92
 
 
 
93
  ```bibtex
94
  @inproceedings{deitke2023objaverse,
95
  title={Objaverse: A Universe of Annotated 3D Objects},
96
+ author={Deitke, Matt and others},
97
  booktitle={CVPR},
98
  year={2023}
99
  }
100
 
101
  @inproceedings{deitke2022procthor,
102
  title={ProcTHOR: Large-Scale Embodied AI Using Procedural Generation},
103
+ author={Deitke, Matt and others},
104
  booktitle={NeurIPS},
105
  year={2022}
106
  }
107
  ```