Datasets:
Update README.md
Browse files
README.md
CHANGED
|
@@ -59,10 +59,9 @@ configs:
|
|
| 59 |
|
| 60 |
## Overview
|
| 61 |
|
| 62 |
-
BlockGen-3D is a large-scale dataset of voxelized 3D models with accompanying text descriptions, specifically designed for text-to-3D generation tasks. By processing and voxelizing models from the [Objaverse dataset](https://huggingface.co/datasets/allenai/objaverse), we have created a standardized representation that is particularly suitable for training 3D
|
| 63 |
-
|
| 64 |
-
Our dataset provides two types of representations: shape-only models represented as binary occupancy grids, and colored models with full RGBA information. Each model is represented in a 32×32×32 voxel grid, striking a balance between detail preservation and computational efficiency. This uniform representation makes the dataset especially suitable for training diffusion models and other deep learning architectures for 3D generation.
|
| 65 |
|
|
|
|
| 66 |
The dataset contains 542,292 total samples, with:
|
| 67 |
- 515,177 training samples
|
| 68 |
- 27,115 test samples
|
|
@@ -160,7 +159,9 @@ dataloader = DataLoader(
|
|
| 160 |
|
| 161 |
## Visualization Examples
|
| 162 |
|
| 163 |
-
|
|
|
|
|
|
|
| 164 |
|
| 165 |
## Dataset Creation Process
|
| 166 |
|
|
@@ -184,21 +185,25 @@ Our dataset was created through the following steps:
|
|
| 184 |
|
| 185 |
If you use this dataset in your research, please cite:
|
| 186 |
|
|
|
|
| 187 |
@misc{blockgen2024,
|
| 188 |
title={BlockGen-3D: A Large-Scale Dataset for Text-to-3D Voxel Generation},
|
| 189 |
author={Peter A. Massih},
|
| 190 |
year={2024},
|
| 191 |
publisher={Hugging Face}
|
| 192 |
}
|
|
|
|
| 193 |
|
| 194 |
Please also cite the original Objaverse dataset:
|
| 195 |
|
|
|
|
| 196 |
@article{objaverse2023,
|
| 197 |
title={Objaverse: A Universe of Annotated 3D Objects},
|
| 198 |
author={Matt Deitke and Dustin Schwenk and Jordi Salvador and Luca Weihs and Oscar Michel and Eli VanderBilt and Ludwig Schmidt and Kiana Ehsani and Aniruddha Kembhavi and Ali Farhadi},
|
| 199 |
journal={arXiv preprint arXiv:2304.02643},
|
| 200 |
year={2023}
|
| 201 |
}
|
|
|
|
| 202 |
|
| 203 |
## Limitations
|
| 204 |
|
|
|
|
| 59 |
|
| 60 |
## Overview
|
| 61 |
|
| 62 |
+
BlockGen-3D is a large-scale dataset of voxelized 3D models with accompanying text descriptions, specifically designed for text-to-3D generation tasks. By processing and voxelizing models from the [Objaverse dataset](https://huggingface.co/datasets/allenai/objaverse), we have created a standardized representation that is particularly suitable for training 3D diffusion models.
|
|
|
|
|
|
|
| 63 |
|
| 64 |
+
Our dataset provides two types of representations: shape-only models represented as binary occupancy grids, and colored models with full RGBA information. Each model is represented in a 32×32×32 voxel grid.
|
| 65 |
The dataset contains 542,292 total samples, with:
|
| 66 |
- 515,177 training samples
|
| 67 |
- 27,115 test samples
|
|
|
|
| 159 |
|
| 160 |
## Visualization Examples
|
| 161 |
|
| 162 |
+

|
| 163 |
+

|
| 164 |
+
*Dataset Usage Example - Showing voxelization process*
|
| 165 |
|
| 166 |
## Dataset Creation Process
|
| 167 |
|
|
|
|
| 185 |
|
| 186 |
If you use this dataset in your research, please cite:
|
| 187 |
|
| 188 |
+
```
|
| 189 |
@misc{blockgen2024,
|
| 190 |
title={BlockGen-3D: A Large-Scale Dataset for Text-to-3D Voxel Generation},
|
| 191 |
author={Peter A. Massih},
|
| 192 |
year={2024},
|
| 193 |
publisher={Hugging Face}
|
| 194 |
}
|
| 195 |
+
```
|
| 196 |
|
| 197 |
Please also cite the original Objaverse dataset:
|
| 198 |
|
| 199 |
+
```
|
| 200 |
@article{objaverse2023,
|
| 201 |
title={Objaverse: A Universe of Annotated 3D Objects},
|
| 202 |
author={Matt Deitke and Dustin Schwenk and Jordi Salvador and Luca Weihs and Oscar Michel and Eli VanderBilt and Ludwig Schmidt and Kiana Ehsani and Aniruddha Kembhavi and Ali Farhadi},
|
| 203 |
journal={arXiv preprint arXiv:2304.02643},
|
| 204 |
year={2023}
|
| 205 |
}
|
| 206 |
+
```
|
| 207 |
|
| 208 |
## Limitations
|
| 209 |
|