Datasets:
ArXiv:
License:
| #!/usr/bin/env python3 | |
| """ | |
| Download and unpack the objaverse_sims dataset from HuggingFace. | |
| This script downloads the dataset and extracts all tar archives to restore | |
| the original directory structure expected by downstream consumers. | |
| Usage: | |
| python unpack.py [target_directory] | |
| Example: | |
| python unpack.py ./objaverse_sims | |
| python unpack.py /data/datasets/objaverse_sims | |
| """ | |
| import os | |
| import sys | |
| import tarfile | |
| from pathlib import Path | |
| def unpack_dataset(target_dir: str = "./objaverse_sims"): | |
| """Download and unpack the dataset to the target directory.""" | |
| # Import here to give helpful error if not installed | |
| try: | |
| from huggingface_hub import snapshot_download | |
| except ImportError: | |
| print("Error: huggingface_hub not installed.") | |
| print("Install with: pip install huggingface_hub[hf_transfer]") | |
| sys.exit(1) | |
| target = Path(target_dir).resolve() | |
| print(f"Target directory: {target}") | |
| # Enable hf_transfer for faster downloads (if available) | |
| os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1" | |
| # Download dataset | |
| print("\n[1/3] Downloading dataset from HuggingFace...") | |
| print(" (This may take a while for ~30GB)") | |
| snapshot_download( | |
| repo_id="ellisbrown/objaverse_sims", | |
| repo_type="dataset", | |
| local_dir=str(target), | |
| local_dir_use_symlinks=False | |
| ) | |
| print(" Done!") | |
| # Unpack processed shards | |
| processed_dir = target / "processed" | |
| if processed_dir.exists(): | |
| print("\n[2/3] Unpacking processed objects...") | |
| shards = sorted(processed_dir.glob("shard_*.tar")) | |
| total_shards = len(shards) | |
| for i, shard in enumerate(shards, 1): | |
| print(f" [{i}/{total_shards}] Extracting {shard.name}...") | |
| with tarfile.open(shard) as tar: | |
| tar.extractall(processed_dir) | |
| shard.unlink() # Remove tar after extraction | |
| # Remove manifest (no longer needed) | |
| manifest = processed_dir / "manifest.json" | |
| if manifest.exists(): | |
| manifest.unlink() | |
| print(" Done!") | |
| else: | |
| print("\n[2/3] Skipping processed objects (directory not found)") | |
| # Unpack houses individual files | |
| houses_dir = target / "houses" | |
| if houses_dir.exists(): | |
| print("\n[3/3] Unpacking house files...") | |
| for split in ["train", "test", "val"]: | |
| tar_file = houses_dir / f"{split}_individual.tar" | |
| if tar_file.exists(): | |
| print(f" Extracting {tar_file.name}...") | |
| split_dir = houses_dir / split | |
| split_dir.mkdir(exist_ok=True) | |
| with tarfile.open(tar_file) as tar: | |
| tar.extractall(split_dir) | |
| tar_file.unlink() # Remove tar after extraction | |
| print(" Done!") | |
| else: | |
| print("\n[3/3] Skipping house files (directory not found)") | |
| # Summary | |
| print("\n" + "=" * 60) | |
| print("Dataset unpacked successfully!") | |
| print("=" * 60) | |
| print(f"\nLocation: {target}") | |
| print("\nStructure:") | |
| print(" processed/ - ~40K 3D object directories") | |
| print(" houses/ - train/test/val house layouts") | |
| print(" procthor_databases/ - asset databases") | |
| print(" 0.json - sample house") | |
| def main(): | |
| if len(sys.argv) > 1: | |
| if sys.argv[1] in ["-h", "--help"]: | |
| print(__doc__) | |
| sys.exit(0) | |
| target = sys.argv[1] | |
| else: | |
| target = "./objaverse_sims" | |
| unpack_dataset(target) | |
| if __name__ == "__main__": | |
| main() | |