The dataset viewer is not available for this subset.
Exception: SplitsNotFoundError
Message: The split names could not be parsed from the dataset config.
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 286, in get_dataset_config_info
for split_generator in builder._split_generators(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/webdataset/webdataset.py", line 82, in _split_generators
raise ValueError(
ValueError: The TAR archives of the dataset should be in WebDataset format, but the files in the archive don't share the same prefix or the same types.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/split_names.py", line 65, in compute_split_names_from_streaming_response
for split in get_dataset_split_names(
^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 340, in get_dataset_split_names
info = get_dataset_config_info(
^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/inspect.py", line 291, in get_dataset_config_info
raise SplitsNotFoundError("The split names could not be parsed from the dataset config.") from err
datasets.inspect.SplitsNotFoundError: The split names could not be parsed from the dataset config.Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
DUSt3R Preprocessed Training Data
Preprocessed training datasets for DUSt3R: Geometric 3D Vision Made Easy (CVPR 2024).
These datasets have been preprocessed using the scripts provided in the official DUSt3R repository (datasets_preprocess/) and are ready for training.
Datasets
| Dataset | Original Size | Parts | Description | Original License |
|---|---|---|---|---|
arkitscenes_processed |
153 GB | 8 (aa~ah) | ARKitScenes - Apple's indoor 3D scene dataset | CC BY-NC-SA 4.0 |
blendedmvs_processed |
79 GB | 4 (aa~ad) | BlendedMVS - Large-scale multi-view stereo dataset | CC BY 4.0 |
co3d_processed |
119 GB | 6 (aa~af) | CO3Dv2 - Common Objects in 3D | CC BY-NC 4.0 |
habitat_processed |
611 GB | 29 (aa~bc) | Habitat-Sim - Photorealistic 3D simulator | See Habitat |
megadepth_processed |
42 GB | 3 (aa~ac) | MegaDepth - Single-view depth prediction from internet photos | See MegaDepth |
scannetpp_processed |
36 GB | 2 (aa~ab) | ScanNet++ - High-fidelity indoor scene dataset | Non-commercial |
staticthings3d_processed |
43 GB | 3 (aa~ac) | StaticThings3D - Synthetic stereo dataset | See StaticThings3D |
waymo_processed |
24 GB | 2 (aa~ab) | Waymo Open Dataset - Autonomous driving dataset | Non-commercial |
wildrgbd_processed |
38 GB | 2 (aa~ab) | WildRGB-D - Wild RGB-D dataset | See WildRGB-D |
Total: ~1.1 TB (59 parts)
File Structure
Each dataset is split into 20 GB parts with the following naming convention:
<dataset_name>.tar.part_aa
<dataset_name>.tar.part_ab
<dataset_name>.tar.part_ac
...
<dataset_name>.sha256 # SHA256 checksums for integrity verification
Download
Download All (bash)
# Using git lfs
git lfs install
git clone https://huggingface.co/datasets/Yong-Hoon/dust3r-dataset
# Or using huggingface-cli
huggingface-cli download Yong-Hoon/dust3r-dataset --repo-type dataset --local-dir ./data
Download a Specific Dataset (bash)
huggingface-cli download Yong-Hoon/dust3r-dataset --repo-type dataset --include "arkitscenes_processed.*" --local-dir ./data
Download All (Python)
from huggingface_hub import snapshot_download
# Download entire dataset
snapshot_download(
repo_id="Yong-Hoon/dust3r-dataset",
repo_type="dataset",
local_dir="./data",
)
Download a Specific Dataset (Python)
from huggingface_hub import snapshot_download
# Download only a specific dataset (e.g., arkitscenes_processed)
snapshot_download(
repo_id="Yong-Hoon/dust3r-dataset",
repo_type="dataset",
local_dir="./data",
allow_patterns=["arkitscenes_processed.*"],
)
Download Multiple Datasets (Python)
from huggingface_hub import snapshot_download
# Choose datasets to download
datasets_to_download = [
"arkitscenes_processed",
"co3d_processed",
"megadepth_processed",
]
patterns = [f"{name}.*" for name in datasets_to_download]
snapshot_download(
repo_id="Yong-Hoon/dust3r-dataset",
repo_type="dataset",
local_dir="./data",
allow_patterns=patterns,
)
Checksum Verification
Each dataset includes a .sha256 file for integrity verification.
Verify (bash)
cd data
# Verify a specific dataset
sha256sum -c arkitscenes_processed.sha256
# Verify all datasets
for f in *.sha256; do
echo "Verifying $f ..."
sha256sum -c "$f"
done
Verify (Python)
import hashlib
from pathlib import Path
def verify_dataset(data_dir: str, dataset_name: str) -> bool:
"""Verify SHA256 checksums for a dataset."""
data_path = Path(data_dir)
sha256_file = data_path / f"{dataset_name}.sha256"
if not sha256_file.exists():
print(f"Checksum file not found: {sha256_file}")
return False
all_ok = True
for line in sha256_file.read_text().strip().splitlines():
expected_hash, filename = line.split()
filepath = data_path / filename.strip()
if not filepath.exists():
print(f"MISSING: {filename}")
all_ok = False
continue
sha256 = hashlib.sha256()
with open(filepath, "rb") as f:
for chunk in iter(lambda: f.read(8192 * 1024), b""):
sha256.update(chunk)
if sha256.hexdigest() == expected_hash:
print(f"OK: {filename}")
else:
print(f"FAILED: {filename}")
all_ok = False
return all_ok
# Verify a specific dataset
verify_dataset("./data", "arkitscenes_processed")
# Verify all datasets
datasets = [
"arkitscenes_processed", "blendedmvs_processed", "co3d_processed",
"habitat_processed", "megadepth_processed", "scannetpp_processed",
"staticthings3d_processed", "waymo_processed", "wildrgbd_processed",
]
for name in datasets:
print(f"\n=== {name} ===")
verify_dataset("./data", name)
Decompression
Decompress a Single Dataset
# Merge split parts and extract
cat arkitscenes_processed.tar.part_* | tar xf -
Decompress All Datasets
cd data
for name in arkitscenes blendedmvs co3d habitat megadepth scannetpp staticthings3d waymo wildrgbd; do
echo "Extracting ${name}_processed ..."
cat ${name}_processed.tar.part_* | tar xf -
done
Decompress a Specific Dataset to a Custom Directory
cat arkitscenes_processed.tar.part_* | tar xf - -C /path/to/output/
After extraction, the directory structure will be:
data/
arkitscenes_processed/
blendedmvs_processed/
co3d_processed/
habitat_processed/
megadepth_processed/
scannetpp_processed/
staticthings3d_processed/
waymo_processed/
wildrgbd_processed/
License
- DUSt3R code: CC BY-NC-SA 4.0 (Naver Corporation)
- Each dataset has its own license as listed in the table above. Please make sure to agree to the license of each dataset before use.
- This data is provided for non-commercial research purposes only.
Citation
@inproceedings{dust3r_cvpr24,
title={DUSt3R: Geometric 3D Vision Made Easy},
author={Shuzhe Wang and Vincent Leroy and Yohann Cabon and Boris Chidlovskii and Jerome Revaud},
booktitle = {CVPR},
year = {2024}
}
Acknowledgements
This dataset collection is based on the preprocessing scripts and pair lists provided by the DUSt3R. All original datasets are the property of their respective authors and institutions.
- Downloads last month
- 11