# Porting Large Datasets to LeRobot Dataset v3.0 This tutorial explains how to port large-scale robotic datasets to the LeRobot Dataset v3.0 format. We'll use the **DROID 1.0.1** dataset as our primary example, which demonstrates handling multi-terabyte datasets with thousands of shards across SLURM clusters. ## File Organization: v2.1 vs v3.0 Dataset v3.0 fundamentally changes how data is organized and stored: **v2.1 Structure (Episode-based)**: ``` dataset/ ├── data/chunk-000/episode_000000.parquet ├── data/chunk-000/episode_000001.parquet ├── videos/chunk-000/camera/episode_000000.mp4 └── meta/episodes.jsonl ``` **v3.0 Structure (File-based)**: ``` dataset/ ├── data/chunk-000/file-000.parquet # Multiple episodes per file ├── videos/camera/chunk-000/file-000.mp4 # Consolidated video chunks └── meta/episodes/chunk-000/file-000.parquet # Structured metadata ``` This transition from individual episode files to file-based chunks dramatically improves performance and reduces storage overhead. ## What's New in Dataset v3.0 Dataset v3.0 introduces significant improvements for handling large datasets: ### 🏗️ **Enhanced File Organization** - **File-based structure**: Episodes are now grouped into chunked files rather than individual episode files - **Configurable file sizes**: for data and video files - **Improved storage efficiency**: Better compression and reduced overhead ### 📊 **Modern Metadata Management** - **Parquet-based metadata**: Replaced JSON Lines with efficient parquet format - **Structured episode access**: Direct pandas DataFrame access via `dataset.meta.episodes` - **Per-episode statistics**: Enhanced statistics tracking at episode level ### 🚀 **Performance Enhancements** - **Memory-mapped access**: Improved RAM usage through PyArrow memory mapping - **Faster loading**: Significantly reduced dataset initialization time - **Better scalability**: Designed for datasets with millions of episodes ## Prerequisites Before porting large datasets, ensure you have: - **LeRobot installed** with v3.0 support. Follow our [Installation Guide](./installation). - **Sufficient storage**: Raw datasets can be very large (e.g., DROID requires 2TB) - **Cluster access** (recommended for large datasets): SLURM or similar job scheduler - **Dataset-specific dependencies**: For DROID, you'll need TensorFlow Dataset utilities ## Understanding the DROID Dataset [DROID 1.0.1](https://droid-dataset.github.io/droid/the-droid-dataset) is an excellent example of a large-scale robotic dataset: - **Size**: 1.7TB (RLDS format), 8.7TB (raw data) - **Structure**: 2048 pre-defined TensorFlow dataset shards - **Content**: 76,000+ robot manipulation trajectories from Franka Emika Panda robots - **Scope**: Real-world manipulation tasks across multiple environments and objects - **Format**: Originally in TensorFlow Records/RLDS format, requiring conversion to LeRobot format - **Hosting**: Google Cloud Storage with public access via `gsutil` The dataset contains diverse manipulation demonstrations with: - Multiple camera views (wrist camera, exterior cameras) - Natural language task descriptions - Robot proprioceptive state and actions - Success/failure annotations ### DROID Features Schema ```python DROID_FEATURES = { # Episode markers "is_first": {"dtype": "bool", "shape": (1,)}, "is_last": {"dtype": "bool", "shape": (1,)}, "is_terminal": {"dtype": "bool", "shape": (1,)}, # Language instructions "language_instruction": {"dtype": "string", "shape": (1,)}, "language_instruction_2": {"dtype": "string", "shape": (1,)}, "language_instruction_3": {"dtype": "string", "shape": (1,)}, # Robot state "observation.state.gripper_position": {"dtype": "float32", "shape": (1,)}, "observation.state.cartesian_position": {"dtype": "float32", "shape": (6,)}, "observation.state.joint_position": {"dtype": "float32", "shape": (7,)}, # Camera observations "observation.images.wrist_left": {"dtype": "image"}, "observation.images.exterior_1_left": {"dtype": "image"}, "observation.images.exterior_2_left": {"dtype": "image"}, # Actions "action.gripper_position": {"dtype": "float32", "shape": (1,)}, "action.cartesian_position": {"dtype": "float32", "shape": (6,)}, "action.joint_position": {"dtype": "float32", "shape": (7,)}, # Standard LeRobot format "observation.state": {"dtype": "float32", "shape": (8,)}, # joints + gripper "action": {"dtype": "float32", "shape": (8,)}, # joints + gripper } ``` ## Approach 1: Single Computer Porting ### Step 1: Install Dependencies For DROID specifically: ```bash pip install tensorflow pip install tensorflow_datasets ``` For other datasets, install the appropriate readers for your source format. ### Step 2: Download Raw Data Download DROID from Google Cloud Storage using `gsutil`: ```bash # Install Google Cloud SDK if not already installed # https://cloud.google.com/sdk/docs/install # Download the full RLDS dataset (1.7TB) gsutil -m cp -r gs://gresearch/robotics/droid/1.0.1 /your/data/ # Or download just the 100-episode sample (2GB) for testing gsutil -m cp -r gs://gresearch/robotics/droid_100 /your/data/ ``` > [!WARNING] > Large datasets require substantial time and storage: > > - **Full DROID (1.7TB)**: Several days to download depending on bandwidth > - **Processing time**: 7+ days for local porting of full dataset > - **Upload time**: 3+ days to push to Hugging Face Hub > - **Local storage**: ~400GB for processed LeRobot format ### Step 3: Port the Dataset ```bash python examples/port_datasets/port_droid.py \ --raw-dir /your/data/droid/1.0.1 \ --repo-id your_id/droid_1.0.1 \ --push-to-hub ``` ### Development and Testing For development, you can port a single shard: ```bash python examples/port_datasets/port_droid.py \ --raw-dir /your/data/droid/1.0.1 \ --repo-id your_id/droid_1.0.1_test \ --num-shards 2048 \ --shard-index 0 ``` This approach works for smaller datasets or testing, but large datasets require cluster computing. ## Approach 2: SLURM Cluster Porting (Recommended) For large datasets like DROID, parallel processing across multiple nodes dramatically reduces processing time. ### Step 1: Install Cluster Dependencies ```bash pip install datatrove # Hugging Face's distributed processing library ``` ### Step 2: Configure Your SLURM Environment Find your partition information: ```bash sinfo --format="%R" # List available partitions sinfo -N -p your_partition -h -o "%N cpus=%c mem=%m" # Check resources ``` Choose a **CPU partition** - no GPU needed for dataset porting. ### Step 3: Launch Parallel Porting Jobs ```bash python examples/port_datasets/slurm_port_shards.py \ --raw-dir /your/data/droid/1.0.1 \ --repo-id your_id/droid_1.0.1 \ --logs-dir /your/logs \ --job-name port_droid \ --partition your_partition \ --workers 2048 \ --cpus-per-task 8 \ --mem-per-cpu 1950M ``` #### Parameter Guidelines - **`--workers`**: Number of parallel jobs (max 2048 for DROID's shard count) - **`--cpus-per-task`**: 8 CPUs recommended for frame encoding parallelization - **`--mem-per-cpu`**: ~16GB total RAM (8×1950M) for loading raw frames > [!TIP] > Start with fewer workers (e.g., 100) to test your cluster configuration before launching thousands of jobs. ### Step 4: Monitor Progress Check running jobs: ```bash squeue -u $USER ``` Monitor overall progress: ```bash jobs_status /your/logs ``` Inspect individual job logs: ```bash less /your/logs/port_droid/slurm_jobs/JOB_ID_WORKER_ID.out ``` Debug failed jobs: ```bash failed_logs /your/logs/port_droid ``` ### Step 5: Aggregate Shards Once all porting jobs complete: ```bash python examples/port_datasets/slurm_aggregate_shards.py \ --repo-id your_id/droid_1.0.1 \ --logs-dir /your/logs \ --job-name aggr_droid \ --partition your_partition \ --workers 2048 \ --cpus-per-task 8 \ --mem-per-cpu 1950M ``` ### Step 6: Upload to Hub ```bash python examples/port_datasets/slurm_upload.py \ --repo-id your_id/droid_1.0.1 \ --logs-dir /your/logs \ --job-name upload_droid \ --partition your_partition \ --workers 50 \ --cpus-per-task 4 \ --mem-per-cpu 1950M ``` > [!NOTE] > Upload uses fewer workers (50) since it's network-bound rather than compute-bound. ## Dataset v3.0 File Structure Your completed dataset will have this modern structure: ``` dataset/ ├── meta/ │ ├── episodes/ │ │ └── chunk-000/ │ │ └── file-000.parquet # Episode metadata │ ├── tasks.parquet # Task definitions │ ├── stats.json # Aggregated statistics │ └── info.json # Dataset information ├── data/ │ └── chunk-000/ │ └── file-000.parquet # Consolidated episode data └── videos/ └── camera_key/ └── chunk-000/ └── file-000.mp4 # Consolidated video files ``` This replaces the old episode-per-file structure with efficient, optimally-sized chunks. ## Migrating from Dataset v2.1 If you have existing datasets in v2.1 format, use the migration tool: ```bash python src/lerobot/datasets/v30/convert_dataset_v21_to_v30.py \ --repo-id your_id/existing_dataset ``` This automatically: - Converts file structure to v3.0 format - Migrates metadata from JSON Lines to parquet - Aggregates statistics and creates per-episode stats - Updates version information ## Performance Benefits Dataset v3.0 provides significant improvements for large datasets: - **Faster loading**: 3-5x reduction in initialization time - **Memory efficiency**: Better RAM usage through memory mapping - **Scalable processing**: Handles millions of episodes efficiently - **Storage optimization**: Reduced file count and improved compression