--- license: apache-2.0 task_categories: - robotics tags: - LeRobot configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description This dataset contains the first set of teleoperated demonstrations collected during a two-day hackathon using the LeRobot library and SO-101 robot arms in a leader–follower setup. Each episode shows the follower arm picking one colored cube and placing it onto the matching colored cross inside a 2×2 grid. Two synchronized RGB cameras were used: - **Top camera**: overhead, provides a full 2D view of the workspace (arm, cube, grid). - **Front/low camera**: slightly above ground level, facing the arm and grid to capture z-axis cues and arm self-pose. The background was masked with cardboard panels, but ambient lighting varied throughout the day; this variation is preserved and is useful for robustness studies. Intended for vision-based imitation learning, multi-view fusion, and tabletop manipulation research. ### Use Cases - **Imitation Learning**: Behavior cloning from teleop demonstrations. - **Multiview Perception**: Fusing top + front perspectives for depth inference without explicit depth sensors. - **Robustness to Lighting**: Evaluating policy sensitivity to illumination drift. - **State–Action Alignment**: Leveraging synchronized proprioception and images. - **Policy Bootstrapping for curricula**: pretrain on single-cube before multi-cube tasks. ## Data Collection ### Teleoperation & Hardware - **Leader–Follower teleop**: human drives a leader arm; follower SO-101 replicates to produce demonstrations. - **Workspace**: Tabletop with 2×2 grid; only one cell has a colored cross. One cube is placed in its matching cross per episode. - **Cameras**: - **Front**: static overhead. - **Left**: static frontal view emphasizing depth. - **Environment**: Cardboard background; illumination changes across time are present in the data. ### Episode Protocol 1- Move to pre-grasp and visually localize the target cube. 2- Approach and grasp the cube. 3- Transport and align over the colored cross. 4- Place, release, and return to neutral. ## Known Limitations Lighting drift: Varying brightness/temperature across episodes; apply color constancy, normalization, or photometric augmentation. Occlusions: Hand/gripper and cube may occlude from the front camera during close approaches. No depth sensor: Only RGB; consider multi-view fusion or learned depth cues. Action semantics: Confirm whether actions are delta-pose or joint velocities in each metadata.json. Early-phase variability: Being the first batch, some episodes may include exploratory motions, hesitations, or failed initial grasps that later recover—useful for learning robustness but consider filtering for clean BC. ## Additional Information - **Homepage:** [deel-ai](https://www.irt-saintexupery.com/deel/) - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v3.0", "robot_type": "so101_follower", "total_episodes": 206, "total_frames": 84098, "total_tasks": 1, "chunks_size": 1000, "data_files_size_in_mb": 100, "video_files_size_in_mb": 500, "fps": 30, "splits": { "train": "0:206" }, "data_path": "data/chunk-{chunk_index:03d}/file-{file_index:03d}.parquet", "video_path": "videos/{video_key}/chunk-{chunk_index:03d}/file-{file_index:03d}.mp4", "features": { "action": { "dtype": "float32", "names": [ "shoulder_pan.pos", "shoulder_lift.pos", "elbow_flex.pos", "wrist_flex.pos", "wrist_roll.pos", "gripper.pos" ], "shape": [ 6 ] }, "observation.state": { "dtype": "float32", "names": [ "shoulder_pan.pos", "shoulder_lift.pos", "elbow_flex.pos", "wrist_flex.pos", "wrist_roll.pos", "gripper.pos" ], "shape": [ 6 ] }, "observation.images.left": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 30, "video.channels": 3, "has_audio": false } }, "observation.images.front": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channels" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 30, "video.channels": 3, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ```