--- license: cc-by-4.0 task_categories: - robotics tags: - LeRobot - Robotic manipulation pretty_name: BridgeData V2 Scripted Demos size_categories: - 100K For the teleoperated trajectories with language annotation, see [jnogga/bridge_data_v2_teleop](https://huggingface.co/datasets/jnogga/bridge_data_v2_teleop). ## Dataset Structure Note that the available cameras vary between episodes. Missing camera perspectives are padded, and the corresponding *_available* sample fields serve as a mask. [meta/info.json](meta/info.json): ```json { "codebase_version": "v3.0", "robot_type": "widow_x", "fps": 5, "data_files_size_in_mb": 100.0, "video_files_size_in_mb": 200.0, "chunks_size": 1000, "total_episodes": 9701, "total_frames": 456260, "total_tasks": 9701, "splits": { "train": "0:9701" }, "data_path": "data/chunk-{chunk_index:03d}/file_{file_index:03d}.parquet", "video_path": "videos/{video_key}/chunk-{chunk_index:03d}/file_{file_index:03d}.mp4", "features": { "action.cartesian": { "dtype": "float32", "shape": [ 7 ], "names": [ "position.x", "position.y", "position.z", "quaternion.w", "quaternion.x", "quaternion.y", "quaternion.z" ], "fps": 5 }, "action.gripper_position": { "dtype": "float32", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.cartesian": { "dtype": "float32", "shape": [ 7 ], "names": [ "position.x", "position.y", "position.z", "quaternion.w", "quaternion.x", "quaternion.y", "quaternion.z" ], "fps": 5 }, "observation.gripper_position": { "dtype": "float32", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.eef_transform": { "dtype": "float32", "shape": [ 7 ], "names": [ "position.x", "position.y", "position.z", "quaternion.w", "quaternion.x", "quaternion.y", "quaternion.z" ], "fps": 5 }, "observation.joint_position": { "dtype": "float32", "shape": [ 6 ], "names": [ "joint_0", "joint_1", "joint_2", "joint_3", "joint_4", "joint_5" ], "fps": 5 }, "observation.joint_velocity": { "dtype": "float32", "shape": [ 6 ], "names": [ "joint_0", "joint_1", "joint_2", "joint_3", "joint_4", "joint_5" ], "fps": 5 }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null, "fps": 5 }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null, "fps": 5 }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null, "fps": 5 }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null, "fps": 5 }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.images.camera_0_available": { "dtype": "bool", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.images.camera_1_available": { "dtype": "bool", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.images.camera_2_available": { "dtype": "bool", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.images.camera_3_available": { "dtype": "bool", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.images.camera_4_available": { "dtype": "bool", "shape": [ 1 ], "names": null, "fps": 5 }, "observation.images.camera_0": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 5, "video.channels": 3, "has_audio": false }, "fps": 5 }, "observation.images.camera_1": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 5, "video.channels": 3, "has_audio": false }, "fps": 5 }, "observation.images.camera_2": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 5, "video.channels": 3, "has_audio": false }, "fps": 5 }, "observation.images.camera_3": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 5, "video.channels": 3, "has_audio": false }, "fps": 5 }, "observation.images.camera_4": { "dtype": "video", "shape": [ 480, 640, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.height": 480, "video.width": 640, "video.codec": "h264", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "video.fps": 5, "video.channels": 3, "has_audio": false }, "fps": 5 } } } ``` ## Getting started ```py # pip install lerobot from lerobot.datasets.lerobot_dataset import LeRobotDataset dataset = LeRobotDataset("jnogga/bridge_data_v2_scripted") ``` See [bridge_example.ipynb](bridge_example.ipynb) for a more detailed example. ## Citation All credit goes to the original authors of BridgeData V2. If you find their work helpful, please cite **BibTeX:** ```bibtex @inproceedings{walke2023bridgedata, title={BridgeData V2: A Dataset for Robot Learning at Scale}, author={Walke, Homer and Black, Kevin and Lee, Abraham and Kim, Moo Jin and Du, Max and Zheng, Chongyi and Zhao, Tony and Hansen-Estruch, Philippe and Vuong, Quan and He, Andre and Myers, Vivek and Fang, Kuan and Finn, Chelsea and Levine, Sergey}, booktitle={Conference on Robot Learning (CoRL)}, year={2023} } ```