Kai0 / README.md
OpenDriveLab-org's picture
Update README.md
ba73931 verified
---
license: mit
task_categories:
- robotics
tags:
- LeRobot
configs:
- config_name: default
data_files: FlattenFold/base/data/chunk-000/episode_000000.parquet
---
# KAI0
<div align="center">
<a href="">
<img src="https://img.shields.io/badge/GitHub-grey?logo=GitHub" alt="GitHub Badge">
</a>
<a href="https://mmlab.hk/research/kai0">
<img src="https://img.shields.io/badge/Research_Blog-grey?style=flat" alt="Research Blog Badge">
</a>
</div>
# TODO
- [ ] The advantage label will be coming soon.
## Contents
- [About the Dataset](#about-the-dataset)
- [Load the Dataset](#get-started)
- [Download the Dataset](#download-the-dataset)
- [Dataset Structure](#dataset-structure)
- [Folder hierarchy](#folder-hierarchy)
- [Details](#details)
- [License and Citation](#license-and-citation)
## [About the Dataset](#contents)
- **~134 hours** real world scenarios
- **Main Tasks**
- ***FlattenFold***
- Single task
- Initial state: T-shirts are randomly tossed onto the table, presenting random crumpled configurations
- Manipulation task: Operate the robotic arm to unfold the garment, then fold it
- ***HangCloth***
- Single task
- Initial state: Hanger is randomly placed, garment is randomly positioned on the table
- Manipulation task: Operate the robotic arm to thread the hanger through the garment, then hang it on the rod
- ***TeeShirtSort***
- Garment classification and arrangement task
- Initial state: Randomly pick a garment from the laundry basket
- Classification: Determine whether the garment is a T-shirt or a dress shirt
- Manipulation task:
- If it is a T-shirt, fold the garment
- If it is a dress shirt, expose the collar, then push it to one side of the table
- **Count of the dataset**
| Task | Base (episodes count/hours) | DAgger (episodes count/hours) | Total(episodes count/hours) |
|------|-----------------------------|-------------------------------|-----------------------------|
| FlattenFold | 3,055/~42 hours | 3,457/ ~13 Hours | 6,512 /~55 hours |
| HangCloth | 6954/~61 hours | 686/~12 hours | 7640/~73 hours |
| TeeShirtSort | 5988/~31 hours | 769/~22 hours | 6757/~53 hours |
| **Total** | **15,997/~134 hours** | **4,912/~47 hours** | **20,909/~181 hours** |
## [Load the dataset](#contents)
- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot)
- The dataset's version is LeRobotDataset v2.1
### For LeRobot version < 0.4.0
Choose the appropriate import based on your version:
| Version | Import Path |
|------------------------|-------------|
| `<= 0.1.0` | `from lerobot.common.datasets.lerobot_dataset import LeRobotDataset` |
| `> 0.1.0` and `< 0.4.0` | `from lerobot.datasets.lerobot_dataset import LeRobotDataset` |
```python
# For version <= 0.1.0
from lerobot.common.datasets.lerobot_dataset import LeRobotDataset
# For version > 0.1.0 and < 0.4.0
from lerobot.datasets.lerobot_dataset import LeRobotDataset
# Load the dataset
dataset = LeRobotDataset(repo_id='where/the/dataset/you/stored')
```
### For LeRobot version >= 0.4.0
You need to migrate the dataset from v2.1 to v3.0 first. See the official documentation: [Migrate the dataset from v2.1 to v3.0](https://huggingface.co/docs/lerobot/lerobot-dataset-v3)
```bash
python -m lerobot.datasets.v30.convert_dataset_v21_to_v30 --repo-id=<HF_USER/DATASET_ID>
```
## [Download the Dataset](#contents)
### Python Script
```python
from huggingface_hub import hf_hub_download, snapshot_download
from datasets import load_dataset
# Download a single file
hf_hub_download(
repo_id="OpenDriveLab-org/kai0",
filename="episodes.jsonl",
subfolder="meta",
repo_type="dataset",
local_dir="where/you/want/to/save"
)
# Download a specific folder
snapshot_download(
repo_id="OpenDriveLab-org/kai0",
local_dir="/where/you/want/to/save",
repo_type="dataset",
allow_patterns=["data/*"]
)
# Load the entire dataset
dataset = load_dataset("OpenDriveLab-org/kai0")
```
### Terminal (CLI)
```bash
# Download a single file
hf download OpenDriveLab-org/kai0 \
--include "meta/info.json" \
--repo-type dataset \
--local-dir "/where/you/want/to/save"
# Download a specific folder
hf download OpenDriveLab-org/kai0 \
--repo-type dataset \
--include "meta/*" \
--local-dir "/where/you/want/to/save"
# Download the entire dataset
hf download OpenDriveLab-org/kai0 \
--repo-type dataset \
--local-dir "/where/you/want/to/save"
```
## [Dataset Structure](#contents)
### [Folder hierarchy](#contents)
Under each task directory, data is partitioned into two subsets: base and dagger.
- base
contains
original demonstration trajectories of robotic arm manipulation for garment arrangement tasks.
- dagger
contains on-policy recovery trajectories collected via iterative DAgger, designed to populate failure recovery modes absent in static demonstrations.
```text
Kai0-data/
├── FlattenFold/
│ ├── base/
│ │ ├── data/
│ │ │ ├── chunk-000/
│ │ │ │ ├── episode_000000.parquet
│ │ │ │ ├── episode_000001.parquet
│ │ │ │ └── ...
│ │ │ └── ...
│ │ ├── videos/
│ │ │ ├── chunk-000/
│ │ │ │ ├── observation.images.hand_left/
│ │ │ │ │ ├── episode_000000.mp4
│ │ │ │ │ ├── episode_000001.mp4
│ │ │ │ │ └── ...
│ │ │ │ ├── observation.images.hand_right/
│ │ │ │ │ ├── episode_000000.mp4
│ │ │ │ │ ├── episode_000001.mp4
│ │ │ │ │ └── ...
│ │ │ │ ├── observation.images.top_head/
│ │ │ │ │ ├── episode_000000.mp4
│ │ │ │ │ ├── episode_000001.mp4
│ │ │ │ │ └── ...
│ │ │ │ └── ...
│ │ │ └── ...
│ │ └── meta/
│ │ ├── info.json
│ │ ├── episodes.jsonl
│ │ ├── tasks.jsonl
│ │ └── episodes_stats.jsonl
│ └── dagger/
├── HangCloth/
│ ├── base/
│ └── dagger/
├── TeeShirtSort/
│ ├── base/
│ └── dagger/
└── README.md
```
<a id='Details'></a>
### [Details](#contents)
#### info.json
the basic struct of the [info.json](#meta/info.json)
```json
{
"codebase_version": "v2.1",
"robot_type": "agilex",
"total_episodes": ..., # the total episodes in the dataset
"total_frames": ..., # The total number of video frames in any single camera perspective
"total_tasks": ..., # Total number of tasks
"total_videos": ..., # The total number of videos from all camera perspectives in the dataset
"total_chunks": ..., # The number of chunks in the dataset
"chunks_size": ..., # The max number of episodes in a chunk
"fps": ..., # Video frame rate per second
"splits": { # how to split the dataset
"train": ...
},
"data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet",
"video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4",
"features": {
"observation.images.top_head": { # the camera perspective
"dtype": "video",
"shape": [
480,
640,
3
],
"names": [
"height",
"width",
"channel"
],
"info": {
"video.height": 480,
"video.width": 640,
"video.codec": "av1",
"video.pix_fmt": "yuv420p",
"video.is_depth_map": false,
"video.fps": 30,
"video.channels": 3,
"has_audio": false
}
},
"observation.images.hand_left": { # the camera perspective
...
},
"observation.images.hand_right": { # the camera perspective
...
},
"observation.state": {
"dtype": "float32",
"shape": [
14
],
"names": null
},
"action": {
"dtype": "float32",
"shape": [
14
],
"names": null
},
"timestamp": {
"dtype": "float32",
"shape": [
1
],
"names": null
},
"frame_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"episode_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"task_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
}
}
}
```
#### [Parquet file format](#contents)
| Field Name | shape | Meaning |
|------------|-------------|-------------|
| observation.state | [N, 14] |left `[:, :6]`, right `[:, 7:13]`, joint angle<br> left`[:, 6]`, right `[:, 13]` , gripper open range|
| action | [N, 14] |left `[:, :6]`, right `[:, 7:13]`, joint angle<br>left`[:, 6]`, right `[:, 13]` , gripper open range |
| timestamp | [N, 1] | Time elapsed since the start of the episode (in seconds) |
| frame_index | [N, 1] | Index of this frame within the current episode (0-indexed) |
| episode_index | [N, 1] | Index of the episode this frame belongs to |
| index | [N, 1] | Global unique index across all frames in the dataset |
| task_index | [N, 1] | Index identifying the task type being performed |
### [tasks.jsonl](#FlattenFold/meta/tasks.jsonl)
Contains task language prompts (natural language instructions) that specify the manipulation task to be performed. Each entry maps a task_index to its corresponding task description, which can be used for language-conditioned policy training.
# License and Citation
All the data and code within this repo are under [](). Please consider citing our project if it helps your research.
```BibTeX
@misc{,
title={},
author={},
howpublished={\url{}},
year={}
}