nielsr's picture
nielsr HF Staff
Improve dataset card: Add paper/code links, description, sample usage, citation, and language tag
224b3ab verified
|
raw
history blame
6.12 kB
---
license: apache-2.0
task_categories:
- robotics
language:
- en
tags:
- LeRobot
configs:
- config_name: default
data_files: data/*/*.parquet
---
This dataset was created using [LeRobot](https://github.com/huggingface/lerobot).
## Dataset Description
This dataset provides the robotic trajectories and observations used in the paper [VITA: Vision-to-Action Flow Matching Policy](https://huggingface.co/papers/2507.13231). VITA introduces a noise-free and conditioning-free policy learning framework that directly maps visual representations to latent actions using flow matching, enabling faster inference for robotic manipulation tasks. The datasets are built on [LeRobot](https://github.com/huggingface/lerobot) Hugging Face formats and optimized into offline `zarr` for faster training.
- **Homepage:** https://ucd-dare.github.io/VITA/
- **Paper:** https://huggingface.co/papers/2507.13231
- **Code:** https://github.com/ucd-dare/VITA
- **License:** apache-2.0
## Dataset Structure
[meta/info.json](meta/info.json):
```json
{
"codebase_version": "v2.1",
"robot_type": null,
"total_episodes": 175,
"total_frames": 26266,
"total_tasks": 1,
"total_videos": 350,
"total_chunks": 1,
"chunks_size": 1000,
"fps": 20,
"splits": {
"train": "0:175"
},
"data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet",
"video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4",
"features": {
"action": {
"dtype": "float32",
"shape": [
7
],
"names": null
},
"action.delta": {
"dtype": "float32",
"shape": [
7
],
"names": null
},
"action.absolute": {
"dtype": "float32",
"shape": [
7
],
"names": null
},
"observation.state": {
"dtype": "float32",
"shape": [
43
],
"names": null
},
"observation.environment_state": {
"dtype": "float32",
"shape": [
14
],
"names": null
},
"observation.images.agentview_image": {
"dtype": "video",
"shape": [
256,
256,
3
],
"names": [
"height",
"width",
"channel"
],
"info": {
"video.fps": 20.0,
"video.height": 256,
"video.width": 256,
"video.channels": 3,
"video.codec": "av1",
"video.pix_fmt": "yuv420p",
"video.is_depth_map": false,
"has_audio": false
}
},
"observation.images.robot0_eye_in_hand_image": {
"dtype": "video",
"shape": [
256,
256,
3
],
"names": [
"height",
"width",
"channel"
],
"info": {
"video.fps": 20.0,
"video.height": 256,
"video.width": 256,
"video.channels": 3,
"video.codec": "av1",
"video.pix_fmt": "yuv420p",
"video.is_depth_map": false,
"has_audio": false
}
},
"timestamp": {
"dtype": "float32",
"shape": [
1
],
"names": null
},
"frame_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"episode_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"index": {
"dtype": "int64",
"shape": [
1
],
"names": null
},
"task_index": {
"dtype": "int64",
"shape": [
1
],
"names": null
}
}
}
```
## Sample Usage
This dataset is designed to be used with the VITA codebase, which extends [LeRobot](https://github.com/huggingface/lerobot). Below are examples for converting datasets to an optimized `zarr` format and training a VITA policy.
First, ensure the VITA repository is cloned and setup, and the `FLARE_DATASETS_DIR` environment variable is set as described in the [VITA GitHub repository](https://github.com/ucd-dare/VITA#%EF%B8%8F-setup).
### Dataset Preprocessing
To list available datasets:
```bash
cd gym-av-aloha/scripts
python convert.py --ls
```
To convert a HuggingFace dataset to an offline `zarr` format (e.g., `av_aloha_sim_hook_package`):
```bash
python convert.py -r iantc104/av_aloha_sim_hook_package
```
### Training a VITA Policy
Once the dataset is converted, you can train a VITA policy using the `flare` module from the VITA codebase:
```bash
python flare/train.py policy=vita task=hook_package session=test
```
You can override default configurations as needed:
```bash
# Example: Use a specific GPU
python flare/train.py policy=vita task=hook_package session=test device=cuda:2
# Example: Change online validation frequency and episodes
python flare/train.py policy=vita task=hook_package session=test \
val.val_online_freq=2000 val.eval_n_episodes=10
# Example: Run an ablation
python flare/train.py policy=vita task=hook_package session=ablate \
policy.vita.decode_flow_latents=False wandb.notes=ablation
```
## Citation
**BibTeX:**
```bibtex
@article{gao2025vita,
title={VITA: Vision-to-Action Flow Matching Policy},
author={Gao, Dechen and Zhao, Boqi and Lee, Andrew and Chuang, Ian and Zhou, Hanchu and Wang, Hang and Zhao, Zhe and Zhang, Junshan and Soltani, Iman},
journal={arXiv preprint arXiv:2507.13231},
year={2025}
}
```