--- license: apache-2.0 task_categories: - robotics language: - en tags: - LeRobot configs: - config_name: default data_files: data/*/*.parquet --- This dataset was created using [LeRobot](https://github.com/huggingface/lerobot). ## Dataset Description This dataset provides the robotic trajectories and observations used in the paper [VITA: Vision-to-Action Flow Matching Policy](https://huggingface.co/papers/2507.13231). VITA introduces a noise-free and conditioning-free policy learning framework that directly maps visual representations to latent actions using flow matching, enabling faster inference for robotic manipulation tasks. The datasets are built on [LeRobot](https://github.com/huggingface/lerobot) Hugging Face formats and optimized into offline `zarr` for faster training. - **Homepage:** https://ucd-dare.github.io/VITA/ - **Paper:** https://huggingface.co/papers/2507.13231 - **Code:** https://github.com/ucd-dare/VITA - **License:** apache-2.0 ## Dataset Structure [meta/info.json](meta/info.json): ```json { "codebase_version": "v2.1", "robot_type": null, "total_episodes": 175, "total_frames": 26266, "total_tasks": 1, "total_videos": 350, "total_chunks": 1, "chunks_size": 1000, "fps": 20, "splits": { "train": "0:175" }, "data_path": "data/chunk-{episode_chunk:03d}/episode_{episode_index:06d}.parquet", "video_path": "videos/chunk-{episode_chunk:03d}/{video_key}/episode_{episode_index:06d}.mp4", "features": { "action": { "dtype": "float32", "shape": [ 7 ], "names": null }, "action.delta": { "dtype": "float32", "shape": [ 7 ], "names": null }, "action.absolute": { "dtype": "float32", "shape": [ 7 ], "names": null }, "observation.state": { "dtype": "float32", "shape": [ 43 ], "names": null }, "observation.environment_state": { "dtype": "float32", "shape": [ 14 ], "names": null }, "observation.images.agentview_image": { "dtype": "video", "shape": [ 256, 256, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.fps": 20.0, "video.height": 256, "video.width": 256, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "observation.images.robot0_eye_in_hand_image": { "dtype": "video", "shape": [ 256, 256, 3 ], "names": [ "height", "width", "channel" ], "info": { "video.fps": 20.0, "video.height": 256, "video.width": 256, "video.channels": 3, "video.codec": "av1", "video.pix_fmt": "yuv420p", "video.is_depth_map": false, "has_audio": false } }, "timestamp": { "dtype": "float32", "shape": [ 1 ], "names": null }, "frame_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "episode_index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "index": { "dtype": "int64", "shape": [ 1 ], "names": null }, "task_index": { "dtype": "int64", "shape": [ 1 ], "names": null } } } ``` ## Sample Usage This dataset is designed to be used with the VITA codebase, which extends [LeRobot](https://github.com/huggingface/lerobot). Below are examples for converting datasets to an optimized `zarr` format and training a VITA policy. First, ensure the VITA repository is cloned and setup, and the `FLARE_DATASETS_DIR` environment variable is set as described in the [VITA GitHub repository](https://github.com/ucd-dare/VITA#%EF%B8%8F-setup). ### Dataset Preprocessing To list available datasets: ```bash cd gym-av-aloha/scripts python convert.py --ls ``` To convert a HuggingFace dataset to an offline `zarr` format (e.g., `av_aloha_sim_hook_package`): ```bash python convert.py -r iantc104/av_aloha_sim_hook_package ``` ### Training a VITA Policy Once the dataset is converted, you can train a VITA policy using the `flare` module from the VITA codebase: ```bash python flare/train.py policy=vita task=hook_package session=test ``` You can override default configurations as needed: ```bash # Example: Use a specific GPU python flare/train.py policy=vita task=hook_package session=test device=cuda:2 # Example: Change online validation frequency and episodes python flare/train.py policy=vita task=hook_package session=test \ val.val_online_freq=2000 val.eval_n_episodes=10 # Example: Run an ablation python flare/train.py policy=vita task=hook_package session=ablate \ policy.vita.decode_flow_latents=False wandb.notes=ablation ``` ## Citation **BibTeX:** ```bibtex @article{gao2025vita, title={VITA: Vision-to-Action Flow Matching Policy}, author={Gao, Dechen and Zhao, Boqi and Lee, Andrew and Chuang, Ian and Zhou, Hanchu and Wang, Hang and Zhao, Zhe and Zhang, Junshan and Soltani, Iman}, journal={arXiv preprint arXiv:2507.13231}, year={2025} } ```