| --- |
| license: mit |
| task_categories: |
| - robotics |
| tags: |
| - robotics |
| - manipulation |
| - contact-rich-manipulation |
| - force-torque |
| - imitation-learning |
| - flow-matching |
| - zarr |
| pretty_name: ForceFlow Dataset |
| size_categories: |
| - 10G<n<100G |
| --- |
| |
| # ForceFlow Dataset |
|
|
| **ForceFlow: Learning to Feel and Act via Contact-Driven Flow Matching** |
|
|
| [[Project Page](https://jokeresc.github.io/ForceFlow-page)] | [[Code](https://github.com/JokerESC/ForceFlow)] |
|
|
|  |
|
|
| Contact-rich manipulation remains one of the hardest problems in robot learning: vision alone cannot capture the high-frequency contact dynamics that determine whether a plug seats correctly, a stamp triggers cleanly, or a wipe exerts consistent pressure. This dataset is collected to support **ForceFlow**, a force-aware reactive framework built on flow matching that addresses this gap. |
|
|
| ForceFlow fuses temporal force/torque history with visual observations through an asymmetric multimodal design — force history acts as a global regulation signal to prevent it from being overshadowed by high-dimensional image features, while a hybrid action space jointly predicts end-effector motion and expected next-step contact force. To handle spatial generalization, ForceFlow introduces a **Vision-to-Force (V2F) handover**: a VLM first localizes the target in the scene, then control passes to the force-aware policy for precise local contact interaction. |
|
|
| This dataset contains **7 real-robot teleoperated demonstration tasks** spanning two categories of contact-rich manipulation, collected on a UFACTORY xArm6 equipped with a 6-axis wrist F/T sensor and dual Intel RealSense cameras. |
|
|
| --- |
|
|
| ## Tasks |
|
|
| **Short-horizon contact** — tasks requiring precise force application at a specific moment: |
|
|
| | Task | Episodes | Total Steps | Key Challenge | |
| |---|---|---|---| |
| | `stamp` | 100 | 45,867 | Visual ambiguity in paper thickness; force-triggered stamping | |
| | `plug` | 100 | 50,107 | Coarse visual alignment with force-guided insertion | |
| | `press_button` | 50 | 23,396 | Varying spring constants and trigger depths | |
| | `insert` | 50 | 25,032 | Sub-millimeter tolerance and geometric jamming | |
|
|
| **Continuous contact** — tasks requiring sustained force regulation throughout execution: |
|
|
| | Task | Episodes | Total Steps | Key Challenge | |
| |---|---|---|---| |
| | `clean_whiteboard` | 100 | 56,810 | Stable normal force tracking on a planar surface | |
| | `clean_vase` | 50 | 85,478 | Adaptive force regulation on a curved, non-linear surface | |
| | `peel` | 50 | 38,564 | Consistent peel force on adhesive tape | |
|
|
| --- |
|
|
| ## Data Format |
|
|
| Each task is provided in two formats: |
|
|
| - **`<task>.zarr/`** — Zarr v2 directory store, ready for direct training use |
| - **`<task>.zip`** — Zipped archive of the same zarr store |
| - **`<task>_normalizer.json`** — Pre-computed normalizer statistics (mean/std) for all fields |
| |
| ### Zarr Structure |
| |
| ``` |
| <task>.zarr/ |
| ├── data/ |
| │ ├── action (N, 6) float32 — end-effector delta pose (6-DOF) |
| │ ├── pos (N, 6) float32 — end-effector absolute pose |
| │ ├── force (N, 6) float32 — raw F/T sensor readings |
| │ ├── delta_force (N, 6) float32 — force delta (not in `peel`) |
| │ ├── gripper_action (N, 1) float32 — gripper command (0=open, 1=close) |
| │ ├── gripper_state (N, 1) float32 — gripper current state |
| │ ├── rgb_arm (N, 3, 240, 320) uint8 — wrist camera (JPEG-compressed) |
| │ └── rgb_fix (N, 3, 240, 320) uint8 — fixed camera (JPEG-compressed) |
| └── meta/ |
| └── episode_ends (E,) uint32 — cumulative step index at each episode end |
| ``` |
| |
| > **Note:** The `peel` task does not contain the `delta_force` field. |
| |
| RGB arrays are stored with a custom JPEG codec. To read them, install [image_codecs](https://github.com/JokerESC/ForceFlow/tree/main/CleanDiffuser/image_codecs) from the ForceFlow repo and register the codec before opening the zarr store. |
| |
| --- |
| |
| ## Usage |
| |
| ### Prerequisites |
| |
| ```bash |
| git clone --recurse-submodules https://github.com/JokerESC/ForceFlow.git |
| cd ForceFlow |
| pip install -r requirements.txt |
| pip install -e CleanDiffuser/ |
| ``` |
| |
| ### Load a dataset |
| |
| ```python |
| import sys |
| sys.path.insert(0, 'path/to/ForceFlow/CleanDiffuser') |
| |
| import numcodecs |
| import image_codecs |
| numcodecs.register_codec(image_codecs.jpeg) |
| |
| import zarr |
| import numpy as np |
| |
| z = zarr.open('plug.zarr', 'r') |
| |
| episode_ends = z['meta/episode_ends'][:] # shape (100,) |
| actions = z['data/action'][:] # shape (50107, 6) |
| forces = z['data/force'][:] # shape (50107, 6) |
| rgb_arm = z['data/rgb_arm'][:] # shape (50107, 3, 240, 320) |
| |
| # Reconstruct per-episode slices |
| starts = np.concatenate([[0], episode_ends[:-1]]) |
| for ep_idx, (s, e) in enumerate(zip(starts, episode_ends)): |
| ep_actions = actions[s:e] # (T, 6) |
| ep_forces = forces[s:e] # (T, 6) |
| ``` |
| |
| ### Training with ForceFlow |
| |
| ```bash |
| # Edit configs/xarm.yaml to point to the downloaded data |
| python -m pipeline.train --config configs/xarm.yaml |
| ``` |
| |
| --- |
| |
| ## Hardware |
| |
| | Component | Details | |
| |---|---| |
| | Robot arm | UFACTORY xArm6 | |
| | F/T sensor | 6-axis wrist force/torque sensor | |
| | Wrist camera | Intel RealSense D435 | |
| | Fixed camera | Intel RealSense L515 | |
| | Teleoperation | 3Dconnexion SpaceMouse | |
| |
| --- |
| |
| ## License |
| |
| MIT — see [LICENSE](https://github.com/JokerESC/ForceFlow/blob/main/LICENSE). |
| |
| --- |
| |
| ## Citation |
| |
| If you use this dataset, please cite: |
| |
| ```bibtex |
| @misc{forceflow2025, |
| title = {ForceFlow: Learning to Feel and Act via Contact-Driven Flow Matching}, |
| author = {JokerESC}, |
| year = {2025}, |
| url = {https://github.com/JokerESC/ForceFlow} |
| } |
| ``` |
| |