Bl4ckJack777 commited on
Commit
a00ec58
·
verified ·
1 Parent(s): 2116b11

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +143 -43
README.md CHANGED
@@ -1,45 +1,3 @@
1
- ---
2
- language:
3
- - en
4
- - zh
5
- tags:
6
- - robotics
7
- - manipulation
8
- - vla
9
- - trajectory-data
10
- - multimodal
11
- - vision-language-action
12
- license: other
13
- task_categories:
14
- - robotics
15
- - reinforcement-learning
16
- - computer-vision
17
- multimodal: vision+language+action
18
- dataset_info:
19
- features:
20
- - name: rgb_images
21
- dtype: image
22
- description: Multi-view RGB images
23
- - name: slam_poses
24
- sequence: float32
25
- description: SLAM pose trajectories
26
- - name: vive_poses
27
- sequence: float32
28
- description: Vive tracking system poses
29
- - name: point_clouds
30
- sequence: float32
31
- description: Time-of-Flight point cloud data
32
- - name: clamp_data
33
- sequence: float32
34
- description: Clamp sensor readings
35
- - name: merged_trajectory
36
- sequence: float32
37
- description: Fused trajectory data
38
- configs:
39
- - config_name: default
40
- data_files: "**/*"
41
- ---
42
-
43
  # FastUMI Pro Dataset
44
 
45
  ## Project Description
@@ -67,4 +25,146 @@ huggingface-cli download FastUMIPro/example_data_fastumi_pro_raw --repo-type dat
67
 
68
  # Mirror acceleration solution
69
  export HF_ENDPOINT=https://hf-mirror.com
70
- huggingface-cli download --repo-type dataset --resume-download FastUMIPro/example_data_fastumi_pro_raw --local-dir ~/fastumi_data/
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # FastUMI Pro Dataset
2
 
3
  ## Project Description
 
25
 
26
  # Mirror acceleration solution
27
  export HF_ENDPOINT=https://hf-mirror.com
28
+ huggingface-cli download --repo-type dataset --resume-download FastUMIPro/example_data_fastumi_pro_raw --local-dir ~/fastumi_data/
29
+
30
+ ## Data Structure
31
+ FastUMI PRO uses raw format containing various types of raw sensor data, which can be easily converted to other formats. The raw format facilitates querying and validating original sensor outputs for rapid problem identification.
32
+ DATA/
33
+ └── device_label_xv_serial/
34
+ └── session_timestamp/
35
+ ├── RGB_Images/
36
+ │ ├── timestamps.csv
37
+ │ └── Frames/
38
+ │ ├── frame_000001.jpg
39
+ │ ├── frame_000002.jpg
40
+ │ └── ...
41
+ ├── SLAM_Poses/
42
+ │ └── slam_raw.txt
43
+ ├── Vive_Poses/
44
+ │ └── vive_data_tum.txt
45
+ ├── ToF_PointClouds/
46
+ │ ├── timestamps.csv
47
+ │ └── PointClouds/
48
+ │ ├── pointcloud_000001.pcd
49
+ │ ├── pointcloud_000002.pcd
50
+ │ └── ...
51
+ ├── Clamp_Data/
52
+ │ └── clamp_data_tum.txt
53
+ └── Merged_Trajectory/
54
+ ├── merged_trajectory.txt
55
+ └── merge_stats.txt
56
+
57
+ ### Directory Descriptions
58
+ session_xxx: Individual data collection session
59
+
60
+ RGB_Images: Frame images supporting multiple viewpoints; supports both Images and Videos
61
+
62
+ SLAM_Poses: UMI pose data
63
+
64
+ Vive_Poses: Vive tracking system pose data
65
+
66
+ ToF_PointClouds: Time-of-Flight point cloud raw data (depth)
67
+
68
+ Merged_Trajectory: Trajectory data
69
+
70
+ ### Data Specifications
71
+
72
+ Attributes
73
+ sim:
74
+
75
+ False: Real environment data
76
+
77
+ True: Simulation data
78
+
79
+ Observations
80
+ observations/images/: Camera image data
81
+
82
+ Default camera name: front
83
+
84
+ Shape: (frames, 1920, 1080, 3)
85
+
86
+ Data type: uint8
87
+
88
+ Compression: gzip (level 4)
89
+
90
+ observations/qpos:
91
+
92
+ Type: Floating point dataset
93
+
94
+ Shape: (timesteps, 7)
95
+
96
+ Meaning: Robot end-effector position + quaternion orientation
97
+
98
+ Order: [Pos X, Pos Y, Pos Z, Q_X, Q_Y, Q_Z, Q_W]
99
+
100
+ Actions
101
+ Type: Floating point dataset
102
+
103
+ Shape: (timesteps, 7)
104
+
105
+ Meaning: Actions (same structure as qpos, typically mirroring qpos)
106
+
107
+ Data Conversion
108
+ Supports one-click export to specific formats via web toolchain, or conversion between formats using tools like:
109
+
110
+ Any4lerobot: GitHub - Tavish9/any4lerobot
111
+
112
+ Conversion paths supported:
113
+
114
+ hdf5 → lerobot v3.0
115
+
116
+ hdf5 → lerobot(Pi0) v2.0
117
+
118
+ hdf5 → rlds
119
+
120
+ Model Performance
121
+ Preliminary experiments show that models trained on this dataset demonstrate significant multi-task generalization capabilities in universal manipulation tasks:
122
+
123
+ VLA Models: Including PI-O models with language understanding and action planning capabilities, exhibiting excellent generalization and execution stability in multi-task language-conditioned control
124
+
125
+ VA Models: Classical visual control architectures like ACT, DP also show significant improvements, particularly in complex operation sequences, viewpoint perturbations, and fine motion tracking with enhanced robustness
126
+
127
+ Related Links
128
+ Project Homepage: https://fastumi.com/pro/
129
+
130
+ FastUMI Project: https://fastumi.com
131
+
132
+ Hugging Face Dataset: https://huggingface.co/datasets/IPE...
133
+
134
+ Research Paper: [2409.19499] FastUMI: A Scalable and...
135
+
136
+ Open Source Toolchain:
137
+
138
+ Demo Replay: GitHub - Loki-Lu/FastUMI_replay_sin...
139
+
140
+ Dual-arm Demo: GitHub - Loki-Lu/FastUMI_replay_du...
141
+
142
+ Hardware SDK: GitHub - FastUMIRobotics/FastUMI_...
143
+
144
+ Monitoring Tools: GitHub - FastUMIRobotics/FastUMI_...
145
+
146
+ Data Collection Tools: GitHub - FastUMIRobotics/FastUMI_...
147
+
148
+ Related Research
149
+ [2508.10538] MLM: Learning Multi-ta...
150
+
151
+ PIO (FastUMI Lightweight Adaptation, Version V0) Full Tutorial: PIO (FastUMI数据轻量级适配,版本V0)…
152
+
153
+ Citation
154
+ If you use this dataset in your research, please cite the relevant papers:
155
+
156
+ bibtex
157
+ @article{fastumi2024,
158
+ title={FastUMI: A Scalable and Hardware-Agnostic Framework for Robot Manipulation Learning},
159
+ author={FastUMI Team},
160
+ journal={arXiv preprint},
161
+ year={2024}
162
+ }
163
+ Contact
164
+ For any questions or suggestions, please contact the development team:
165
+
166
+ Lead: [Name]
167
+
168
+ Email: [Email Address]
169
+
170
+ WeChat: [WeChat ID]