socialbank commited on
Commit
be4f098
·
verified ·
1 Parent(s): b9eaa70

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +346 -3
README.md CHANGED
@@ -1,3 +1,346 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ annotations_creators:
3
+ - expert-generated
4
+ language_creators:
5
+ - found
6
+ language:
7
+ - ja
8
+ license: cc-by-nc-4.0
9
+ multilinguality:
10
+ - monolingual
11
+ size_categories:
12
+ - 1K<n<10K
13
+ source_datasets:
14
+ - original
15
+ task_categories:
16
+ - robotics
17
+ - object-detection
18
+ task_ids:
19
+ - imitation-learning
20
+ - action-segmentation
21
+ - pose-estimation
22
+ paperswithcode_id: null
23
+ pretty_name: FieldData Outdoor Labor Motion Dataset
24
+ tags:
25
+ - physical-ai
26
+ - embodied-ai
27
+ - motion-capture
28
+ - manipulation
29
+ - outdoor-robotics
30
+ - painting
31
+ - graffiti-removal
32
+ - construction
33
+ - human-robot-learning
34
+ ---
35
+
36
+ # Dataset Card: FieldData Outdoor Labor Motion Dataset (FOLMD)
37
+
38
+ ## Table of Contents
39
+
40
+ - [Dataset Description](#dataset-description)
41
+ - [Dataset Summary](#dataset-summary)
42
+ - [Supported Tasks](#supported-tasks)
43
+ - [Data Collection](#data-collection)
44
+ - [Dataset Structure](#dataset-structure)
45
+ - [Annotations](#annotations)
46
+ - [Data Fields](#data-fields)
47
+ - [Data Splits](#data-splits)
48
+ - [Dataset Creation](#dataset-creation)
49
+ - [Licensing Information](#licensing-information)
50
+ - [Citation](#citation)
51
+ - [Contact](#contact)
52
+
53
+ ---
54
+
55
+ ## Dataset Description
56
+
57
+ - **Homepage:** https://fielddata.jp
58
+ - **Repository:** https://huggingface.co/datasets/fielddata-jp/folmd
59
+ - **Paper:** [Coming soon]
60
+ - **Point of Contact:** contact@fielddata.jp
61
+
62
+ ### Dataset Summary
63
+
64
+ **FieldData Outdoor Labor Motion Dataset (FOLMD)** is a multi-modal motion dataset of professional outdoor workers performing real-world manual labor tasks in urban environments in Japan.
65
+
66
+ This dataset is designed to support the development of **physical AI foundation models** and **embodied robotic systems** targeting outdoor maintenance and construction tasks. All data was collected from professional workers at actual job sites — not from staged laboratory environments.
67
+
68
+ The v1.0 release covers a complete two-phase urban fence restoration task:
69
+ 1. **Graffiti removal** using chemical solvent and rag wiping
70
+ 2. **Repainting** with black paint using brushes at multiple heights and postures
71
+
72
+ Data modalities include egocentric RGB video, depth maps (Intel RealSense D435i), full-body IMU (8-point, Mbientlab MetaMotionS), glove-based grasp force (Tekscan Grip System), and annotated 3D skeleton sequences (SMPL format).
73
+
74
+ ---
75
+
76
+ ## Why This Dataset?
77
+
78
+ | Challenge for Robots | What This Dataset Provides |
79
+ |---|---|
80
+ | Vertical surface contact control | Sustained brush pressure on metal slat fence (9–25 N) |
81
+ | Low-posture precision work | Deep crouch / kneeling brushwork at ground level |
82
+ | Boundary-aware manipulation | Painting within 10 mm of masking tape edges |
83
+ | Multi-posture task transitions | Standing → crouching → kneeling within a single session |
84
+ | Visual quality assessment | Workers pausing to inspect coverage, detecting missed spots |
85
+ | Multi-agent coordination | 3–4 workers dividing fence height and working in parallel |
86
+
87
+ These tasks involve sustained contact with irregular vertical surfaces, fine motor control under physical load, and situated decision-making — all critical challenges for next-generation manipulation models.
88
+
89
+ ---
90
+
91
+ ## Supported Tasks
92
+
93
+ - **Imitation Learning (IL):** Action-annotated demonstrations suitable for behavior cloning and inverse reinforcement learning
94
+ - **Vision-Language-Action (VLA):** Paired egocentric video and action labels for multimodal training
95
+ - **Pose Estimation:** Full-body SMPL skeleton data across 6 posture categories
96
+ - **Action Segmentation:** 7-class taxonomy with fine-grained sub-actions and temporal boundaries
97
+ - **Force Control Learning:** Grasp pressure profiles for 5 grip types across different tools and surfaces
98
+ - **Contact-Rich Manipulation:** Depth + force data for learning compliant surface following
99
+
100
+ ---
101
+
102
+ ## Data Collection
103
+
104
+ ### Sensor Setup
105
+
106
+ | Modality | Device | Specs | Sample Rate |
107
+ |---|---|---|---|
108
+ | Egocentric RGB | GoPro Hero 12 | 1920×1080 | 60 fps |
109
+ | Depth + RGB | Intel RealSense D435i | 848×480 | 30 fps |
110
+ | Full-body IMU (×8) | Mbientlab MetaMotionS | Accel + Gyro + Euler | 200 Hz |
111
+ | Grasp Force Glove | Tekscan Grip System | 24 cells, right hand | 100 Hz |
112
+ | 3D Skeleton | Computed from IMU | SMPL format | 30 fps |
113
+ | External cameras | GoPro Hero 12 (×3) | 1920×1080 | 60 fps |
114
+
115
+ **Temporal synchronization:** All sensors synchronized via GPS-PPS hardware sync signal. Timestamp accuracy: ±0.8 ms.
116
+
117
+ ### Sensor Placement
118
+
119
+ ```
120
+ Head: GoPro (front) + RealSense (side)
121
+ R Wrist: IMU
122
+ L Wrist: IMU
123
+ R Elbow: IMU
124
+ L Elbow: IMU
125
+ Waist: IMU
126
+ R Knee: IMU
127
+ L Knee: IMU
128
+ R Hand: Force glove (Tekscan)
129
+ ```
130
+
131
+ ### Collection Protocol
132
+
133
+ - Workers are professional outdoor maintenance staff with 2–10 years of experience
134
+ - Written informed consent obtained from all participants
135
+ - Data collected at real urban job sites (Tokyo metropolitan area, Japan)
136
+ - Session begins with 5-minute sensor calibration and zeroing
137
+ - Workers perform tasks naturally without scripted movements
138
+ - All sessions include workers of mixed skill levels (novice / intermediate / experienced)
139
+
140
+ ---
141
+
142
+ ## Dataset Structure
143
+
144
+ ```
145
+ folmd-v1.0/
146
+ ├── sessions/
147
+ │ └── session_20210327_001/
148
+ │ ├── raw/
149
+ │ │ ├── gopro_head.mp4 # Egocentric RGB (60 fps)
150
+ │ │ ├── realsense_rgb.mp4 # RealSense RGB (30 fps)
151
+ │ │ ├── realsense_depth.bag # Depth stream (ROS bag)
152
+ │ │ ├── imu_r_wrist.csv # Right wrist IMU
153
+ │ │ ├── imu_l_wrist.csv # Left wrist IMU
154
+ │ │ ├── imu_r_elbow.csv # Right elbow IMU
155
+ │ │ ├── imu_l_elbow.csv # Left elbow IMU
156
+ │ │ ├── imu_waist.csv # Waist IMU
157
+ │ │ ├── imu_r_knee.csv # Right knee IMU
158
+ │ │ ├── imu_l_knee.csv # Left knee IMU
159
+ │ │ ├── imu_head.csv # Head IMU
160
+ │ │ └── force_glove_right.csv # Right hand force (24 cells)
161
+ │ ├── processed/
162
+ │ │ ├── skeleton_smpl/
163
+ │ │ │ └── frame_*.json # Per-keyframe SMPL skeleton
164
+ │ │ ├── depth_pointcloud/
165
+ │ │ │ └── frame_*.pcd # Per-keyframe 3D point cloud
166
+ │ │ └── action_segments.json # Full annotation (Tier 2)
167
+ │ └── metadata/
168
+ │ ├── session_info.json
169
+ │ ├── sensor_calibration.json
170
+ │ └── quality_report.pdf
171
+ ├── dataset_card.md # This file
172
+ ├── annotation_guidelines.pdf # Labeling rulebook (EN)
173
+ └── sample/
174
+ └── 30min_annotated_sample/ # Free sample — see Licensing
175
+ ```
176
+
177
+ ---
178
+
179
+ ## Annotations
180
+
181
+ ### Action Taxonomy (7 Primary Labels)
182
+
183
+ | Action Label | Description | Avg Duration (sec) |
184
+ |---|---|---|
185
+ | `solvent_application` | Applying chemical solvent to dissolve graffiti using rag | 45–180 |
186
+ | `scrub_with_rag` | Physical scrubbing to remove dissolved paint | 10–60 |
187
+ | `paint_application` | Applying paint with brush/roller at mid-to-upper height | 30–300 |
188
+ | `detail_paint_application` | Precision brushwork at edges, slat gaps, base boundaries | 10–120 |
189
+ | `visual_inspection` | Standing back to scan surface for quality assessment | 5–30 |
190
+ | `material_preparation` | Handling tools, mixing paint, preparing supplies | 5–60 |
191
+ | `masking` | Applying or adjusting masking tape | 30–180 |
192
+
193
+ ### Posture Labels (6 Categories)
194
+
195
+ ```
196
+ standing_upright — full height, minimal trunk lean
197
+ standing_slight_lean — 10–20° trunk forward flexion
198
+ crouching_moderate — knee flexion 60–90°
199
+ crouching_deep — knee flexion > 90°
200
+ kneeling_single_knee — one knee on ground
201
+ kneeling_double_knee — both knees on ground
202
+ ```
203
+
204
+ ### Force Labels (Qualitative)
205
+
206
+ ```
207
+ minimal < 5 N (passive hold / inspection)
208
+ low 5–10 N (precision brush / fine detail)
209
+ medium 10–20 N (normal painting stroke)
210
+ high 20–30 N (scrubbing / removal force)
211
+ very_high > 30 N (heavy scrub on resistant surface)
212
+ ```
213
+
214
+ ### Annotation Layers
215
+
216
+ | Layer | Content | Completeness |
217
+ |---|---|---|
218
+ | Layer 1 | Action segmentation (start/end time + label) | 100% |
219
+ | Layer 2 | Object + surface + environment labels | 100% |
220
+ | Layer 3 | Skill quality label (expert/intermediate/novice) | 100% |
221
+ | Layer 4 | Failure cases and recovery actions | 100% |
222
+
223
+ **Inter-annotator agreement (Cohen's κ):** 0.94
224
+
225
+ ---
226
+
227
+ ## Data Fields
228
+
229
+ ### `action_segments.json`
230
+
231
+ | Field | Type | Description |
232
+ |---|---|---|
233
+ | `session_id` | string | Unique session identifier |
234
+ | `frame_id` | string | Keyframe identifier |
235
+ | `timestamp_sec` | float | Time from session start (seconds) |
236
+ | `phase` | string | `graffiti_removal` or `repaint_application` |
237
+ | `action_label` | string | Primary action (see taxonomy above) |
238
+ | `sub_action` | string | Fine-grained sub-action description |
239
+ | `worker_id` | string | Worker identifier (W01–W04) |
240
+ | `body_posture` | string | Posture category |
241
+ | `working_height` | string | Vertical zone on fence being worked |
242
+
243
+ ### `imu_*.csv`
244
+
245
+ | Field | Type | Unit | Description |
246
+ |---|---|---|---|
247
+ | `timestamp_sec` | float | s | Synchronized timestamp |
248
+ | `accel_x/y/z_g` | float | g | Linear acceleration |
249
+ | `gyro_x/y/z_dps` | float | °/s | Angular velocity |
250
+ | `euler_roll/pitch/yaw_deg` | float | ° | Fused orientation estimate |
251
+
252
+ ### `force_glove_right.csv`
253
+
254
+ | Field | Type | Unit | Description |
255
+ |---|---|---|---|
256
+ | `timestamp_sec` | float | s | Synchronized timestamp |
257
+ | `cell_{name}_N` | float | N | Force at each of 24 sensor cells |
258
+ | `total_force_N` | float | N | Sum across all active cells |
259
+ | `grip_type` | string | — | Inferred grip classification |
260
+
261
+ ### `realsense_depth.bag` (ROS bag)
262
+
263
+ | Topic | Format | Description |
264
+ |---|---|---|
265
+ | `/camera/depth/image_rect_raw` | sensor_msgs/Image | 16-bit depth (mm) |
266
+ | `/camera/color/image_raw` | sensor_msgs/Image | RGB aligned to depth |
267
+ | `/camera/depth/color/points` | sensor_msgs/PointCloud2 | Fused RGBD point cloud |
268
+
269
+ ---
270
+
271
+ ## Data Splits
272
+
273
+ | Split | Sessions | Hours | Notes |
274
+ |---|---|---|---|
275
+ | Sample (free) | 1 | 0.5 | Available without license agreement |
276
+ | Full v1.0 | 10 | 50 | Commercial license required |
277
+ | Planned v2.0 | 50+ | 300+ | Additional tasks: garbage collection, weeding, building cleaning |
278
+
279
+ ---
280
+
281
+ ## Dataset Creation
282
+
283
+ ### Motivation
284
+
285
+ Physical AI and embodied robot models are rapidly advancing, but publicly available training data for **outdoor real-world manipulation tasks** remains extremely scarce. Existing motion datasets focus primarily on laboratory manipulation, household tasks, or driving — leaving a large gap in construction, maintenance, and urban service work.
286
+
287
+ This dataset addresses that gap by collecting data from professional Japanese workers performing skilled outdoor maintenance tasks. Japan's workforce is recognized globally for precision, safety discipline, and consistent execution — making it an ideal source of high-quality demonstration data.
288
+
289
+ ### Collection Methodology
290
+
291
+ - Task protocols defined in cooperation with professional site supervisors
292
+ - Sessions conducted at real operational job sites
293
+ - Multiple skill levels included to capture expert vs. novice motion differences
294
+ - Both "successful" and "recovery from failure" motion captured and labeled
295
+ - All data cleaned, synchronized, and formatted to be pipeline-ready
296
+
297
+ ### Known Limitations
298
+
299
+ - All data collected in Tokyo metropolitan area — urban environment bias
300
+ - Nighttime and rainy-weather sessions represent < 15% of current dataset
301
+ - Force glove covers right hand only
302
+ - Skeleton estimation accuracy degrades at extreme joint angles (> 110°)
303
+
304
+ ---
305
+
306
+ ## Licensing Information
307
+
308
+ This dataset is released under **Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0)**.
309
+
310
+ - ✅ Free to use for academic and non-commercial research
311
+ - ✅ Free 30-minute sample available without registration
312
+ - ✅ Redistribution permitted with attribution
313
+ - ❌ Commercial use requires a separate commercial license
314
+
315
+ For commercial licensing inquiries, please contact: **contact@fielddata.jp**
316
+
317
+ ---
318
+
319
+ ## Citation
320
+
321
+ If you use this dataset in your research, please cite:
322
+
323
+ ```bibtex
324
+ @dataset{fielddata2025folmd,
325
+ author = {FieldData Japan},
326
+ title = {FieldData Outdoor Labor Motion Dataset (FOLMD) v1.0},
327
+ year = {2025},
328
+ publisher = {HuggingFace},
329
+ url = {https://huggingface.co/datasets/fielddata-jp/folmd},
330
+ note = {Multi-modal motion dataset for physical AI training — outdoor maintenance tasks, Tokyo, Japan}
331
+ }
332
+ ```
333
+
334
+ ---
335
+
336
+ ## Contact
337
+
338
+ **FieldData Japan**
339
+ - Email: contact@fielddata.jp
340
+ - HuggingFace: [fielddata-jp](https://huggingface.co/fielddata-jp)
341
+ - X (Twitter): [@fielddata_jp](https://twitter.com/fielddata_jp)
342
+
343
+ We welcome collaboration inquiries from robotics researchers and companies interested in:
344
+ - Custom data collection for specific task domains
345
+ - Expanding the dataset to new outdoor task categories
346
+ - Joint research and co-authorship opportunities