File size: 13,077 Bytes
be4f098
 
 
 
 
 
 
 
 
 
 
 
 
 
 
aff8a55
be4f098
 
aff8a55
 
be4f098
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
---
annotations_creators:
  - expert-generated
language_creators:
  - found
language:
  - ja
license: cc-by-nc-4.0
multilinguality:
  - monolingual
size_categories:
  - 1K<n<10K
source_datasets:
  - original
task_categories:
  - other
task_ids:
  - pose-estimation
  - grasping
  - task-planning
paperswithcode_id: null
pretty_name: FieldData Outdoor Labor Motion Dataset
tags:
  - physical-ai
  - embodied-ai
  - motion-capture
  - manipulation
  - outdoor-robotics
  - painting
  - graffiti-removal
  - construction
  - human-robot-learning
---

# Dataset Card: FieldData Outdoor Labor Motion Dataset (FOLMD)

## Table of Contents

- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks)
- [Data Collection](#data-collection)
- [Dataset Structure](#dataset-structure)
- [Annotations](#annotations)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Licensing Information](#licensing-information)
- [Citation](#citation)
- [Contact](#contact)

---

## Dataset Description

- **Homepage:** https://fielddata.jp
- **Repository:** https://huggingface.co/datasets/fielddata-jp/folmd
- **Paper:** [Coming soon]
- **Point of Contact:** contact@fielddata.jp

### Dataset Summary

**FieldData Outdoor Labor Motion Dataset (FOLMD)** is a multi-modal motion dataset of professional outdoor workers performing real-world manual labor tasks in urban environments in Japan.

This dataset is designed to support the development of **physical AI foundation models** and **embodied robotic systems** targeting outdoor maintenance and construction tasks. All data was collected from professional workers at actual job sites — not from staged laboratory environments.

The v1.0 release covers a complete two-phase urban fence restoration task:
1. **Graffiti removal** using chemical solvent and rag wiping
2. **Repainting** with black paint using brushes at multiple heights and postures

Data modalities include egocentric RGB video, depth maps (Intel RealSense D435i), full-body IMU (8-point, Mbientlab MetaMotionS), glove-based grasp force (Tekscan Grip System), and annotated 3D skeleton sequences (SMPL format).

---

## Why This Dataset?

| Challenge for Robots | What This Dataset Provides |
|---|---|
| Vertical surface contact control | Sustained brush pressure on metal slat fence (9–25 N) |
| Low-posture precision work | Deep crouch / kneeling brushwork at ground level |
| Boundary-aware manipulation | Painting within 10 mm of masking tape edges |
| Multi-posture task transitions | Standing → crouching → kneeling within a single session |
| Visual quality assessment | Workers pausing to inspect coverage, detecting missed spots |
| Multi-agent coordination | 3–4 workers dividing fence height and working in parallel |

These tasks involve sustained contact with irregular vertical surfaces, fine motor control under physical load, and situated decision-making — all critical challenges for next-generation manipulation models.

---

## Supported Tasks

- **Imitation Learning (IL):** Action-annotated demonstrations suitable for behavior cloning and inverse reinforcement learning
- **Vision-Language-Action (VLA):** Paired egocentric video and action labels for multimodal training
- **Pose Estimation:** Full-body SMPL skeleton data across 6 posture categories
- **Action Segmentation:** 7-class taxonomy with fine-grained sub-actions and temporal boundaries
- **Force Control Learning:** Grasp pressure profiles for 5 grip types across different tools and surfaces
- **Contact-Rich Manipulation:** Depth + force data for learning compliant surface following

---

## Data Collection

### Sensor Setup

| Modality | Device | Specs | Sample Rate |
|---|---|---|---|
| Egocentric RGB | GoPro Hero 12 | 1920×1080 | 60 fps |
| Depth + RGB | Intel RealSense D435i | 848×480 | 30 fps |
| Full-body IMU (×8) | Mbientlab MetaMotionS | Accel + Gyro + Euler | 200 Hz |
| Grasp Force Glove | Tekscan Grip System | 24 cells, right hand | 100 Hz |
| 3D Skeleton | Computed from IMU | SMPL format | 30 fps |
| External cameras | GoPro Hero 12 (×3) | 1920×1080 | 60 fps |

**Temporal synchronization:** All sensors synchronized via GPS-PPS hardware sync signal. Timestamp accuracy: ±0.8 ms.

### Sensor Placement

```
Head:     GoPro (front) + RealSense (side)
R Wrist:  IMU
L Wrist:  IMU
R Elbow:  IMU
L Elbow:  IMU
Waist:    IMU
R Knee:   IMU
L Knee:   IMU
R Hand:   Force glove (Tekscan)
```

### Collection Protocol

- Workers are professional outdoor maintenance staff with 2–10 years of experience
- Written informed consent obtained from all participants
- Data collected at real urban job sites (Tokyo metropolitan area, Japan)
- Session begins with 5-minute sensor calibration and zeroing
- Workers perform tasks naturally without scripted movements
- All sessions include workers of mixed skill levels (novice / intermediate / experienced)

---

## Dataset Structure

```
folmd-v1.0/
├── sessions/
│   └── session_20210327_001/
│       ├── raw/
│       │   ├── gopro_head.mp4            # Egocentric RGB (60 fps)
│       │   ├── realsense_rgb.mp4         # RealSense RGB (30 fps)
│       │   ├── realsense_depth.bag       # Depth stream (ROS bag)
│       │   ├── imu_r_wrist.csv           # Right wrist IMU
│       │   ├── imu_l_wrist.csv           # Left wrist IMU
│       │   ├── imu_r_elbow.csv           # Right elbow IMU
│       │   ├── imu_l_elbow.csv           # Left elbow IMU
│       │   ├── imu_waist.csv             # Waist IMU
│       │   ├── imu_r_knee.csv            # Right knee IMU
│       │   ├── imu_l_knee.csv            # Left knee IMU
│       │   ├── imu_head.csv              # Head IMU
│       │   └── force_glove_right.csv     # Right hand force (24 cells)
│       ├── processed/
│       │   ├── skeleton_smpl/
│       │   │   └── frame_*.json          # Per-keyframe SMPL skeleton
│       │   ├── depth_pointcloud/
│       │   │   └── frame_*.pcd           # Per-keyframe 3D point cloud
│       │   └── action_segments.json      # Full annotation (Tier 2)
│       └── metadata/
│           ├── session_info.json
│           ├── sensor_calibration.json
│           └── quality_report.pdf
├── dataset_card.md                       # This file
├── annotation_guidelines.pdf            # Labeling rulebook (EN)
└── sample/
    └── 30min_annotated_sample/           # Free sample — see Licensing
```

---

## Annotations

### Action Taxonomy (7 Primary Labels)

| Action Label | Description | Avg Duration (sec) |
|---|---|---|
| `solvent_application` | Applying chemical solvent to dissolve graffiti using rag | 45–180 |
| `scrub_with_rag` | Physical scrubbing to remove dissolved paint | 10–60 |
| `paint_application` | Applying paint with brush/roller at mid-to-upper height | 30–300 |
| `detail_paint_application` | Precision brushwork at edges, slat gaps, base boundaries | 10–120 |
| `visual_inspection` | Standing back to scan surface for quality assessment | 5–30 |
| `material_preparation` | Handling tools, mixing paint, preparing supplies | 5–60 |
| `masking` | Applying or adjusting masking tape | 30–180 |

### Posture Labels (6 Categories)

```
standing_upright          — full height, minimal trunk lean
standing_slight_lean      — 10–20° trunk forward flexion
crouching_moderate        — knee flexion 60–90°
crouching_deep            — knee flexion > 90°
kneeling_single_knee      — one knee on ground
kneeling_double_knee      — both knees on ground
```

### Force Labels (Qualitative)

```
minimal    < 5 N    (passive hold / inspection)
low        5–10 N   (precision brush / fine detail)
medium     10–20 N  (normal painting stroke)
high       20–30 N  (scrubbing / removal force)
very_high  > 30 N   (heavy scrub on resistant surface)
```

### Annotation Layers

| Layer | Content | Completeness |
|---|---|---|
| Layer 1 | Action segmentation (start/end time + label) | 100% |
| Layer 2 | Object + surface + environment labels | 100% |
| Layer 3 | Skill quality label (expert/intermediate/novice) | 100% |
| Layer 4 | Failure cases and recovery actions | 100% |

**Inter-annotator agreement (Cohen's κ):** 0.94

---

## Data Fields

### `action_segments.json`

| Field | Type | Description |
|---|---|---|
| `session_id` | string | Unique session identifier |
| `frame_id` | string | Keyframe identifier |
| `timestamp_sec` | float | Time from session start (seconds) |
| `phase` | string | `graffiti_removal` or `repaint_application` |
| `action_label` | string | Primary action (see taxonomy above) |
| `sub_action` | string | Fine-grained sub-action description |
| `worker_id` | string | Worker identifier (W01–W04) |
| `body_posture` | string | Posture category |
| `working_height` | string | Vertical zone on fence being worked |

### `imu_*.csv`

| Field | Type | Unit | Description |
|---|---|---|---|
| `timestamp_sec` | float | s | Synchronized timestamp |
| `accel_x/y/z_g` | float | g | Linear acceleration |
| `gyro_x/y/z_dps` | float | °/s | Angular velocity |
| `euler_roll/pitch/yaw_deg` | float | ° | Fused orientation estimate |

### `force_glove_right.csv`

| Field | Type | Unit | Description |
|---|---|---|---|
| `timestamp_sec` | float | s | Synchronized timestamp |
| `cell_{name}_N` | float | N | Force at each of 24 sensor cells |
| `total_force_N` | float | N | Sum across all active cells |
| `grip_type` | string | — | Inferred grip classification |

### `realsense_depth.bag` (ROS bag)

| Topic | Format | Description |
|---|---|---|
| `/camera/depth/image_rect_raw` | sensor_msgs/Image | 16-bit depth (mm) |
| `/camera/color/image_raw` | sensor_msgs/Image | RGB aligned to depth |
| `/camera/depth/color/points` | sensor_msgs/PointCloud2 | Fused RGBD point cloud |

---

## Data Splits

| Split | Sessions | Hours | Notes |
|---|---|---|---|
| Sample (free) | 1 | 0.5 | Available without license agreement |
| Full v1.0 | 10 | 50 | Commercial license required |
| Planned v2.0 | 50+ | 300+ | Additional tasks: garbage collection, weeding, building cleaning |

---

## Dataset Creation

### Motivation

Physical AI and embodied robot models are rapidly advancing, but publicly available training data for **outdoor real-world manipulation tasks** remains extremely scarce. Existing motion datasets focus primarily on laboratory manipulation, household tasks, or driving — leaving a large gap in construction, maintenance, and urban service work.

This dataset addresses that gap by collecting data from professional Japanese workers performing skilled outdoor maintenance tasks. Japan's workforce is recognized globally for precision, safety discipline, and consistent execution — making it an ideal source of high-quality demonstration data.

### Collection Methodology

- Task protocols defined in cooperation with professional site supervisors
- Sessions conducted at real operational job sites
- Multiple skill levels included to capture expert vs. novice motion differences
- Both "successful" and "recovery from failure" motion captured and labeled
- All data cleaned, synchronized, and formatted to be pipeline-ready

### Known Limitations

- All data collected in Tokyo metropolitan area — urban environment bias
- Nighttime and rainy-weather sessions represent < 15% of current dataset
- Force glove covers right hand only
- Skeleton estimation accuracy degrades at extreme joint angles (> 110°)

---

## Licensing Information

This dataset is released under **Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0)**.

- ✅ Free to use for academic and non-commercial research
- ✅ Free 30-minute sample available without registration
- ✅ Redistribution permitted with attribution
- ❌ Commercial use requires a separate commercial license

For commercial licensing inquiries, please contact: **contact@fielddata.jp**

---

## Citation

If you use this dataset in your research, please cite:

```bibtex
@dataset{fielddata2025folmd,
  author    = {FieldData Japan},
  title     = {FieldData Outdoor Labor Motion Dataset (FOLMD) v1.0},
  year      = {2025},
  publisher = {HuggingFace},
  url       = {https://huggingface.co/datasets/fielddata-jp/folmd},
  note      = {Multi-modal motion dataset for physical AI training — outdoor maintenance tasks, Tokyo, Japan}
}
```

---

## Contact

**FieldData Japan**
- Email: contact@fielddata.jp
- HuggingFace: [fielddata-jp](https://huggingface.co/fielddata-jp)
- X (Twitter): [@fielddata_jp](https://twitter.com/fielddata_jp)

We welcome collaboration inquiries from robotics researchers and companies interested in:
- Custom data collection for specific task domains
- Expanding the dataset to new outdoor task categories
- Joint research and co-authorship opportunities