File size: 13,077 Bytes
be4f098 aff8a55 be4f098 aff8a55 be4f098 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 | ---
annotations_creators:
- expert-generated
language_creators:
- found
language:
- ja
license: cc-by-nc-4.0
multilinguality:
- monolingual
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- other
task_ids:
- pose-estimation
- grasping
- task-planning
paperswithcode_id: null
pretty_name: FieldData Outdoor Labor Motion Dataset
tags:
- physical-ai
- embodied-ai
- motion-capture
- manipulation
- outdoor-robotics
- painting
- graffiti-removal
- construction
- human-robot-learning
---
# Dataset Card: FieldData Outdoor Labor Motion Dataset (FOLMD)
## Table of Contents
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks](#supported-tasks)
- [Data Collection](#data-collection)
- [Dataset Structure](#dataset-structure)
- [Annotations](#annotations)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Licensing Information](#licensing-information)
- [Citation](#citation)
- [Contact](#contact)
---
## Dataset Description
- **Homepage:** https://fielddata.jp
- **Repository:** https://huggingface.co/datasets/fielddata-jp/folmd
- **Paper:** [Coming soon]
- **Point of Contact:** contact@fielddata.jp
### Dataset Summary
**FieldData Outdoor Labor Motion Dataset (FOLMD)** is a multi-modal motion dataset of professional outdoor workers performing real-world manual labor tasks in urban environments in Japan.
This dataset is designed to support the development of **physical AI foundation models** and **embodied robotic systems** targeting outdoor maintenance and construction tasks. All data was collected from professional workers at actual job sites — not from staged laboratory environments.
The v1.0 release covers a complete two-phase urban fence restoration task:
1. **Graffiti removal** using chemical solvent and rag wiping
2. **Repainting** with black paint using brushes at multiple heights and postures
Data modalities include egocentric RGB video, depth maps (Intel RealSense D435i), full-body IMU (8-point, Mbientlab MetaMotionS), glove-based grasp force (Tekscan Grip System), and annotated 3D skeleton sequences (SMPL format).
---
## Why This Dataset?
| Challenge for Robots | What This Dataset Provides |
|---|---|
| Vertical surface contact control | Sustained brush pressure on metal slat fence (9–25 N) |
| Low-posture precision work | Deep crouch / kneeling brushwork at ground level |
| Boundary-aware manipulation | Painting within 10 mm of masking tape edges |
| Multi-posture task transitions | Standing → crouching → kneeling within a single session |
| Visual quality assessment | Workers pausing to inspect coverage, detecting missed spots |
| Multi-agent coordination | 3–4 workers dividing fence height and working in parallel |
These tasks involve sustained contact with irregular vertical surfaces, fine motor control under physical load, and situated decision-making — all critical challenges for next-generation manipulation models.
---
## Supported Tasks
- **Imitation Learning (IL):** Action-annotated demonstrations suitable for behavior cloning and inverse reinforcement learning
- **Vision-Language-Action (VLA):** Paired egocentric video and action labels for multimodal training
- **Pose Estimation:** Full-body SMPL skeleton data across 6 posture categories
- **Action Segmentation:** 7-class taxonomy with fine-grained sub-actions and temporal boundaries
- **Force Control Learning:** Grasp pressure profiles for 5 grip types across different tools and surfaces
- **Contact-Rich Manipulation:** Depth + force data for learning compliant surface following
---
## Data Collection
### Sensor Setup
| Modality | Device | Specs | Sample Rate |
|---|---|---|---|
| Egocentric RGB | GoPro Hero 12 | 1920×1080 | 60 fps |
| Depth + RGB | Intel RealSense D435i | 848×480 | 30 fps |
| Full-body IMU (×8) | Mbientlab MetaMotionS | Accel + Gyro + Euler | 200 Hz |
| Grasp Force Glove | Tekscan Grip System | 24 cells, right hand | 100 Hz |
| 3D Skeleton | Computed from IMU | SMPL format | 30 fps |
| External cameras | GoPro Hero 12 (×3) | 1920×1080 | 60 fps |
**Temporal synchronization:** All sensors synchronized via GPS-PPS hardware sync signal. Timestamp accuracy: ±0.8 ms.
### Sensor Placement
```
Head: GoPro (front) + RealSense (side)
R Wrist: IMU
L Wrist: IMU
R Elbow: IMU
L Elbow: IMU
Waist: IMU
R Knee: IMU
L Knee: IMU
R Hand: Force glove (Tekscan)
```
### Collection Protocol
- Workers are professional outdoor maintenance staff with 2–10 years of experience
- Written informed consent obtained from all participants
- Data collected at real urban job sites (Tokyo metropolitan area, Japan)
- Session begins with 5-minute sensor calibration and zeroing
- Workers perform tasks naturally without scripted movements
- All sessions include workers of mixed skill levels (novice / intermediate / experienced)
---
## Dataset Structure
```
folmd-v1.0/
├── sessions/
│ └── session_20210327_001/
│ ├── raw/
│ │ ├── gopro_head.mp4 # Egocentric RGB (60 fps)
│ │ ├── realsense_rgb.mp4 # RealSense RGB (30 fps)
│ │ ├── realsense_depth.bag # Depth stream (ROS bag)
│ │ ├── imu_r_wrist.csv # Right wrist IMU
│ │ ├── imu_l_wrist.csv # Left wrist IMU
│ │ ├── imu_r_elbow.csv # Right elbow IMU
│ │ ├── imu_l_elbow.csv # Left elbow IMU
│ │ ├── imu_waist.csv # Waist IMU
│ │ ├── imu_r_knee.csv # Right knee IMU
│ │ ├── imu_l_knee.csv # Left knee IMU
│ │ ├── imu_head.csv # Head IMU
│ │ └── force_glove_right.csv # Right hand force (24 cells)
│ ├── processed/
│ │ ├── skeleton_smpl/
│ │ │ └── frame_*.json # Per-keyframe SMPL skeleton
│ │ ├── depth_pointcloud/
│ │ │ └── frame_*.pcd # Per-keyframe 3D point cloud
│ │ └── action_segments.json # Full annotation (Tier 2)
│ └── metadata/
│ ├── session_info.json
│ ├── sensor_calibration.json
│ └── quality_report.pdf
├── dataset_card.md # This file
├── annotation_guidelines.pdf # Labeling rulebook (EN)
└── sample/
└── 30min_annotated_sample/ # Free sample — see Licensing
```
---
## Annotations
### Action Taxonomy (7 Primary Labels)
| Action Label | Description | Avg Duration (sec) |
|---|---|---|
| `solvent_application` | Applying chemical solvent to dissolve graffiti using rag | 45–180 |
| `scrub_with_rag` | Physical scrubbing to remove dissolved paint | 10–60 |
| `paint_application` | Applying paint with brush/roller at mid-to-upper height | 30–300 |
| `detail_paint_application` | Precision brushwork at edges, slat gaps, base boundaries | 10–120 |
| `visual_inspection` | Standing back to scan surface for quality assessment | 5–30 |
| `material_preparation` | Handling tools, mixing paint, preparing supplies | 5–60 |
| `masking` | Applying or adjusting masking tape | 30–180 |
### Posture Labels (6 Categories)
```
standing_upright — full height, minimal trunk lean
standing_slight_lean — 10–20° trunk forward flexion
crouching_moderate — knee flexion 60–90°
crouching_deep — knee flexion > 90°
kneeling_single_knee — one knee on ground
kneeling_double_knee — both knees on ground
```
### Force Labels (Qualitative)
```
minimal < 5 N (passive hold / inspection)
low 5–10 N (precision brush / fine detail)
medium 10–20 N (normal painting stroke)
high 20–30 N (scrubbing / removal force)
very_high > 30 N (heavy scrub on resistant surface)
```
### Annotation Layers
| Layer | Content | Completeness |
|---|---|---|
| Layer 1 | Action segmentation (start/end time + label) | 100% |
| Layer 2 | Object + surface + environment labels | 100% |
| Layer 3 | Skill quality label (expert/intermediate/novice) | 100% |
| Layer 4 | Failure cases and recovery actions | 100% |
**Inter-annotator agreement (Cohen's κ):** 0.94
---
## Data Fields
### `action_segments.json`
| Field | Type | Description |
|---|---|---|
| `session_id` | string | Unique session identifier |
| `frame_id` | string | Keyframe identifier |
| `timestamp_sec` | float | Time from session start (seconds) |
| `phase` | string | `graffiti_removal` or `repaint_application` |
| `action_label` | string | Primary action (see taxonomy above) |
| `sub_action` | string | Fine-grained sub-action description |
| `worker_id` | string | Worker identifier (W01–W04) |
| `body_posture` | string | Posture category |
| `working_height` | string | Vertical zone on fence being worked |
### `imu_*.csv`
| Field | Type | Unit | Description |
|---|---|---|---|
| `timestamp_sec` | float | s | Synchronized timestamp |
| `accel_x/y/z_g` | float | g | Linear acceleration |
| `gyro_x/y/z_dps` | float | °/s | Angular velocity |
| `euler_roll/pitch/yaw_deg` | float | ° | Fused orientation estimate |
### `force_glove_right.csv`
| Field | Type | Unit | Description |
|---|---|---|---|
| `timestamp_sec` | float | s | Synchronized timestamp |
| `cell_{name}_N` | float | N | Force at each of 24 sensor cells |
| `total_force_N` | float | N | Sum across all active cells |
| `grip_type` | string | — | Inferred grip classification |
### `realsense_depth.bag` (ROS bag)
| Topic | Format | Description |
|---|---|---|
| `/camera/depth/image_rect_raw` | sensor_msgs/Image | 16-bit depth (mm) |
| `/camera/color/image_raw` | sensor_msgs/Image | RGB aligned to depth |
| `/camera/depth/color/points` | sensor_msgs/PointCloud2 | Fused RGBD point cloud |
---
## Data Splits
| Split | Sessions | Hours | Notes |
|---|---|---|---|
| Sample (free) | 1 | 0.5 | Available without license agreement |
| Full v1.0 | 10 | 50 | Commercial license required |
| Planned v2.0 | 50+ | 300+ | Additional tasks: garbage collection, weeding, building cleaning |
---
## Dataset Creation
### Motivation
Physical AI and embodied robot models are rapidly advancing, but publicly available training data for **outdoor real-world manipulation tasks** remains extremely scarce. Existing motion datasets focus primarily on laboratory manipulation, household tasks, or driving — leaving a large gap in construction, maintenance, and urban service work.
This dataset addresses that gap by collecting data from professional Japanese workers performing skilled outdoor maintenance tasks. Japan's workforce is recognized globally for precision, safety discipline, and consistent execution — making it an ideal source of high-quality demonstration data.
### Collection Methodology
- Task protocols defined in cooperation with professional site supervisors
- Sessions conducted at real operational job sites
- Multiple skill levels included to capture expert vs. novice motion differences
- Both "successful" and "recovery from failure" motion captured and labeled
- All data cleaned, synchronized, and formatted to be pipeline-ready
### Known Limitations
- All data collected in Tokyo metropolitan area — urban environment bias
- Nighttime and rainy-weather sessions represent < 15% of current dataset
- Force glove covers right hand only
- Skeleton estimation accuracy degrades at extreme joint angles (> 110°)
---
## Licensing Information
This dataset is released under **Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0)**.
- ✅ Free to use for academic and non-commercial research
- ✅ Free 30-minute sample available without registration
- ✅ Redistribution permitted with attribution
- ❌ Commercial use requires a separate commercial license
For commercial licensing inquiries, please contact: **contact@fielddata.jp**
---
## Citation
If you use this dataset in your research, please cite:
```bibtex
@dataset{fielddata2025folmd,
author = {FieldData Japan},
title = {FieldData Outdoor Labor Motion Dataset (FOLMD) v1.0},
year = {2025},
publisher = {HuggingFace},
url = {https://huggingface.co/datasets/fielddata-jp/folmd},
note = {Multi-modal motion dataset for physical AI training — outdoor maintenance tasks, Tokyo, Japan}
}
```
---
## Contact
**FieldData Japan**
- Email: contact@fielddata.jp
- HuggingFace: [fielddata-jp](https://huggingface.co/fielddata-jp)
- X (Twitter): [@fielddata_jp](https://twitter.com/fielddata_jp)
We welcome collaboration inquiries from robotics researchers and companies interested in:
- Custom data collection for specific task domains
- Expanding the dataset to new outdoor task categories
- Joint research and co-authorship opportunities
|