curdie / README.md
astridwilde1's picture
Update README.md
9385c2e verified
---
license: apache-2.0
task_categories:
- image-to-3d
- object-detection
- mask-generation
- image-segmentation
- depth-estimation
- keypoint-detection
tags:
- pose-estimation
- synthetic-data
- computer-vision
- unreal-engine
- coco-keypoints
- instance-segmentation
- depth-maps
size_categories:
- 1M<n<10M
pretty_name: CURDIE3
---
# Dataset Card for CURDIE3
## Dataset Details
### Dataset Description
**CURDIE3** (Crafted Unreal Renders of Detailed Individuals in 3 Environments) is a synthetic pose estimation dataset featuring photorealistic RGB images, metric depth maps, instance segmentation masks, and multi-format keypoint annotations rendered in Unreal Engine.
The dataset contains **1,408,410 images** with ground truth labels rendered at **30 fps**, featuring 100 distinct actor models across 1,400 distinct animations in three environment maps: `field`, `house`, and `urban`. Each sample provides comprehensive annotations including COCO-17 body keypoints, MediaPipe hand keypoints, foot keypoints, camera intrinsics/extrinsics, and per-actor instance segmentation masks in COCO RLE format.
![Example1](https://cdn-uploads.huggingface.co/production/uploads/68bb195c6e97e5a7f855f80f/RPQWKxzBI7IWNQRLdIfzq.jpeg)
![Example2](https://cdn-uploads.huggingface.co/production/uploads/68bb195c6e97e5a7f855f80f/DBOzvVg29SwmmUspZTA5f.jpeg)
![Example3](https://cdn-uploads.huggingface.co/production/uploads/68bb195c6e97e5a7f855f80f/NSpXDX4AFy360CLkF0CnI.jpeg)
- **Curated by:** Brendan Wilson, Astrid Wilde, Chris Johnson, Michael Listo, Audrey Ito, Anna Shive, Sherry Wallace
- **License:** Apache 2.0
### Dataset Sources
- **Paper:** Wilson et al., "CURDIE3: Crafted Unreal Renders of Detailed Individuals in 3 Environments" (2026)
## Uses
### Direct Use
CURDIE3 is designed for training and evaluating models on the following tasks:
- 2D/3D human pose estimation
- Instance segmentation
- Multi-person detection and tracking
- Hand pose estimation
- Keypoint detection and localization
### Out-of-Scope Use
Due to known limitations, CURDIE3 is **not suitable** for:
- **Depth estimation from RGB:** Ground-truth depth is ambiguous at mesh intersection points
- **Occlusion reasoning and amodal completion:** Intersection boundaries are physically impossible
- **Physics-based pose estimation or motion prediction:** Some configurations are non-physical
- **Contact and interaction detection:** Difficult to distinguish intentional contact from erroneous intersection
- **Sim-to-real transfer:** Where physical plausibility is required
- **Facial expression recognition or emotion detection:** All actors maintain neutral expressions
- **Lip-sync or speech-related tasks:** No facial animations present
- **Social interaction modeling:** No coordinated interactions between subjects
- **Group activity recognition:** Actors are animated independently
- **Ground contact estimation or footskate detection:** No foot-ground contact labels; feet may penetrate floors or float
## Dataset Structure
| Attribute | Value |
|-----------|-------|
| Total Images | 1,408,410 |
| Frame Rate | 30 fps |
| Actor Models | 100 |
| Animations | 1,400 |
| Environments | 3 (`field`, `house`, `urban`) |
### Camera Movement Types
The dataset contains four types of camera movements:
| Movement Type | Description | Tag |
|---------------|-------------|-----|
| **Stationary** | Fixed camera position and orientation | `free` or `orbit` |
| **Dolly** | Linear translation in a constant direction for the duration of the animation | `free` |
| **Arc** | Linear translation combined with constant rotation for the duration of the animation | `free` |
| **Orbit** | Camera pivots around a specified center point in 3D space | `orbit` |
**Tag Reference:**
- All **dolly** and **arc** movements are tagged as `free`
- All **orbit** movements are tagged as `orbit`
- **Stationary** cameras may be tagged as either `free` or `orbit`
### File Format
The dataset is distributed as split 7z archives with 40GB volumes. Each sample consists of paired files with a globally unique 12-digit identifier:
- `000000000042.png` — RGB image
- `000000000042.npz` — Compressed NumPy archive containing all annotations
**Extraction:**
```bash
7z x curdie3.7z.001 # Automatically processes all volumes
```
**Loading NPZ files:**
```python
import numpy as np
data = np.load(path, allow_pickle=True) # allow_pickle=True required for RLE masks
```
### Coordinate Systems
| Space | Convention |
|-------|------------|
| World 3D | UE4: X=Forward, Y=Right, Z=Up (left-handed) |
| Screen 2D | Origin at top-left, X=right, Y=down (pixels) |
| Rotation | Degrees: Pitch (X), Yaw (Z), Roll (Y) |
### NPZ Field Reference
#### Frame Metadata
- `frame_id` (str): Original frame identifier
- `dataset` (str): Source dataset name
- `image_shape` (uint16, shape 3): [height, width, channels]
#### Depth
- `depth_m` (float16, shape H×W): Per-pixel depth in meters (range ~0–30m)
#### Camera Intrinsics & Extrinsics
- `K` (float32, shape 3×3): Camera intrinsic matrix
- `camera_location` (float16, shape 3): World-space position in meters
- `camera_rotation` (float16, shape 3): Orientation [Pitch, Yaw, Roll] in degrees
- `camera_direction` (float16, shape 3): Unit forward vector
#### Sensor Settings
- `sensor_width_mm`, `sensor_height_mm` (float16): Physical sensor dimensions
- `focal_length_mm` (float16): Lens focal length
- `hfov_deg` (float16): Horizontal field of view
- `exposure_compensation` (float16): Exposure compensation stops
#### Scene Metadata
- `selected_map` (str): Environment name (`field`, `house`, or `urban`)
- `time_of_day` (str): In-scene time as "HH:MM:SS"
- `capture_mode` (str): Camera movement tag (`free` or `orbit`)
#### Instance Segmentation
- `actor_mask_rles` (object, shape N): Array of COCO RLE bytes, one per actor
- `actor_names` (object, shape N): Actor identifiers
- `num_actors` (int): Number of actors in frame
#### Per-Actor Annotations
- `actor_world_locations` (float16, shape N×3): World position in meters
- `actor_world_rotations` (float16, shape N×3): Orientation in degrees
- `actor_bbox_xywh` (int32, shape N×4): COCO-format bounding boxes
#### Keypoints
- `keypoints_body_3d` (float16, shape N×17×4): COCO-17 body keypoints, world space
- `keypoints_body_2d` (float16, shape N×17×3): COCO-17 body keypoints, image space
- `keypoints_foot_3d` / `keypoints_foot_2d`: 6 foot keypoints per actor
- `keypoints_lefthand_3d` / `keypoints_lefthand_2d`: 21 MediaPipe hand keypoints
- `keypoints_righthand_3d` / `keypoints_righthand_2d`: 21 MediaPipe hand keypoints
### Keypoint Definitions
**COCO-17 Body Keypoints:**
```
0: nose, 1: left_eye, 2: right_eye, 3: left_ear, 4: right_ear,
5: left_shoulder, 6: right_shoulder, 7: left_elbow, 8: right_elbow,
9: left_wrist, 10: right_wrist, 11: left_hip, 12: right_hip,
13: left_knee, 14: right_knee, 15: left_ankle, 16: right_ankle
```
**MediaPipe Hand (21 points):**
```
0: wrist, 1-4: thumb, 5-8: index, 9-12: middle, 13-16: ring, 17-20: pinky
```
## Dataset Creation
### Curation Rationale
CURDIE3 was created to provide a large-scale synthetic dataset for human pose estimation research, offering pixel-perfect ground truth annotations that are difficult or impossible to obtain from real-world data collection.
### Source Data
#### Data Collection and Processing
All data was rendered in Unreal Engine using 100 distinct actor models and 1,400 animations across three environment maps. Camera positions include stationary, dolly, arc, and orbit capture modes with varying focal lengths and exposure settings.
#### Who are the source data producers?
The dataset was synthetically generated; no real human subjects were involved.
### Annotations
#### Annotation process
All annotations are automatically generated during the rendering process, providing pixel-perfect ground truth without manual labeling.
#### Personal and Sensitive Information
This dataset contains only synthetic data with no real human subjects. No personal or sensitive information is present.
## Bias, Risks, and Limitations
### Known Limitations
1. **Mesh Intersections:** Actor meshes may intersect with each other and environment geometry throughout the dataset.
2. **Neutral Facial Expressions:** All actors maintain neutral resting faces—no facial expressions are present.
3. **No Social Interactions:** Actors are animated independently with no coordinated interactions (similar to BEDLAM).
4. **No Ground Contact Labels:** Foot-ground contact is not annotated; feet may penetrate floors or float above ground.
5. **Limited Camera Movement Variety:** Only four camera movement types are present (stationary, dolly, arc, orbit), with all movements maintaining constant velocity and rotation rate throughout each animation.
### Recommendations
Users should carefully consider the known limitations when selecting CURDIE3 for their research. The dataset is best suited for pose estimation tasks that do not require physical plausibility, facial expressions, or social interaction understanding.
## Citation
**BibTeX:**
```bibtex
@inproceedings{wilson2026curdie3,
title={{CURDIE}3: Crafted Unreal Renders of Detailed Individuals in 3 Environments},
author={Brendan Wilson and Astrid Wilde and Chris Johnson and Michael Listo and Audrey Ito and Anna Shive and Sherry Wallace},
year={2026}
}
```
## Technical Requirements
### Dependencies
```
numpy
Pillow
pycocotools
```
### Example Usage
```python
import numpy as np
from pycocotools import mask as mask_utils
from PIL import Image
data = np.load("000000000042.npz", allow_pickle=True)
rgb = np.array(Image.open("000000000042.png"))
# Access depth
depth_m = data['depth_m'].astype(np.float32)
# Access camera intrinsics
K = data['K']
fx, fy = K[0, 0], K[1, 1]
cx, cy = K[0, 2], K[1, 2]
# Check camera movement type
capture_mode = str(data['capture_mode']) # 'free' or 'orbit'
# Iterate over actors
h, w = data['image_shape'][:2]
for i in range(data['num_actors']):
# Decode instance mask
rle = {'size': [int(h), int(w)], 'counts': data['actor_mask_rles'][i]}
mask = mask_utils.decode(rle)
# Get 2D body keypoints
kpts_2d = data['keypoints_body_2d'][i] # (17, 3)
```
## Dataset Card Authors
Brendan Wilson, Astrid Wilde, Chris Johnson, Michael Listo, Audrey Ito, Anna Shive, Sherry Wallace
### Contact Us
If you're interested in similar datasets with greater diversity (more actors, more environments, realistic camera motions, Human On Human Interaction, Human Object Interactions) and richer labels, or substituting humans with robots,
[Contact Us](mailto:astrid@kikitora.com)