File size: 4,261 Bytes
d352621
 
 
6e803ac
 
d352621
6e803ac
 
 
 
 
 
 
 
 
d352621
 
6e803ac
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d352621
 
6e803ac
 
 
 
 
 
 
 
 
 
 
 
d352621
 
 
6e803ac
d352621
 
6e803ac
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
---
license: apache-2.0
task_categories:
  - robotics
  - reinforcement-learning
tags:
  - robotics
  - imitation-learning
  - diffusion-policy
  - manipulation
  - fetch
  - mujoco
  - lerobot
size_categories:
  - 1K<n<10K
---

# DiffPick: Fetch Pick-and-Place Demonstrations

A clean dataset of **200 successful pick-and-place demonstrations** collected from a scripted expert policy in the [`FetchPickAndPlace-v4`](https://robotics.farama.org/envs/fetch/pick_and_place/) MuJoCo environment. Designed for training **vision-based imitation learning** policies (Diffusion Policy, ACT, BC).

Part of the [DiffPick project](https://github.com/e-cagan/diffpick) — a from-scratch implementation of a Diffusion Policy pipeline with ROS2 deployment.

## Dataset Stats

| Property | Value |
|---|---|
| Episodes | 200 |
| Total frames | 5,489 |
| Mean episode length | 27.4 steps |
| Min / Max length | 20 / 35 steps |
| FPS | 25 |
| Image resolution | 96×96 RGB |
| Success rate (during collection) | 97.1% (200 of 206 attempts kept) |

## Features

| Key | Shape | Type | Description |
|---|---|---|---|
| `observation.image` | (3, 96, 96) | float32 | Front-view RGB camera |
| `observation.state` | (10,) | float32 | Robot proprioception only (gripper xyz, finger widths, velocities). **No object pose** — must be inferred from image. |
| `action` | (4,) | float32 | End-effector delta (dx, dy, dz) ∈ [-1,1] + gripper command (-1 close, +1 open) |
| `task` | string | — | "Pick up the block and place it at the target location." |

### Why proprioception-only state?

The state vector deliberately **excludes object position**. This forces a learned policy to develop visual grounding rather than copying ground-truth coordinates. The result: policies trained on this dataset must actually *see* the object in the RGB stream to succeed — closer to a real-world deployment scenario where object pose isn't directly observable.

## Expert Policy

Demonstrations were generated by a hand-crafted state machine controller:

```
APPROACH (gripper open, hover above object)

DESCEND (gripper open, lower to object)

GRASP (close gripper, hold for 8 steps)

PLACE (move to target, gripper closed)
```

Proportional control in end-effector space (no IK required, since the env exposes a 4-D end-effector action interface). Episodes ending in success too quickly (< 15 steps, indicating object near target at reset) were filtered out.

Source: [`data_collection/scripted_policy.py`](https://github.com/e-cagan/diffpick/blob/main/data_collection/scripted_policy.py)

## Usage

```python
from lerobot.datasets.lerobot_dataset import LeRobotDataset

dataset = LeRobotDataset("e-cagan/diffpick")
sample = dataset[0]

print(sample["observation.image"].shape)  # torch.Size([3, 96, 96])
print(sample["observation.state"].shape)  # torch.Size([10])
print(sample["action"].shape)             # torch.Size([4])
```

## Reproduction

```bash
git clone https://github.com/e-cagan/diffpick
cd diffpick
pip install -r requirements.txt

# Collect raw demos
python -m data_collection.collect --n_episodes 200

# Convert to LeRobotDataset format
python -m data_collection.to_lerobot_dataset \
    --raw_dir data/raw_demos \
    --repo_id <your-username>/diffpick \
    --fps 25
```

## Intended Use

- Training **Diffusion Policy** for vision-conditioned manipulation
- Benchmarking imitation learning algorithms (BC vs ACT vs DP)
- Learning resource for ROS2 + MuJoCo + LeRobot integration

## Limitations

- Single environment seed family (`FetchPickAndPlace-v4` defaults). No domain randomization for backgrounds, lighting, or distractors.
- Single front-facing 96×96 camera. No wrist cam, no depth.
- Scripted expert is deterministic given a seed — no behavioral diversity (no left-hand/right-hand approach modes, etc.). This may limit the multi-modal advantages of Diffusion Policy.
- Object is a single blue cube. No category generalization.

## Citation

If you use this dataset, please cite:

```bibtex
@misc{apaydin2026diffpick,
  author = {Apaydın, Emin Çağan},
  title = {DiffPick: A Diffusion Policy Pipeline for Fetch Pick-and-Place},
  year = {2026},
  publisher = {GitHub},
  url = {https://github.com/e-cagan/diffpick}
}
```

## License

Apache 2.0