File size: 4,378 Bytes
5a059c0
 
 
 
 
 
 
 
 
 
 
59fd801
8cc7e10
5a059c0
 
 
 
 
 
 
 
 
694df88
5a059c0
 
694df88
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5a059c0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
414b449
5a059c0
 
414b449
5a059c0
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
---
license: cc-by-4.0
language:
- en
tags:
- medical
---
# SutureBot: A Precision Framework and Benchmark for Autonomous End-to-End Suturing

This dataset is part of the **SutureBot** project, a benchmark for developing and evaluating autonomous surgical robotic policies for end-to-end suturing. It contains high-resolution, multi-camera video and robot kinematics from real robot demonstrations on tissue phantoms, supporting research in imitation learning, vision-language-action modeling, and surgical autonomy.

[project website](https://suturebot.github.io/) | [code repository](https://github.com/SutureBot/SutureBot/tree/ACT).

## Dataset Summary

The dataset includes:

- **Stereo endoscope images** (`left_img_dir`, `right_img_dir`)
- **Wrist camera views** (`endo_psm1`, `endo_psm2`)
- **Low-level kinematics** (end-effector pose, joint angles, jaw state, etc.) in `ee_csv.csv`
- **Multiple tissues** and **multiple task types** (e.g., needle pick-up, suturing)
- All data is organized by:
tissue_[id]/[task_name]/[episode_timestamp]/


The following is a visualization of the dataset folder structure:
 ``` 
 $PATH_TO_DATASET
β”œβ”€β”€ [DATASET_NAME]       # the dataset base dir
|   └── tissue_1                      # data subset
|   |   β”œβ”€β”€ 1_[task_name]             # task name
|   |   |   β”œβ”€β”€ [episode]             # should be timestamp when the data was recorded
|   |   |   |      β”œβ”€β”€ left_img_dir   # left endoscope cam images (frame000000_left.jpg)
|   |   |   |      β”œβ”€β”€ right_img_dir  # right endoscope cam images (frame000000_right.jpg)
|   |   |   |      β”œβ”€β”€ endo_psm1      # right wrist cam images (frame000000_psm1.jpg)
|   |   |   |      β”œβ”€β”€ endo_psm2      # left wrist cam images (frame000000_psm2.jpg)
|   |   |   |      └── ee_csv.csv     # kinematics
|   └── tissue_2                      # data subset
|   |   β”œβ”€β”€ 1_[task_name]             # task name
|   |   |   β”œβ”€β”€ [episode]             # should be timestamp when the data was recorded
|   |   |   |      β”œβ”€β”€ left_img_dir   # left endoscope cam images (frame000000_left.jpg)
|   |   |   |      β”œβ”€β”€ right_img_dir  # right endoscope cam images (frame000000_right.jpg)
|   |   |   |      β”œβ”€β”€ endo_psm1      # right wrist cam images (frame000000_psm1.jpg)
|   |   |   |      β”œβ”€β”€ endo_psm2      # left wrist cam images (frame000000_psm2.jpg)
|   |   |   |      └── ee_csv.csv     # kinematics
...
 ``` 

Each episode is a complete trajectory from a real robotic execution using a da Vinci Research Kit (dVRK) platform.

---

## Usage

This dataset is designed for training and evaluating autonomous policies that take as input multi-view RGB images (optionally language-conditioned), and output continuous control actions.

It supports a wide range of applications including:
- Behavioral cloning and offline RL
- Multi-modal policy learning (images + kinematics + language)
- Visual servoing and tool tracking
- Surgical robot benchmarking

---

## File Structure

Each episode contains:
- `left_img_dir/`: Left endoscope camera images
- `right_img_dir/`: Right endoscope camera images
- `endo_psm1/`: Right wrist (PSM1) camera images
- `endo_psm2/`: Left wrist (PSM2) camera images
- `ee_csv.csv`: Kinematic data including joint states, poses, and jaw angles

---

## Format

- Image format: `.jpg`, RGB
- Kinematics: CSV with columns such as:
- `timestamp`
- `psm1_pose.position.{x,y,z}`
- `psm1_pose.orientation.{x,y,z,w}`
- `psm1_js[0-5]`, `psm2_js[0-5]`, etc.
- `ecm_pose.*`, `suj_pose.*`, etc.

A complete list of columns is available in the `dataset.croissant.json` schema.

---

## Citation

If you use this dataset, please cite:

```bibtex
@misc{suturebot2025,
title       = {SutureBot: A Precision Framework and Benchmark for Autonomous End-to-End Suturing},
author      = {Jesse Haworth, Juo-Tung Chen, Nigel Nelson, Ji Woong Kim, Masoud Moghani, Chelsea Finn, Axel Krieger},
year        = {2025},
note        = {Under review at NeurIPS 2025 Datasets and Benchmarks Track},
howpublished = {\url{https://huggingface.co/datasets/jchen396/suturebot}}
}
```
## License
This dataset is licensed under CC-BY-4.0. Please credit the authors when using this dataset.

## Contact
For questions or collaboration inquiries, please contact:

Juo-Tung Chen
Johns Hopkins University
πŸ“§ jchen396@jh.edu