File size: 6,304 Bytes
7b863d2 60290ff 7b863d2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 |
---
license: apache-2.0
tags:
- egocentric
- exotenric
- surgery
- or
- scene-graph
- activity-understanding
- gaze
- hand
---
# EgoExOR-HQ: An Ego-Exo-Centric Operating Room Dataset for Surgical Activity Understanding
[](https://huggingface.co/datasets/TUM/EgoExOR)
[](https://github.com/ardamamur/EgoExOR)
[](https://neurips.cc/)
**EgoExOR-HQ** — This repository hosts the **enriched high-quality release** of the EgoExOR dataset. For scene graph generation code, benchmarks, and pretrained models, see the [main EgoExOR repository](https://github.com/ardamamur/EgoExOR).
**Authors:** Ege Özsoy, Arda Mamur, Felix Tristram, Chantal Pellegrini, Magdalena Wysocki, Benjamin Busam, Nassir Navab
## ✨ What's New in EgoExOR-HQ
This release adds:
- **High-quality images** — 1344×1344 resolution (instead of 336×336)
- **Raw depth images** — From external RGB-D cameras (instead of pre-merged point clouds), so you can build merged or per-camera point clouds for your use case
- **Per-device audios** — Separate audio streams per microphone
## Overview
Operating rooms (ORs) demand precise coordination among surgeons, nurses, and equipment in a fast-paced, occlusion-heavy environment, necessitating advanced perception models to enhance safety and efficiency. Existing datasets either provide partial egocentric views or sparse exocentric multi-view context, but do not explore the comprehensive combination of both.
We introduce **EgoExOR**, the first OR dataset and accompanying benchmark to fuse first-person and third-person perspectives. Spanning 94 minutes (84,553 frames at 15 FPS) of two emulated spine procedures—*Ultrasound-Guided Needle Insertion* and *Minimally Invasive Spine Surgery*—EgoExOR integrates:
- **Egocentric:** RGB, gaze, hand tracking, audio from wearable glasses
- **Exocentric:** RGB and depth from RGB-D cameras, ultrasound imagery
- **Annotations:** 36 entities, 22 relations (568,235 triplets) for scene graph generation
This dataset sets a new foundation for OR perception, offering a rich, multimodal resource for next-generation clinical perception.
## 🌟 Key Features
- **Multiple modalities** — RGB video, audio (full waveform + per-frame snippets, per-device), eye gaze, hand tracking, raw depth, and scene graph annotations
- **Time-synchronized streams** — All modalities aligned on a common timeline for precise cross-modal correlation
- **High-resolution RGB** — 1344×1344 frames for fine-grained visual analysis
- **Raw depth** — Build custom point clouds or depth-based models; depth from external RGB-D cameras only
- **Per-device audio** — Separate microphone streams for spatial or multi-channel audio processing
## 📂 Dataset Structure
The dataset is distributed as **phase-level HDF5 files** for efficient download:
| File | Description |
|------|-------------|
| `miss_1.h5` | MISS procedure, phase 1 |
| `miss_2.h5` | MISS procedure, phase 2 |
| `miss_3.h5` | MISS procedure, phase 3 |
| `miss_4.h5` | MISS procedure, phase 4 |
To obtain a single merged file (including splits), use the merge utility from the [main EgoExOR repository](https://github.com/ardamamur/EgoExOR) (see `data/README.md`).
### HDF5 Schema
```
/metadata
/vocabulary/entity — Entity names and IDs (instruments, anatomy, etc.)
/vocabulary/relation — Relation names and IDs (holding, cutting, etc.)
/sources/sources — Camera/source names and IDs (head_surgeon, external_1, etc.)
/dataset — version, creation_date, title
/procedures/{procedure}/phases/{phase}/takes/{take}/
/sources — source_count, source_0, source_1, … (camera roles)
/frames/rgb — (num_frames, num_cameras, H, W, 3) uint8 — 1344×1344
/eye_gaze/coordinates — (num_frames, num_ego_cameras, 3) float32 — gaze 2D + camera ID
/eye_gaze_depth/values — (num_frames, num_ego_cameras) float32
/hand_tracking/positions — (num_frames, num_ego_cameras, 17) float32
/audio/waveform — Full stereo waveform
/audio/snippets — 1-second snippets aligned to frames
/audio/per_device/ — Per-microphone waveform and snippets
/point_cloud/depth/values — Raw depth images (external cameras; others zero-filled)
/point_cloud/merged/ — Not populated; use raw depth to build point clouds yourself
/annotations/ — Scene graph annotations (frame_idx, rel_annotations, scene_graph)
/splits
train, validation, test — Split tables (procedure, phase, take, frame_id)
```
**Note:** Camera/source IDs in `eye_gaze/coordinates` map to `metadata/sources` for correct source names.
## ⚙️ Efficiency and Usability
- **HDF5** — Hierarchical structure, partial loading, gzip compression
- **Chunking** — Efficient access to frame ranges for sequence-based training
- **Logical layout** — `procedures → phases → takes → modality` for easy navigation
## 📜 License
Released under the [Apache 2.0 License](https://www.apache.org/licenses/LICENSE-2.0). Free for academic and commercial use with attribution.
## 📚 Citation
```bibtex
@misc{özsoy2025egoexoregoexocentricoperatingroom,
title={EgoExOR: An Ego-Exo-Centric Operating Room Dataset for Surgical Activity Understanding},
author={Ege Özsoy and Arda Mamur and Felix Tristram and Chantal Pellegrini and Magdalena Wysocki and Benjamin Busam and Nassir Navab},
year={2025},
eprint={2505.24287},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2505.24287},
}
```
## 🔗 Related Resources
- **Original EgoExOR (v1)** — [ardamamur/EgoExOR](https://huggingface.co/datasets/ardamamur/EgoExOR) — 336×336 images, pre-merged point clouds, merged audio
- **Code, benchmarks, pretrained model** — [github.com/ardamamur/EgoExOR](https://github.com/ardamamur/EgoExOR)
---
**Dataset:** [TUM/EgoExOR](https://huggingface.co/datasets/TUM/EgoExOR) · **Last Updated:** February 2025 |