Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,128 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
annotations_creators:
|
| 2 |
+
- other
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
language_creators:
|
| 6 |
+
- other
|
| 7 |
+
license:
|
| 8 |
+
- odc-by
|
| 9 |
+
multilinguality:
|
| 10 |
+
- monolingual
|
| 11 |
+
pretty_name: 'RGB-D-SegmentEgocentricBodies '
|
| 12 |
+
size_categories:
|
| 13 |
+
- 1K<n<10K
|
| 14 |
+
source_datasets:
|
| 15 |
+
- original
|
| 16 |
+
tags:
|
| 17 |
+
- egocentric segmentation
|
| 18 |
+
- extended reality
|
| 19 |
+
- xr
|
| 20 |
+
- human-body
|
| 21 |
+
- mixed-reality
|
| 22 |
+
- avatar
|
| 23 |
+
task_categories:
|
| 24 |
+
- image-segmentation
|
| 25 |
+
- depth-estimation
|
| 26 |
+
task_ids:
|
| 27 |
+
- semantic-segmentation
|
| 28 |
+
- features:
|
| 29 |
+
- name: image
|
| 30 |
+
dtype: image
|
| 31 |
+
- name: depth
|
| 32 |
+
dtype: image
|
| 33 |
+
- name: mask
|
| 34 |
+
dtype: image
|
| 35 |
+
- name: synthetic_depth
|
| 36 |
+
dtype: image
|
| 37 |
+
-splits:
|
| 38 |
+
- name: train
|
| 39 |
+
num_examples: 8005
|
| 40 |
+
- name: validation
|
| 41 |
+
num_examples: 1069
|
| 42 |
+
|
| 43 |
+
# RGB-D Segment Egocentric Bodies Dataset
|
| 44 |
+
|
| 45 |
+
## Overview
|
| 46 |
+
|
| 47 |
+
The **RGB-D Segment Egocentric Bodies Dataset** is a multi-modal dataset designed for **egocentric body segmentation and depth-aware perception**. It contains synchronized **RGB images**, **real depth maps**, **segmentation masks**, and **synthetic depth data**, captured from an egocentric point of view.
|
| 48 |
+
The dataset is intended to support research in **XR/VR/AR**, **human–computer interaction**, and **depth-aware computer vision**.
|
| 49 |
+
|
| 50 |
+
## Dataset Description
|
| 51 |
+
|
| 52 |
+
The dataset is an extension of the EgoBodies Dataset (please refer to https://arxiv.org/pdf/2207.01296 for more information), with depth frames. We provide two versions of depth: real depth images acquired with different sensors:
|
| 53 |
+
RealSense D435, Realsense L515. Synthetic detph were estimated using Depth-Anything by Yang et al (2024). It is composed of more than 40 different users, in wild scenarios.
|
| 54 |
+
|
| 55 |
+
## Dataset Structure
|
| 56 |
+
|
| 57 |
+
RGB-D-SegmentEgocentricBodies/
|
| 58 |
+
│
|
| 59 |
+
├── train/ # ~3.11 GB
|
| 60 |
+
│ ├── images/ # RGB frames
|
| 61 |
+
│ ├── depths/ # Real depth maps
|
| 62 |
+
│ ├── masks/ # Segmentation masks
|
| 63 |
+
│ └── synthetic_depths/ # Synthetic or enhanced depth maps
|
| 64 |
+
│
|
| 65 |
+
├── val/ # ~401 MB
|
| 66 |
+
│ ├── images/
|
| 67 |
+
│ ├── depths/
|
| 68 |
+
│ ├── masks/
|
| 69 |
+
│ └── synthetic_depths/
|
| 70 |
+
│
|
| 71 |
+
└── .gitattributes # Git LFS configuration
|
| 72 |
+
|
| 73 |
+
## Intended Use
|
| 74 |
+
|
| 75 |
+
This dataset is suitable for:
|
| 76 |
+
|
| 77 |
+
- Egocentric human / body-part segmentation
|
| 78 |
+
- Depth-aware perception models
|
| 79 |
+
- XR avatar embodiment and telepresence
|
| 80 |
+
- Mixed-reality interaction research
|
| 81 |
+
- Training and benchmarking RGB-D models
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
## Example Usage
|
| 85 |
+
|
| 86 |
+
```python
|
| 87 |
+
from PIL import Image
|
| 88 |
+
import numpy as np
|
| 89 |
+
import os
|
| 90 |
+
|
| 91 |
+
def load_sample(root, split, idx):
|
| 92 |
+
base = os.path.join(root, split)
|
| 93 |
+
rgb = Image.open(os.path.join(base, "images", f"{idx}.png"))
|
| 94 |
+
depth = Image.open(os.path.join(base, "depths", f"{idx}.png"))
|
| 95 |
+
mask = Image.open(os.path.join(base, "masks", f"{idx}.png"))
|
| 96 |
+
synth = Image.open(os.path.join(base, "synthetic_depths", f"{idx}.png"))
|
| 97 |
+
return rgb, depth, mask, synth
|
| 98 |
+
|
| 99 |
+
## Acknowledgements
|
| 100 |
+
|
| 101 |
+
This dataset was created by ExtendedRealityLab and developed in the context of research on egocentric perception and immersive telepresence.
|
| 102 |
+
If you use this dataset in academic work, please cite the following papers:
|
| 103 |
+
|
| 104 |
+
@article{gonzalez2023full,
|
| 105 |
+
title={Full body video-based self-avatars for mixed reality: from e2e system to user study},
|
| 106 |
+
author={Gonzalez Morin, Diego and Gonzalez-Sosa, Ester and Perez, Pablo and Villegas, Alvaro},
|
| 107 |
+
journal={Virtual Reality},
|
| 108 |
+
volume={27},
|
| 109 |
+
number={3},
|
| 110 |
+
pages={2129--2147},
|
| 111 |
+
year={2023},
|
| 112 |
+
publisher={Springer}
|
| 113 |
+
}
|
| 114 |
+
|
| 115 |
+
@article{gonzalez2022real,
|
| 116 |
+
title={Real time egocentric segmentation for video-self avatar in mixed reality},
|
| 117 |
+
author={Gonzalez-Sosa, Ester and Gajic, Andrija and Gonzalez-Morin, Diego and Robledo, Guillermo and Perez, Pablo and Villegas, Alvaro},
|
| 118 |
+
journal={arXiv preprint arXiv:2207.01296},
|
| 119 |
+
year={2022}
|
| 120 |
+
}
|
| 121 |
+
|
| 122 |
+
@article{tobaruela2026egocentricrgbd,
|
| 123 |
+
title={RGB-D Egocentric Segmentation of Human Bodies for XR Applications},
|
| 124 |
+
author={Pedros-Tobaruela, Sofia and Gonzalez-Sosa, Ester and Perez, Pablo and Villegas, Alvaro},
|
| 125 |
+
journal={submitted}
|
| 126 |
+
}
|
| 127 |
+
|
| 128 |
+
|