DISCOVR / README.md
dkillough's picture
Add VRSight citation again
3a96e5e verified
---
license: cc-by-4.0
task_categories:
- object-detection
language:
- en
tags:
- computer-vision
- object-detection
- yolo
- virtual-reality
- vr
- accessibility
- social-vr
pretty_name: DISCOVR - Virtual Reality UI Object Detection Dataset
size_categories:
- 10K<n<100K
dataset_info:
features:
- name: image
dtype: image
- name: objects
struct:
- name: class_id
list: int64
- name: center_x
list: float32
- name: center_y
list: float32
- name: width
list: float32
- name: height
list: float32
splits:
- name: train
num_bytes: 536592914
num_examples: 15207
- name: test
num_bytes: 29938152
num_examples: 839
- name: validation
num_bytes: 59182849
num_examples: 1645
download_size: 613839753
dataset_size: 625713915
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
---
# DIgitial Social Context Objects in VR (DISCOVR): A Social Virtual Reality Object Detection Dataset
## Dataset Description
DISCOVR is an object detection dataset for identifying user interface elements and interactive objects in virtual reality (VR) and social VR environments. The dataset contains **17,691 annotated images** across **30 object classes** commonly found in 17 top social VR applications & VR demos.
This dataset is designed to support research in VR accessibility, automatic UI analysis, and assistive technologies for virtual environments.
### The entire dataset is available to download at once at https://huggingface.co/datasets/UWMadAbility/DISCOVR/blob/main/dataset.zip
### Pretrained YOLOv8 weights are available at https://huggingface.co/UWMadAbility/VRSight
### If you use DISCOVR in your work, please cite our work VRSight for which it was developed:
```bibtex
@inproceedings{killough2025vrsight,
title={VRSight: An AI-Driven Scene Description System to Improve Virtual Reality Accessibility for Blind People},
author={Killough, Daniel and Feng, Justin and Ching, Zheng Xue and Wang, Daniel and Dyava, Rithvik and Tian, Yapeng and Zhao, Yuhang},
booktitle={Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology},
pages={1--17},
year={2025}
}
```
---
### Dataset Summary
- **Total Images:** 17,691
- Training: 15,207 images
- Validation: 1,645 images
- Test: 839 images
- **Classes:** 30 object categories
- **Format:** YOLOv8
- **License:** CC BY 4.0
## Object Classes
The dataset includes 30 classes of VR UI elements and interactive objects:
| ID | Class Name | Description |
|----|------------|-------------|
| 0 | avatar | User representations (human avatars) |
| 1 | avatar-nonhuman | Non-human avatar representations |
| 2 | button | Interactive buttons |
| 3 | campfire | Campfire objects (social gathering points) |
| 4 | chat box | Text chat interface elements |
| 5 | chat bubble | Speech/thought bubbles |
| 6 | controller | VR controller representations |
| 7 | dashboard | VR OS dashboard |
| 8 | guardian | Boundary/guardian system indicators (blue grid/plus signs) |
| 9 | hand | Hand representations |
| 10 | hud | Heads-up display elements |
| 11 | indicator-mute | Mute status indicators |
| 12 | interactable | Generic interactable objects |
| 13 | locomotion-target | Movement/teleportation targets |
| 14 | menu | Menu interfaces |
| 15 | out of bounds | Out-of-bounds warnings (red circle) |
| 16 | portal | Portal/doorway objects |
| 17 | progress bar | Progress indicators |
| 18 | seat-multiple | Multi-person seating |
| 19 | seat-single | Single-person seating |
| 20 | sign-graphic | Graphical signs |
| 21 | sign-text | Text-based signs |
| 22 | spawner | Object spawning points |
| 23 | table | Tables and surfaces |
| 24 | target | Target/aim points |
| 25 | ui-graphic | Graphical UI elements |
| 26 | ui-text | Text UI elements |
| 27 | watch | Watch/time displays |
| 28 | writing surface | Whiteboards/drawable surfaces |
| 29 | writing utensil | Drawing/writing tools |
## Dataset Structure
```
DISCOVR/
├── train/
│ ├── images/ # 15,207 training images (.jpg)
│ └── labels/ # YOLO format annotations (.txt)
├── validation/
│ ├── images/ # 1,645 validation images
│ └── labels/ # YOLO format annotations
├── test/
│ ├── images/ # 839 test images
│ └── labels/ # YOLO format annotations
└── data.yaml # Dataset configuration file
```
### Annotation Format
Annotations are in YOLO format with normalized coordinates:
```
<class_id> <center_x> <center_y> <width> <height>
```
All coordinates are normalized to [0, 1] range relative to image dimensions.
---
## Usage
### With Hugging Face Datasets
```python
from datasets import load_dataset
# Load the full dataset
dataset = load_dataset("UWMadAbility/DISCOVR")
# Access individual splits
train_data = dataset['train']
val_data = dataset['validation']
test_data = dataset['test']
# Example: Get first training image and its annotations
sample = train_data[0]
image = sample['image']
objects = sample['objects']
print(f"Number of objects: {len(objects['class_id'])}")
print(f"Class IDs: {objects['class_id']}")
print(f"Bounding boxes: {list(zip(objects['center_x'], objects['center_y'], objects['width'], objects['height']))}")
```
### With YOLOv8/Ultralytics
First, download the dataset and create a `data.yaml` file:
```yaml
path: ./DISCOVR
train: train/images
val: validation/images
test: test/images
nc: 30
names:
0: avatar
1: avatar-nonhuman
2: button
3: campfire
4: chat box
5: chat bubble
6: controller
7: dashboard
8: guardian
9: hand
10: hud
11: indicator-mute
12: interactable
13: locomotion-target
14: menu
15: out of bounds
16: portal
17: progress bar
18: seat-multiple
19: seat-single
20: sign-graphic
21: sign-text
22: spawner
23: table
24: target
25: ui-graphic
26: ui-text
27: watch
28: writing surface
29: writing utensil
```
Then train a model:
```python
from ultralytics import YOLO
# Load a pretrained model
model = YOLO('yolov8n.pt')
# Train the model
results = model.train(
data='data.yaml',
epochs=100,
imgsz=640,
batch=16
)
# Validate the model
metrics = model.val()
# Make predictions
results = model.predict('path/to/vr_image.jpg')
```
### With Transformers (DETR, etc.)
```python
from datasets import load_dataset
from transformers import AutoImageProcessor, AutoModelForObjectDetection
# Load dataset
dataset = load_dataset("UWMadAbility/DISCOVR")
# Load model and processor
processor = AutoImageProcessor.from_pretrained("facebook/detr-resnet-50")
model = AutoModelForObjectDetection.from_pretrained("facebook/detr-resnet-50")
# Process image
sample = dataset['train'][0]
inputs = processor(images=sample['image'], return_tensors="pt")
# Note: You'll need to convert YOLO format to COCO format for DETR
# YOLO: (center_x, center_y, width, height) normalized
# COCO: (x_min, y_min, width, height) in pixels
```
---
## Applications
This dataset can be used for:
- **VR Accessibility Research**: Automatically detecting and describing UI elements for users with disabilities
- **UI/UX Analysis**: Analyzing VR interface design patterns
- **Assistive Technologies**: Building screen readers and navigation aids for VR
- **Automatic Testing**: Testing VR applications for UI consistency
- **Content Moderation**: Detecting inappropriate content in social VR spaces
- **User Behavior Research**: Understanding how users interact with VR interfaces
## Citation
If you use this dataset in your research, please cite our publication using DISCOVR, called VRSight:
```bibtex
@inproceedings{killough2025vrsight,
title={VRSight: An AI-Driven Scene Description System to Improve Virtual Reality Accessibility for Blind People},
author={Killough, Daniel and Feng, Justin and Ching, Zheng Xue and Wang, Daniel and Dyava, Rithvik and Tian, Yapeng and Zhao, Yuhang},
booktitle={Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology},
pages={1--17},
year={2025}
}
```
## License
This dataset is released under the **Creative Commons Attribution 4.0 International (CC BY 4.0)** license.
You are free to:
- Share — copy and redistribute the material
- Adapt — remix, transform, and build upon the material
Under the following terms:
- Attribution — You must give appropriate credit
## Contact
For questions, issues, or collaborations:
- Main Codebase: https://github.com/MadisonAbilityLab/VRSight
- This Repository: [UWMadAbility/DISCOVR](https://huggingface.co/datasets/UWMadAbility/DISCOVR)
- Organization: UW-Madison Ability Lab
## Acknowledgments
This dataset was created by Daniel K., Justin, Daniel W., ZX, Ricky, Abhinav, and the MadAbility Lab at the University of Wisconsin-Madison to support research in VR accessibility and assistive technologies.