DISCOVR / README.md
dkillough's picture
Add VRSight citation again
3a96e5e verified
metadata
license: cc-by-4.0
task_categories:
  - object-detection
language:
  - en
tags:
  - computer-vision
  - object-detection
  - yolo
  - virtual-reality
  - vr
  - accessibility
  - social-vr
pretty_name: DISCOVR - Virtual Reality UI Object Detection Dataset
size_categories:
  - 10K<n<100K
dataset_info:
  features:
    - name: image
      dtype: image
    - name: objects
      struct:
        - name: class_id
          list: int64
        - name: center_x
          list: float32
        - name: center_y
          list: float32
        - name: width
          list: float32
        - name: height
          list: float32
  splits:
    - name: train
      num_bytes: 536592914
      num_examples: 15207
    - name: test
      num_bytes: 29938152
      num_examples: 839
    - name: validation
      num_bytes: 59182849
      num_examples: 1645
  download_size: 613839753
  dataset_size: 625713915
configs:
  - config_name: default
    data_files:
      - split: train
        path: data/train-*
      - split: test
        path: data/test-*
      - split: validation
        path: data/validation-*

DIgitial Social Context Objects in VR (DISCOVR): A Social Virtual Reality Object Detection Dataset

Dataset Description

DISCOVR is an object detection dataset for identifying user interface elements and interactive objects in virtual reality (VR) and social VR environments. The dataset contains 17,691 annotated images across 30 object classes commonly found in 17 top social VR applications & VR demos.

This dataset is designed to support research in VR accessibility, automatic UI analysis, and assistive technologies for virtual environments.

The entire dataset is available to download at once at https://huggingface.co/datasets/UWMadAbility/DISCOVR/blob/main/dataset.zip

Pretrained YOLOv8 weights are available at https://huggingface.co/UWMadAbility/VRSight

If you use DISCOVR in your work, please cite our work VRSight for which it was developed:

@inproceedings{killough2025vrsight,
  title={VRSight: An AI-Driven Scene Description System to Improve Virtual Reality Accessibility for Blind People},
  author={Killough, Daniel and Feng, Justin and Ching, Zheng Xue and Wang, Daniel and Dyava, Rithvik and Tian, Yapeng and Zhao, Yuhang},
  booktitle={Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology},
  pages={1--17},
  year={2025}
}

Dataset Summary

  • Total Images: 17,691
    • Training: 15,207 images
    • Validation: 1,645 images
    • Test: 839 images
  • Classes: 30 object categories
  • Format: YOLOv8
  • License: CC BY 4.0

Object Classes

The dataset includes 30 classes of VR UI elements and interactive objects:

ID Class Name Description
0 avatar User representations (human avatars)
1 avatar-nonhuman Non-human avatar representations
2 button Interactive buttons
3 campfire Campfire objects (social gathering points)
4 chat box Text chat interface elements
5 chat bubble Speech/thought bubbles
6 controller VR controller representations
7 dashboard VR OS dashboard
8 guardian Boundary/guardian system indicators (blue grid/plus signs)
9 hand Hand representations
10 hud Heads-up display elements
11 indicator-mute Mute status indicators
12 interactable Generic interactable objects
13 locomotion-target Movement/teleportation targets
14 menu Menu interfaces
15 out of bounds Out-of-bounds warnings (red circle)
16 portal Portal/doorway objects
17 progress bar Progress indicators
18 seat-multiple Multi-person seating
19 seat-single Single-person seating
20 sign-graphic Graphical signs
21 sign-text Text-based signs
22 spawner Object spawning points
23 table Tables and surfaces
24 target Target/aim points
25 ui-graphic Graphical UI elements
26 ui-text Text UI elements
27 watch Watch/time displays
28 writing surface Whiteboards/drawable surfaces
29 writing utensil Drawing/writing tools

Dataset Structure

DISCOVR/
├── train/
│   ├── images/        # 15,207 training images (.jpg)
│   └── labels/        # YOLO format annotations (.txt)
├── validation/
│   ├── images/        # 1,645 validation images
│   └── labels/        # YOLO format annotations
├── test/
│   ├── images/        # 839 test images
│   └── labels/        # YOLO format annotations
└── data.yaml          # Dataset configuration file

Annotation Format

Annotations are in YOLO format with normalized coordinates:

<class_id> <center_x> <center_y> <width> <height>

All coordinates are normalized to [0, 1] range relative to image dimensions.


Usage

With Hugging Face Datasets

from datasets import load_dataset

# Load the full dataset
dataset = load_dataset("UWMadAbility/DISCOVR")

# Access individual splits
train_data = dataset['train']
val_data = dataset['validation']
test_data = dataset['test']

# Example: Get first training image and its annotations
sample = train_data[0]
image = sample['image']
objects = sample['objects']

print(f"Number of objects: {len(objects['class_id'])}")
print(f"Class IDs: {objects['class_id']}")
print(f"Bounding boxes: {list(zip(objects['center_x'], objects['center_y'], objects['width'], objects['height']))}")

With YOLOv8/Ultralytics

First, download the dataset and create a data.yaml file:

path: ./DISCOVR
train: train/images
val: validation/images
test: test/images

nc: 30
names:
  0: avatar
  1: avatar-nonhuman
  2: button
  3: campfire
  4: chat box
  5: chat bubble
  6: controller
  7: dashboard
  8: guardian
  9: hand
  10: hud
  11: indicator-mute
  12: interactable
  13: locomotion-target
  14: menu
  15: out of bounds
  16: portal
  17: progress bar
  18: seat-multiple
  19: seat-single
  20: sign-graphic
  21: sign-text
  22: spawner
  23: table
  24: target
  25: ui-graphic
  26: ui-text
  27: watch
  28: writing surface
  29: writing utensil

Then train a model:

from ultralytics import YOLO

# Load a pretrained model
model = YOLO('yolov8n.pt')

# Train the model
results = model.train(
    data='data.yaml',
    epochs=100,
    imgsz=640,
    batch=16
)

# Validate the model
metrics = model.val()

# Make predictions
results = model.predict('path/to/vr_image.jpg')

With Transformers (DETR, etc.)

from datasets import load_dataset
from transformers import AutoImageProcessor, AutoModelForObjectDetection

# Load dataset
dataset = load_dataset("UWMadAbility/DISCOVR")

# Load model and processor
processor = AutoImageProcessor.from_pretrained("facebook/detr-resnet-50")
model = AutoModelForObjectDetection.from_pretrained("facebook/detr-resnet-50")

# Process image
sample = dataset['train'][0]
inputs = processor(images=sample['image'], return_tensors="pt")

# Note: You'll need to convert YOLO format to COCO format for DETR
# YOLO: (center_x, center_y, width, height) normalized
# COCO: (x_min, y_min, width, height) in pixels

Applications

This dataset can be used for:

  • VR Accessibility Research: Automatically detecting and describing UI elements for users with disabilities
  • UI/UX Analysis: Analyzing VR interface design patterns
  • Assistive Technologies: Building screen readers and navigation aids for VR
  • Automatic Testing: Testing VR applications for UI consistency
  • Content Moderation: Detecting inappropriate content in social VR spaces
  • User Behavior Research: Understanding how users interact with VR interfaces

Citation

If you use this dataset in your research, please cite our publication using DISCOVR, called VRSight:

@inproceedings{killough2025vrsight,
  title={VRSight: An AI-Driven Scene Description System to Improve Virtual Reality Accessibility for Blind People},
  author={Killough, Daniel and Feng, Justin and Ching, Zheng Xue and Wang, Daniel and Dyava, Rithvik and Tian, Yapeng and Zhao, Yuhang},
  booktitle={Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology},
  pages={1--17},
  year={2025}
}

License

This dataset is released under the Creative Commons Attribution 4.0 International (CC BY 4.0) license.

You are free to:

  • Share — copy and redistribute the material
  • Adapt — remix, transform, and build upon the material

Under the following terms:

  • Attribution — You must give appropriate credit

Contact

For questions, issues, or collaborations:

Acknowledgments

This dataset was created by Daniel K., Justin, Daniel W., ZX, Ricky, Abhinav, and the MadAbility Lab at the University of Wisconsin-Madison to support research in VR accessibility and assistive technologies.