| # Datasheet: BLV Object Recognition (Synthetic + Real-World) |
|
|
| Following the structure of *Datasheets for Datasets* (Gebru et al., 2018). |
|
|
| ## Motivation |
|
|
| **For what purpose was the dataset created?** |
| To enable training and evaluation of computer-vision models for blind and |
| low-vision (BLV) navigation aids. The dataset focuses on infrastructure |
| objects that BLV travelers must perceive and interact with (signals, doors, |
| escalators, handrails, etc.) and pairs photorealistic synthetic data for |
| training with real-world photographs (split into train / val / test) for |
| evaluation. |
|
|
| **Who funded the creation of the dataset?** |
| DARoS Lab. |
|
|
| ## Composition |
|
|
| **What do instances represent?** |
| Each instance is a single image with a paired pixel-level segmentation mask. |
| Synthetic instances additionally carry 2D bounding boxes per object. |
|
|
| **How many instances are there?** |
| - `syn/train`: 452704 |
| - `real_ours/train`: 3703 |
| - `real_ours/validation`: 396 |
| - `real_ours/test`: 1482 |
| - `real_curated/train`: 36466 |
| - `synthetic_objects/` (3D assets): 500 across 9 classes |
|
|
| **Does the dataset contain all possible instances?** |
| No. The synthetic data is a finite sample drawn from a parameterized |
| generation pipeline; the real-world data is a finite collection of |
| photographs. |
|
|
| **Is there any missing information?** |
| The synthetic-only class `turnstile` has no real-world examples in this |
| release. |
|
|
| **Are there errors, sources of noise, or redundancies?** |
| - Synthetic masks are produced by IsaacSim's Replicator and may contain edge |
| artifacts at sub-pixel object boundaries. |
| - Real-world polygon annotations were authored manually and may have small |
| boundary errors. |
|
|
| ## Collection process |
|
|
| **Synthetic.** Generated in NVIDIA IsaacSim using Replicator. Each trajectory |
| samples an asset, environment, and lighting condition, and records |
| RGB + semantic mask + 2D bounding box per frame. |
|
|
| **Real-world.** Photographs captured by data collectors at distinct physical |
| locations covering the 10 shared object classes; annotated with COCO-format |
| polygon segmentations. |
|
|
| ## Preprocessing / cleaning / labeling |
|
|
| See `README.md` and `scripts/build_hf_dataset.py`. The on-disk layout |
| re-encodes synthetic RGBA-coded masks into a single global palettized format |
| and rasterizes real-world COCO polygons into the same format. Source RGB PNGs |
| are not re-encoded. |
|
|
| ## Uses |
|
|
| **Has the dataset been used for any tasks already?** |
| The dataset will accompany a paper at the NeurIPS 2026 Datasets & Benchmarks |
| track (submission pending). |
|
|
| **What other tasks could the dataset be used for?** |
| Sim-to-real transfer studies, robustness analysis under lighting conditions, |
| multi-task learning combining detection and segmentation. |
|
|
| **Are there tasks for which the dataset should not be used?** |
| The dataset must not be used for surveillance or identification of |
| individuals. The synthetic data does not represent real people; the |
| real-world data was collected in public spaces and is intended only for |
| accessibility research. |
|
|
| ## Distribution |
|
|
| The dataset is hosted on Hugging Face at `NavAble/NeurIPS_2026_BLV` and |
| licensed under CC BY 4.0. |
|
|