Datasets:
Languages:
English
Size:
n<1K
Tags:
robotics
drone-navigation
vision-language-navigation
open-vocabulary-detection
embodied-ai
habitat-sim
License:
File size: 4,737 Bytes
48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f 352d177 48fb42f | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 | ---
license: cc-by-nc-4.0
tags:
- robotics
- drone-navigation
- vision-language-navigation
- open-vocabulary-detection
- embodied-ai
- habitat-sim
- benchmark
- preview-sample
task_categories:
- object-detection
- depth-estimation
- robotics
size_categories:
- n<1K
pretty_name: "Yonder (NeurIPS reviewer sample)"
language:
- en
---
# Yonder Sample — NeurIPS 2026 Reviewer Preview
This is a **~500 MB sample** of the [Yonder](https://huggingface.co/datasets/astralhf/yonder)
drone navigation dataset, intended for fast inspection by NeurIPS reviewers and others
who want to verify the data format and content before downloading the full ~3.3 TB release.
## What's included
- **One HSSD scene:** `hssd-102343992`
- **50 consecutive waypoint NPZs** (`wp0000` through `wp0049`)
- **All 12 yaw orientations** per waypoint
- **All sensor modalities** present in the full dataset:
stereo RGB (left/right), forward depth, landing camera, up-IR, down-IR,
360° LiDAR, position, orientation
- **Semantic segmentation** for every waypoint (50 matching `*_semantics.npz` files)
- ~50 × 10 MB sensor + 50 × ~25 KB semantics ≈ **500 MB** total
## What's NOT in this sample
- **Multiple scenes** — by design. This sample is a single-scene slice. The full
release spans **167 HSSD scenes**, all with semantic annotations. Other scene
sources (ReplicaCAD, Replica, HM3D) considered during early collection have been
excluded from the public release for license-compatibility reasons; see the
[main dataset card](https://huggingface.co/datasets/astralhf/yonder) for details.
- **COCO bounding-box annotations** — derived programmatically from the semantic
channels and shipped per-scene under `annotations/` on the main repo.
## Quick start
```python
from huggingface_hub import snapshot_download
import numpy as np
local = snapshot_download(repo_id="astralhf/yonder-sample", repo_type="dataset")
# Sensor data
data = np.load(f"{local}/hssd-102343992_wp0000.npz")
print(sorted(data.keys())[:10])
# ['down_ir', 'lidar360', 'orientation', 'position', 'up_ir',
# 'yaw000_forward_depth', 'yaw000_landing_cam', 'yaw000_left_rgb',
# 'yaw000_right_rgb', 'yaw001_forward_depth']
print("left_rgb yaw000:", data["yaw000_left_rgb"].shape, data["yaw000_left_rgb"].dtype)
# left_rgb yaw000: (480, 640, 3) uint8
print("forward_depth yaw000:", data["yaw000_forward_depth"].shape, data["yaw000_forward_depth"].dtype)
# forward_depth yaw000: (480, 640) float16
print("lidar360:", data["lidar360"].shape, data["lidar360"].dtype)
# lidar360: (1024, 16) float32
# Semantic segmentation (one file per waypoint, 12 yaw keys)
sem = np.load(f"{local}/hssd-102343992_wp0000_semantics.npz")
print(sorted(sem.keys())[:4])
# ['yaw000_semantic', 'yaw030_semantic', 'yaw060_semantic', 'yaw090_semantic']
print("semantic yaw000:", sem["yaw000_semantic"].shape, sem["yaw000_semantic"].dtype)
# semantic yaw000: (480, 640) uint16 (per-pixel instance ID)
```
## Visualizing a frame
```python
import numpy as np
import matplotlib.pyplot as plt
data = np.load("hssd-102343992_wp0000.npz")
sem = np.load("hssd-102343992_wp0000_semantics.npz")
fig, axes = plt.subplots(1, 4, figsize=(20, 5))
axes[0].imshow(data["yaw000_left_rgb"]); axes[0].set_title("Left RGB")
axes[1].imshow(data["yaw000_right_rgb"]); axes[1].set_title("Right RGB")
axes[2].imshow(data["yaw000_forward_depth"], cmap="plasma")
axes[2].set_title("Forward depth (m)")
axes[3].imshow(sem["yaw000_semantic"], cmap="tab20")
axes[3].set_title("Semantic instance IDs")
for a in axes: a.axis("off")
plt.tight_layout(); plt.savefig("yonder_sample.png", dpi=150)
```
## Going to the full dataset
```python
# Single scene from the full repo (~25 GB sensor + semantics)
snapshot_download(
repo_id="astralhf/yonder",
repo_type="dataset",
allow_patterns=[
"indoor/drone-data/augmented/hssd-102343992/*.npz",
"semantics/hssd-102343992/*.npz",
"annotations/hssd-102343992/*.json",
],
)
# Just the manifests (a few MB) to plan a custom download
snapshot_download(
repo_id="astralhf/yonder",
repo_type="dataset",
allow_patterns="indoor/drone-data/augmented/*/manifest.json",
)
```
## License
CC-BY-NC-4.0 (inheriting HSSD's NonCommercial restriction). See the
[main dataset card](https://huggingface.co/datasets/astralhf/yonder) for full
licensing and Responsible AI considerations.
## Citation
```bibtex
@inproceedings{anonymous2026yonder,
title = {Yonder: A 4.65M-Frame Drone Navigation Dataset and the Cross-Simulator Generalization Gap},
author = {Anonymous Author(s)},
booktitle = {NeurIPS Datasets and Benchmarks Track},
year = {2026},
note = {Anonymized for double-blind review.}
}
```
|