Datasets:
File size: 5,601 Bytes
2b0bec7 352a275 2b0bec7 352a275 2b0bec7 352a275 5bcc0da 2b0bec7 352a275 2b0bec7 352a275 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 | ---
license: cc-by-nc-4.0
size_categories:
- 10GB<n<100GB
pretty_name: SpaceSense-Bench
task_categories:
- object-detection
- image-segmentation
- depth-estimation
- other
tags:
- space
- satellite
- multi-modal
- lidar
- pose-estimation
---
# SpaceSense-Bench: Multi-Modal Spacecraft Perception and Pose Estimation Dataset
[**Project Page**](https://wuaodi.github.io/SpaceSense-Bench/) | [**Paper**](https://huggingface.co/papers/2603.09320) | [**Toolkit & Code**](https://github.com/wuaodi/SpaceSense-Bench)
SpaceSense-Bench is a high-fidelity simulation-based multi-modal (RGB, Depth, LiDAR Point Cloud) dataset for spacecraft component-level semantic understanding, containing **136 satellite models** with synchronized multi-modal data.

## Dataset Overview
| Item | Detail |
|------|--------|
| Satellite Models | 136 (sourced from NASA/ESA 3D models) |
| Data Modalities | RGB, Depth, Semantic Segmentation, LiDAR Point Cloud, 6-DoF Pose |
| Image Resolution | 1024 x 1024 |
| Camera FOV | 50 degrees |
| Semantic Classes | 7 (main_body, solar_panel, dish_antenna, omni_antenna, payload, thruster, adapter_ring) |
| Simulation Platform | Unreal Engine 5.2.0 + AirSim 1.8.1 |
## Sample Usage
The [SpaceSense-Toolkit](https://github.com/wuaodi/SpaceSense-Bench/tree/main/SpaceSense-Toolkit) provides tools for converting raw data to standard formats and visualizing the results.
### Installation
```bash
pip install -r requirements.txt
```
### Conversion and Visualization
```bash
# Visualize the raw data
python SpaceSense-Toolkit/visualize/raw_data_web_visualizer.py --raw-data data_example
# Convert to Semantic-KITTI (3D segmentation)
python SpaceSense-Toolkit/convert/airsim_to_semantickitti.py --raw-data data_example --output output/semantickitti --satellite-json SpaceSense-Toolkit/configs/satellite_descriptions.json
# Convert to MMSegmentation (2D segmentation)
python SpaceSense-Toolkit/convert/airsim_to_mmseg.py --raw-data data_example --output output/mmseg
# Convert to YOLO (Object detection)
python SpaceSense-Toolkit/convert/airsim_to_yolo.py --raw-data data_example --output output/yolo
```
## Data Modalities
| Modality | Format | Unit / Range | Description |
|----------|--------|-------------|-------------|
| RGB | PNG (1024x1024) | 8-bit color | Scene rendering |
| Depth | PNG (1024x1024) | int32, millimeters (0 ~ 10,000,000 mm, background = 10,000 m) | Per-pixel depth map |
| Semantic Segmentation | PNG (1024x1024) | uint8, class ID per pixel (0 = background) | Component-level segmentation mask |
| LiDAR Point Cloud | ASC (x y z per line) | meters, 3 decimal places | Sparse 3D point cloud |
| 6-DoF Pose | CSV | meters + Hamilton quaternion (w,x,y,z) | Camera-to-target relative pose |
## Coordinate System & Units
| Item | Convention |
|------|-----------|
| Camera Frame | X-forward, Y-right, Z-down (right-hand system) |
| World Frame | AirSim NED, target spacecraft fixed at origin |
| Quaternion | Hamilton convention: w + xi + yj + zk |
| Euler Angles | ZYX intrinsic (Yaw-Pitch-Roll) |
| Position | meters (m), 6 decimal places |
| Depth Map | millimeters (mm), int32; deep space background = 10,000 m |
| LiDAR | meters (m), .asc format (x y z), 3 decimal places |
| Timestamp | YYYYMMDDHHMMSSmmm |
## Sensor Configuration
### Camera (cam0)
- Resolution: 1024 x 1024
- FOV: 50 degrees
- Image types captured: RGB (type 0), Segmentation (type 5), Depth (type 2)
- TargetGamma: 1.0
### LiDAR
- Range: 300 m
- Channels: 256
- Vertical FOV: -20 to +20 degrees
- Horizontal FOV: -20 to +20 degrees
- Data frame: SensorLocalFrame
## Data Split (Zero-shot / OOD)
The training and validation sets contain **completely non-overlapping satellite models**, so validation performance reflects zero-shot generalization to unseen spacecraft.
| Split | Satellites | Rule |
|-------|----------:|------|
| Train | 117 | All satellites excluding val and excluded |
| Test | 14 | Every 10th by index: seq 00, 10, 20, ..., 130 |
| Validation | 5 | Seq 131-135, reserved for future testing |
## Data Organization
Each `.tar.gz` file in the `raw/` folder contains data for one satellite:
```
<timestamp>_<satellite_name>/
├── approach_front/
│ ├── rgb/ # RGB images (.png)
│ ├── depth/ # Depth maps (.png, int32, mm)
│ ├── segmentation/ # Semantic masks (.png, uint8)
│ ├── lidar/ # Point clouds (.asc)
│ └── poses.csv # 6-DoF poses
├── approach_back/
├── orbit_xy/
└── ...
```
## Semantic Class Definitions
| Class ID | Name | Description |
|:--------:|------|-------------|
| 0 | background | Deep space background |
| 1 | main_body | Spacecraft main body / bus |
| 2 | solar_panel | Solar panels / solar arrays |
| 3 | dish_antenna | Dish / parabolic antennas |
| 4 | omni_antenna | Omnidirectional antennas / booms |
| 5 | payload | Scientific instruments / payloads |
| 6 | thruster | Thrusters / propulsion systems |
| 7 | adapter_ring | Launch adapter rings |
## License
This dataset is released under the [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license. Non-commercial use only.
## Citation
```bibtex
@article{SpaceSense-Bench,
title={SpaceSense-Bench: A Large-Scale Multi-Modal Benchmark for Spacecraft Perception and Pose Estimation},
author={Aodi Wu, Jianhong Zuo, Zeyuan Zhao, Xubo Luo, Ruisuo Wang, Xue Wan},
year={2026},
url={https://arxiv.org/abs/2603.09320}
}
``` |