| | --- |
| | license: cc-by-nc-4.0 |
| | size_categories: |
| | - 10GB<n<100GB |
| | pretty_name: SpaceSense-Bench |
| | task_categories: |
| | - object-detection |
| | - image-segmentation |
| | - depth-estimation |
| | - other |
| | tags: |
| | - space |
| | - satellite |
| | - multi-modal |
| | - lidar |
| | - pose-estimation |
| | --- |
| | |
| | # SpaceSense-Bench: Multi-Modal Spacecraft Perception and Pose Estimation Dataset |
| |
|
| | [**Project Page**](https://wuaodi.github.io/SpaceSense-Bench/) | [**Paper**](https://huggingface.co/papers/2603.09320) | [**Toolkit & Code**](https://github.com/wuaodi/SpaceSense-Bench) |
| |
|
| | SpaceSense-Bench is a high-fidelity simulation-based multi-modal (RGB, Depth, LiDAR Point Cloud) dataset for spacecraft component-level semantic understanding, containing **136 satellite models** with synchronized multi-modal data. |
| |
|
| |  |
| |
|
| | ## Dataset Overview |
| |
|
| | | Item | Detail | |
| | |------|--------| |
| | | Satellite Models | 136 (sourced from NASA/ESA 3D models) | |
| | | Data Modalities | RGB, Depth, Semantic Segmentation, LiDAR Point Cloud, 6-DoF Pose | |
| | | Image Resolution | 1024 x 1024 | |
| | | Camera FOV | 50 degrees | |
| | | Semantic Classes | 7 (main_body, solar_panel, dish_antenna, omni_antenna, payload, thruster, adapter_ring) | |
| | | Simulation Platform | Unreal Engine 5.2.0 + AirSim 1.8.1 | |
| | |
| | ## Sample Usage |
| | |
| | The [SpaceSense-Toolkit](https://github.com/wuaodi/SpaceSense-Bench/tree/main/SpaceSense-Toolkit) provides tools for converting raw data to standard formats and visualizing the results. |
| | |
| | ### Installation |
| | |
| | ```bash |
| | pip install -r requirements.txt |
| | ``` |
| | |
| | ### Conversion and Visualization |
| | |
| | ```bash |
| | # Visualize the raw data |
| | python SpaceSense-Toolkit/visualize/raw_data_web_visualizer.py --raw-data data_example |
| | |
| | # Convert to Semantic-KITTI (3D segmentation) |
| | python SpaceSense-Toolkit/convert/airsim_to_semantickitti.py --raw-data data_example --output output/semantickitti --satellite-json SpaceSense-Toolkit/configs/satellite_descriptions.json |
| | |
| | # Convert to MMSegmentation (2D segmentation) |
| | python SpaceSense-Toolkit/convert/airsim_to_mmseg.py --raw-data data_example --output output/mmseg |
| |
|
| | # Convert to YOLO (Object detection) |
| | python SpaceSense-Toolkit/convert/airsim_to_yolo.py --raw-data data_example --output output/yolo |
| | ``` |
| | |
| | ## Data Modalities |
| | |
| | | Modality | Format | Unit / Range | Description | |
| | |----------|--------|-------------|-------------| |
| | | RGB | PNG (1024x1024) | 8-bit color | Scene rendering | |
| | | Depth | PNG (1024x1024) | int32, millimeters (0 ~ 10,000,000 mm, background = 10,000 m) | Per-pixel depth map | |
| | | Semantic Segmentation | PNG (1024x1024) | uint8, class ID per pixel (0 = background) | Component-level segmentation mask | |
| | | LiDAR Point Cloud | ASC (x y z per line) | meters, 3 decimal places | Sparse 3D point cloud | |
| | | 6-DoF Pose | CSV | meters + Hamilton quaternion (w,x,y,z) | Camera-to-target relative pose | |
| | |
| | ## Coordinate System & Units |
| | |
| | | Item | Convention | |
| | |------|-----------| |
| | | Camera Frame | X-forward, Y-right, Z-down (right-hand system) | |
| | | World Frame | AirSim NED, target spacecraft fixed at origin | |
| | | Quaternion | Hamilton convention: w + xi + yj + zk | |
| | | Euler Angles | ZYX intrinsic (Yaw-Pitch-Roll) | |
| | | Position | meters (m), 6 decimal places | |
| | | Depth Map | millimeters (mm), int32; deep space background = 10,000 m | |
| | | LiDAR | meters (m), .asc format (x y z), 3 decimal places | |
| | | Timestamp | YYYYMMDDHHMMSSmmm | |
| | |
| | ## Sensor Configuration |
| | |
| | ### Camera (cam0) |
| | |
| | - Resolution: 1024 x 1024 |
| | - FOV: 50 degrees |
| | - Image types captured: RGB (type 0), Segmentation (type 5), Depth (type 2) |
| | - TargetGamma: 1.0 |
| | |
| | ### LiDAR |
| | |
| | - Range: 300 m |
| | - Channels: 256 |
| | - Vertical FOV: -20 to +20 degrees |
| | - Horizontal FOV: -20 to +20 degrees |
| | - Data frame: SensorLocalFrame |
| | |
| | ## Data Split (Zero-shot / OOD) |
| | |
| | The training and validation sets contain **completely non-overlapping satellite models**, so validation performance reflects zero-shot generalization to unseen spacecraft. |
| | |
| | | Split | Satellites | Rule | |
| | |-------|----------:|------| |
| | | Train | 117 | All satellites excluding val and excluded | |
| | | Test | 14 | Every 10th by index: seq 00, 10, 20, ..., 130 | |
| | | Validation | 5 | Seq 131-135, reserved for future testing | |
| | |
| | ## Data Organization |
| | |
| | Each `.tar.gz` file in the `raw/` folder contains data for one satellite: |
| | |
| | ``` |
| | <timestamp>_<satellite_name>/ |
| | ├── approach_front/ |
| | │ ├── rgb/ # RGB images (.png) |
| | │ ├── depth/ # Depth maps (.png, int32, mm) |
| | │ ├── segmentation/ # Semantic masks (.png, uint8) |
| | │ ├── lidar/ # Point clouds (.asc) |
| | │ └── poses.csv # 6-DoF poses |
| | ├── approach_back/ |
| | ├── orbit_xy/ |
| | └── ... |
| | ``` |
| | |
| | ## Semantic Class Definitions |
| | |
| | | Class ID | Name | Description | |
| | |:--------:|------|-------------| |
| | | 0 | background | Deep space background | |
| | | 1 | main_body | Spacecraft main body / bus | |
| | | 2 | solar_panel | Solar panels / solar arrays | |
| | | 3 | dish_antenna | Dish / parabolic antennas | |
| | | 4 | omni_antenna | Omnidirectional antennas / booms | |
| | | 5 | payload | Scientific instruments / payloads | |
| | | 6 | thruster | Thrusters / propulsion systems | |
| | | 7 | adapter_ring | Launch adapter rings | |
| |
|
| | ## License |
| |
|
| | This dataset is released under the [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license. Non-commercial use only. |
| |
|
| | ## Citation |
| |
|
| | ```bibtex |
| | @article{SpaceSense-Bench, |
| | title={SpaceSense-Bench: A Large-Scale Multi-Modal Benchmark for Spacecraft Perception and Pose Estimation}, |
| | author={Aodi Wu, Jianhong Zuo, Zeyuan Zhao, Xubo Luo, Ruisuo Wang, Xue Wan}, |
| | year={2026}, |
| | url={https://arxiv.org/abs/2603.09320} |
| | } |
| | ``` |