File size: 5,601 Bytes
2b0bec7
 
 
352a275
 
 
 
 
 
 
 
 
 
 
 
 
2b0bec7
 
 
 
352a275
2b0bec7
352a275
5bcc0da
 
 
2b0bec7
 
 
 
 
 
 
 
 
 
 
352a275
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2b0bec7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
352a275
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
---
license: cc-by-nc-4.0
size_categories:
- 10GB<n<100GB
pretty_name: SpaceSense-Bench
task_categories:
- object-detection
- image-segmentation
- depth-estimation
- other
tags:
- space
- satellite
- multi-modal
- lidar
- pose-estimation
---

# SpaceSense-Bench: Multi-Modal Spacecraft Perception and Pose Estimation Dataset

[**Project Page**](https://wuaodi.github.io/SpaceSense-Bench/) | [**Paper**](https://huggingface.co/papers/2603.09320) | [**Toolkit & Code**](https://github.com/wuaodi/SpaceSense-Bench)

SpaceSense-Bench is a high-fidelity simulation-based multi-modal (RGB, Depth, LiDAR Point Cloud) dataset for spacecraft component-level semantic understanding, containing **136 satellite models** with synchronized multi-modal data.

![teaser](https://cdn-uploads.huggingface.co/production/uploads/65a5e3a86145ebc6e7fefb20/e8ehYJXlWgR7vhq7LDLwM.png)

## Dataset Overview

| Item | Detail |
|------|--------|
| Satellite Models | 136 (sourced from NASA/ESA 3D models) |
| Data Modalities | RGB, Depth, Semantic Segmentation, LiDAR Point Cloud, 6-DoF Pose |
| Image Resolution | 1024 x 1024 |
| Camera FOV | 50 degrees |
| Semantic Classes | 7 (main_body, solar_panel, dish_antenna, omni_antenna, payload, thruster, adapter_ring) |
| Simulation Platform | Unreal Engine 5.2.0 + AirSim 1.8.1 |

## Sample Usage

The [SpaceSense-Toolkit](https://github.com/wuaodi/SpaceSense-Bench/tree/main/SpaceSense-Toolkit) provides tools for converting raw data to standard formats and visualizing the results.

### Installation

```bash
pip install -r requirements.txt
```

### Conversion and Visualization

```bash
# Visualize the raw data
python SpaceSense-Toolkit/visualize/raw_data_web_visualizer.py --raw-data data_example

# Convert to Semantic-KITTI (3D segmentation)
python SpaceSense-Toolkit/convert/airsim_to_semantickitti.py --raw-data data_example --output output/semantickitti --satellite-json SpaceSense-Toolkit/configs/satellite_descriptions.json

# Convert to MMSegmentation (2D segmentation)
python SpaceSense-Toolkit/convert/airsim_to_mmseg.py --raw-data data_example --output output/mmseg

# Convert to YOLO (Object detection)
python SpaceSense-Toolkit/convert/airsim_to_yolo.py --raw-data data_example --output output/yolo
```

## Data Modalities

| Modality | Format | Unit / Range | Description |
|----------|--------|-------------|-------------|
| RGB | PNG (1024x1024) | 8-bit color | Scene rendering |
| Depth | PNG (1024x1024) | int32, millimeters (0 ~ 10,000,000 mm, background = 10,000 m) | Per-pixel depth map |
| Semantic Segmentation | PNG (1024x1024) | uint8, class ID per pixel (0 = background) | Component-level segmentation mask |
| LiDAR Point Cloud | ASC (x y z per line) | meters, 3 decimal places | Sparse 3D point cloud |
| 6-DoF Pose | CSV | meters + Hamilton quaternion (w,x,y,z) | Camera-to-target relative pose |

## Coordinate System & Units

| Item | Convention |
|------|-----------|
| Camera Frame | X-forward, Y-right, Z-down (right-hand system) |
| World Frame | AirSim NED, target spacecraft fixed at origin |
| Quaternion | Hamilton convention: w + xi + yj + zk |
| Euler Angles | ZYX intrinsic (Yaw-Pitch-Roll) |
| Position | meters (m), 6 decimal places |
| Depth Map | millimeters (mm), int32; deep space background = 10,000 m |
| LiDAR | meters (m), .asc format (x y z), 3 decimal places |
| Timestamp | YYYYMMDDHHMMSSmmm |

## Sensor Configuration

### Camera (cam0)

- Resolution: 1024 x 1024
- FOV: 50 degrees
- Image types captured: RGB (type 0), Segmentation (type 5), Depth (type 2)
- TargetGamma: 1.0

### LiDAR

- Range: 300 m
- Channels: 256
- Vertical FOV: -20 to +20 degrees
- Horizontal FOV: -20 to +20 degrees
- Data frame: SensorLocalFrame

## Data Split (Zero-shot / OOD)

The training and validation sets contain **completely non-overlapping satellite models**, so validation performance reflects zero-shot generalization to unseen spacecraft.

| Split | Satellites | Rule |
|-------|----------:|------|
| Train | 117 | All satellites excluding val and excluded |
| Test | 14 | Every 10th by index: seq 00, 10, 20, ..., 130 |
| Validation | 5 | Seq 131-135, reserved for future testing |

## Data Organization

Each `.tar.gz` file in the `raw/` folder contains data for one satellite:

```
<timestamp>_<satellite_name>/
├── approach_front/
│   ├── rgb/              # RGB images (.png)
│   ├── depth/            # Depth maps (.png, int32, mm)
│   ├── segmentation/     # Semantic masks (.png, uint8)
│   ├── lidar/            # Point clouds (.asc)
│   └── poses.csv         # 6-DoF poses
├── approach_back/
├── orbit_xy/
└── ...
```

## Semantic Class Definitions

| Class ID | Name | Description |
|:--------:|------|-------------|
| 0 | background | Deep space background |
| 1 | main_body | Spacecraft main body / bus |
| 2 | solar_panel | Solar panels / solar arrays |
| 3 | dish_antenna | Dish / parabolic antennas |
| 4 | omni_antenna | Omnidirectional antennas / booms |
| 5 | payload | Scientific instruments / payloads |
| 6 | thruster | Thrusters / propulsion systems |
| 7 | adapter_ring | Launch adapter rings |

## License

This dataset is released under the [CC-BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/) license. Non-commercial use only.

## Citation

```bibtex
@article{SpaceSense-Bench,
	title={SpaceSense-Bench: A Large-Scale Multi-Modal Benchmark for Spacecraft Perception and Pose Estimation},
	author={Aodi Wu, Jianhong Zuo, Zeyuan Zhao, Xubo Luo, Ruisuo Wang, Xue Wan},
	year={2026},
	url={https://arxiv.org/abs/2603.09320}
}
```