Datasets:

ArXiv:
File size: 13,153 Bytes
d46530b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
# Semantic2D Dataset

A 2D lidar semantic segmentation dataset for mobile robotics applications. This is the first publicly available 2D lidar semantic segmentation dataset, featuring point-wise annotations for nine indoor object categories across twelve distinct environments.

**Associated Paper:** *Semantic2D: Enabling Semantic Scene Understanding with 2D Lidar Alone*
**Authors:** Zhanteng Xie, Yipeng Pan, Yinqiang Zhang, Jia Pan, Philip Dames
**Institutions:** The University of Hong Kong, Temple University

**Video:** https://youtu.be/P1Hsvj6WUSY
**GitHub:** https://github.com/TempleRAIL/semantic2d

---

## Dataset Overview

| Property | Value |
|----------|-------|
| Total Data Tuples | ~188,007 |
| Recording Rate | 20 Hz |
| Total Duration | ~131 minutes |
| Environments | 12 indoor environments |
| Buildings | 7 buildings (4 at Temple, 3 at HKU) |
| LiDAR Sensors | 3 types |
| Semantic Classes | 9 + background |

### Semantic Classes

| ID | Class | Description |
|----|-------|-------------|
| 0 | Background/Other | Unclassified objects |
| 1 | Chair | Chairs and seating |
| 2 | Door | Doors and doorways |
| 3 | Elevator | Elevator doors |
| 4 | Person | Dynamic pedestrians |
| 5 | Pillar | Structural pillars |
| 6 | Sofa | Sofas and couches |
| 7 | Table | Tables |
| 8 | Trash bin | Trash cans |
| 9 | Wall | Walls and partitions |

### LiDAR Sensors

| Sensor | Robot | Location | Range (m) | FOV (deg) | Angular Res. (deg) | Points |
|--------|-------|----------|-----------|-----------|-------------------|--------|
| Hokuyo UTM-30LX-EW | Jackal | Temple | [0.1, 60] | 270 | 0.25 | 1,081 |
| WLR-716 | Customized | HKU | [0.15, 25] | 270 | 0.33 | 811 |
| RPLIDAR-S2 | Customized | HKU | [0.2, 30] | 360 | 0.18 | 1,972 |

---

## Directory Structure

```
semantic2d_dataset_2025/
├── README.md

├── rosbags/                               # Original ROS bag files, only used for reference or to extract other data, such as RGB, depth, goal, etc.
│   ├── 2024-04-04-12-16-41.bag            # Temple Engineering Lobby
│   ├── 2024-04-04-13-48-45.bag            # Temple Engineering 6th floor
│   ├── 2024-04-04-14-18-32.bag            # Temple Engineering 9th floor
│   ├── 2024-04-04-14-50-01.bag            # Temple Engineering 8th floor
│   ├── 2024-04-11-14-37-14.bag            # Temple Engineering 4th floor
│   ├── 2024-04-11-15-24-29.bag            # Temple Engineering Corridor
│   ├── 2025-07-08-13-32-08.bag            # Temple SERC Lobby
│   ├── 2025-07-08-14-22-44.bag            # Temple Gladfelter Lobby
│   ├── 2025-07-18-17-43-11.bag            # Temple Mazur Lobby
│   ├── 2025-11-10-15-53-51.bag            # HKU Chow Yei Ching 4th floor
│   ├── 2025-11-11-21-27-17.bag            # HKU Centennial Campus Lobby
│   └── 2025-11-18-22-13-37.bag            # HKU Jockey Club 3rd floor

└── semantic2d_data/                        # Semantic2D Segmentation Data files
    ├── dataset.txt                	    # dataset index for each folder
    ├── 2024-04-04-12-16-41/                # Temple Engineering Lobby 		# Temple environments (Hokuyo)
    ├── 2024-04-04-13-48-45/                # Temple Engineering 6th floor
    ├── 2024-04-04-14-18-32/                # Temple Engineering 9th floor
    ├── 2024-04-04-14-50-01/                # Temple Engineering 8th floor
    ├── 2024-04-11-14-37-14/                # Temple Engineering 4th floor
    ├── 2024-04-11-15-24-29/                # Temple Engineering Corridor
    ├── 2025-07-08-13-32-08/                # Temple SERC Lobby
    ├── 2025-07-08-14-22-44/		    # Temple Gladfelter Lobby
    ├── 2025-07-18-17-43-11/                # Temple Mazur Lobby
    ├── 2025-11-10-15-53-51/                # HKU Chow Yei Ching 4th floor	# HKU environments (WLR-716 + RPLIDAR)
    ├── 2025-11-11-21-27-17/		    # HKU Centennial Campus Lobby
    └── 2025-11-18-22-13-37/	            # HKU Jockey Club 3rd floo
```

---

## Semantic2D Segmentation Data Folder Structure

### Temple University Environments (Hokuyo UTM-30LX-EW)

Folders: `2024-04-04-*` and `2024-04-11-*` (6 environments)

```
2024-04-04-12-16-41/                        # Engineering Lobby
├── train.txt                               # Training split filenames (70%)
├── dev.txt                                 # Validation split filenames (10%)
├── test.txt                                # Test split filenames (20%)

├── scans_lidar/                            # LiDAR range data
│   └── *.npy                               # Shape: (1081,) - range values in meters

├── intensities_lidar/                      # LiDAR intensity data
│   └── *.npy                               # Shape: (1081,) - intensity values

├── line_segments/                          # Extracted line features
│   └── *.npy                               # Line segments [x1,y1,x2,y2] per line

├── positions/                              # Robot poses
│   └── *.npy                               # Shape: (3,) - [x, y, yaw] in map frame

├── velocities/                             # Velocity commands
│   └── *.npy                               # Shape: (2,) - [linear_x, angular_z]

├── semantic_label/                         # Point-wise semantic labels
│   └── *.npy                               # Shape: (1081,) - class ID per point

├── semantic_scan/                          # Semantic scan visualization data
│   └── *.npy

├── final_goals_local/                      # Navigation final goals
│   └── *.npy                               # Goal positions in local frame

├── sub_goals_local/                        # Navigation sub-goals
│   └── *.npy                               # Sub-goal waypoints

└── 202404041210_eng_lobby_map/             # Environment map
    ├── 202404041210_eng_lobby.pgm          # Occupancy grid map (PGM format)
    ├── 202404041210_eng_lobby.yaml         # Map configuration
    └── map_labelme/                        # Semantic labeled map
        ├── img.png                         # Original map image
        ├── label.png                       # Semantic label image
        ├── label_viz.png                   # Colored visualization
        ├── label_names.txt                 # Class name list
        └── map_labelme.json                # LabelMe annotation file
```

### HKU Environments (WLR-716 + RPLIDAR-S2)

Folders: `2025-11-*` (3 environments)

```
2025-11-10-15-53-51/                        # Chow Yei Ching 4th floor
├── train.txt                               # Training split (70%)
├── dev.txt                                 # Validation split (10%)
├── test.txt                                # Test split (20%)

├── # WLR-716 LiDAR data (811 points, 270 deg FOV)
├── scans_lidar_wlr716/                     # Range data
│   └── *.npy                               # Shape: (811,)
├── intensities_lidar_wlr716/               # Intensity data
│   └── *.npy                               # Shape: (811,)
├── line_segments_wlr716/                   # Line segments
│   └── *.npy
├── semantic_label_wlr716/                  # Semantic labels
│   └── *.npy                               # Shape: (811,)

├── # RPLIDAR-S2 data (1972 points, 360 deg FOV)
├── scans_lidar_rplidar/                    # Range data
│   └── *.npy                               # Shape: (1972,)
├── intensities_lidar_rplidar/              # Intensity data
│   └── *.npy                               # Shape: (1972,)
├── line_segments_rplidar/                  # Line segments
│   └── *.npy
├── semantic_label_rplidar/                 # Semantic labels
│   └── *.npy                               # Shape: (1972,)

├── # Shared data (same for both sensors)
├── positions/                              # Robot poses [x, y, yaw]
│   └── *.npy                               # Shape: (3,)
├── velocities/                             # Velocity commands [vx, wz]
│   └── *.npy                               # Shape: (2,)

└── 202511101415_cyc_4th_map/               # Environment map
    ├── *.pgm                               # Occupancy grid
    ├── *.yaml                              # Map configuration
    └── map_labelme/                        # Semantic labels
        └── ...
```

---

## Data Format Details

### LiDAR Scan Data (`.npy`)

```python
import numpy as np

# Load range data
scan = np.load('scans_lidar/0000001.npy')  # Shape: (N,) where N = num_points

# Hokuyo: N=1081, angle_min=-135°, angle_max=135°
# WLR-716: N=811, angle_min=-135°, angle_max=135°
# RPLIDAR-S2: N=1972, angle_min=-180°, angle_max=180°

# Convert to Cartesian coordinates
angles = np.linspace(angle_min, angle_max, num=N)
x = scan * np.cos(angles)
y = scan * np.sin(angles)
```

### Semantic Labels (`.npy`)

```python
# Load semantic labels
labels = np.load('semantic_label/0000001.npy')  # Shape: (N,)

# Each value is a class ID (0-9):
# 0: Background, 1: Chair, 2: Door, 3: Elevator, 4: Person,
# 5: Pillar, 6: Sofa, 7: Table, 8: Trash bin, 9: Wall
```

### Robot Pose (`.npy`)

```python
# Load robot pose
pose = np.load('positions/0000001.npy')  # Shape: (3,)
x, y, yaw = pose[0], pose[1], pose[2]  # Position and orientation in map frame
```

### Dataset Splits (`train.txt`, `dev.txt`, `test.txt`)

```
# Each line contains a .npy filename
0001680.npy
0007568.npy
0009269.npy
...
```

---

## Semantic2D labeling:

### 1. Data Collection (`dataset_collection.py`)

ROS node for collecting data from the robot during teleoperation.

**Key Features:**
- Subscribes to LiDAR scans (`/scan, /wj716_base/scan`, `/rplidar_base/scan`)
- Records range, intensity, line segments, positions, velocities
- Saves data at 20 Hz as `.npy` files


### 2. Semi-Automated Labeling Framework (SALSA)

**Algorithm:**
1. Load pre-labeled semantic map (from LabelMe)
2. For each LiDAR scan:
   - Extract line features for robust alignment
   - Apply ICP to refine robot pose
   - Project LiDAR points to map frame
   - Match points to semantic labels via pixel lookup
   - Points in free space labeled as "Person" (dynamic objects)

**Configuration (modify in script):**
```python
DATASET_ODIR = "/path/to/raw/data"
MAP_ORIGIN = np.array([-82.0, -71.6, 0.0])  # From map YAML
MAP_RESOLUTION = 0.025
POINTS = 811  # Number of LiDAR points
AGNLE_MIN = -2.356  # Min angle (radians)
AGNLE_MAX = 2.356   # Max angle (radians)
```

---

## Environments Summary

| Environment | Location | Folder | LiDAR |
|-------------|----------|--------|-------|
| Engineering Lobby | Temple | `2024-04-04-12-16-41` | Hokuyo |
| Engineering 6th Floor | Temple | `2024-04-04-13-48-45` | Hokuyo |
| Engineering 9th Floor | Temple | `2024-04-04-14-18-32` | Hokuyo |
| Engineering 8th Floor | Temple | `2024-04-04-14-50-01` | Hokuyo |
| Engineering 4th Floor | Temple | `2024-04-11-14-37-14` | Hokuyo |
| Engineering Corridor | Temple | `2024-04-11-15-24-29` | Hokuyo |
| SERC Lobby | Temple | `2025-07-08-13-32-08` | Hokuyo |
| Gladfelter Lobby | Temple | `2025-07-08-14-22-44` | Hokuyo |
| Mazur Lobby | Temple | `2025-07-18-17-43-11` | Hokuyo |
| Chow Yei Ching 4th Floor | HKU | `2025-11-10-15-53-51` | WLR-716/RPLIDAR |
| Centennial Campus Lobby | HKU | `2025-11-11-21-27-17` | WLR-716/RPLIDAR |
| Jockey Club 3rd Floor | HKU | `2025-11-18-22-13-37` | WLR-716/RPLIDAR |

---

## ROS Bag Contents

The original ROS bags in `rosbags/` contain:

| Topic | Message Type | Description |
|-------|--------------|-------------|
| `/scan` or `/*/scan` | `sensor_msgs/LaserScan` | LiDAR scans |
| `/robot_pose` | `geometry_msgs/PoseStamped` | Robot pose |
| `/cmd_vel` | `geometry_msgs/Twist` | Velocity commands |
| `/tf` | `tf2_msgs/TFMessage` | Transforms |
| `/map` | `nav_msgs/OccupancyGrid` | Occupancy map |
| `/camera/*` | `sensor_msgs/Image` | RGB/Depth images |
| `/odom` | `nav_msgs/Odometry` | Odometry |

---

## Related Resources

- **SALSA (Dataset and Labeling Framework):** https://github.com/TempleRAIL/semantic2d
- **S3-Net (Segmentation Algorithm):** https://github.com/TempleRAIL/s3_net
- **Semantic CNN Navigation:** https://github.com/TempleRAIL/semantic_cnn_nav
- **Dataset Zenodo:** DOI: 10.5281/zenodo.18350696

---

## Citation

```bibtex
@article{xie2026semantic2d,
  title={Semantic2D: Enabling Semantic Scene Understanding with 2D Lidar Alone},
  author={Xie, Zhanteng and Pan, Yipeng and Zhang, Yinqiang and Pan, Jia and Dames, Philip},
  journal={arXiv preprint arXiv:2409.09899},
  year={2026}
}
```

---

## License

Please refer to the associated paper and GitHub repository for licensing information.

## Contact

- Zhanteng Xie: zhanteng@hku.hk