# Semantic2D: Enabling Semantic Scene Understanding with 2D Lidar Alone Implementation code for our paper ["Semantic2D: Enabling Semantic Scene Understanding with 2D Lidar Alone"](https://arxiv.org/pdf/2409.09899). Video demos can be found at [multimedia demonstrations](https://youtu.be/P1Hsvj6WUSY). The Semantic2D dataset can be found and downloaded at: https://doi.org/10.5281/zenodo.18350696. --- ## Related Resources - **Dataset Download:** https://doi.org/10.5281/zenodo.18350696 - **SALSA (Dataset and Labeling Framework):** https://github.com/TempleRAIL/semantic2d - **S³-Net (Stochastic Semantic Segmentation):** https://github.com/TempleRAIL/s3_net - **Semantic CNN Navigation:** https://github.com/TempleRAIL/semantic_cnn_nav ## Demos **S³-Net Segmentation** ![S³-Net Segmentation](demos/1.lobby_s3net_segmentation.gif) **Semantic Mapping** ![Semantic Mapping](demos/2.lobby_semantic_mapping.gif) **Semantic Navigation** ![Semantic Navigation](demos/3.lobby_semantic_navigation.gif) ## Table of Contents 1. [Requirements](#requirements) 2. [Installation](#installation) 3. [Dataset Description](#semantic2d-dataset-description) 4. [Semi-Automatic Labeling (SALSA)](#semi-automatic-labeling-usage) - [Step 1: Data Collection](#step-1-data-collection) - [Step 2: Manual Labeling](#step-2-manual-labeling) - [Step 3: Automatic Labeling](#step-3-automatic-labeling) 5. [Customizing for Different LiDAR Sensors](#customizing-for-different-lidar-sensors) 6. [Visualization](#visualization) 7. [Citation](#citation) --- ## Requirements - Ubuntu 20.04 - ROS Noetic - Python 3.8 - Labelme - scikit-learn - tqdm - PyTorch - NumPy - Pillow --- ## Installation ```bash # Clone the repository git clone https://github.com/TempleRAIL/semantic2d.git cd semantic2d # Install Python dependencies pip install labelme scikit-learn tqdm torch numpy pillow # Install LabelMe configuration with pre-defined semantic classes cp salsa/manually_labeling/.labelmerc ~/.labelmerc ``` --- ## Semantic2D Dataset Description The dataset contains the following data types: | Folder | Description | Shape | |--------|-------------|-------| | `scans_lidar/` | 2D LiDAR range data | (N,) array | | `intensities_lidar/` | 2D LiDAR intensity data | (N,) array | | `line_segments/` | Extracted line segments | List of [x1,y1,x2,y2] | | `positions/` | Robot position in map frame | (3,) array [x, y, yaw] | | `velocities/` | Robot velocity commands | (2,) array [Vx, Wz] | | `semantic_label/` | Point-wise semantic labels | (N,) array | ### Semantic Classes | ID | Class | Color (RGB) | |----|-------|-------------| | 0 | Other/Background | - | | 1 | Chair | (109, 0, 156) | | 2 | Door | (0, 46, 221) | | 3 | Elevator | (0, 164, 187) | | 4 | Person | (204, 204, 204) | | 5 | Pillar | (0, 155, 18) | | 6 | Sofa | (0, 225, 0) | | 7 | Table | (203, 249, 0) | | 8 | Trash bin | (255, 173, 0) | | 9 | Wall | (227, 0, 0) | --- ## Semi-Automatic Labeling Usage SALSA (Semi-Automatic Labeling framework for Semantic Annotation) consists of three steps: --- ### Step 1: Data Collection Collect and save data from a rosbag file. **Prerequisites:** You should have already created an environment map using a mapping package (e.g., `gmapping`) and collected raw rosbag data. #### 1.1 Configure Data Collection Edit `salsa/manually_labeling/dataset_collection.py`: ```python ################ CUSTOMIZATION REQUIRED ################ # Number of LiDAR points (must match your sensor) POINTS = 1081 # Hokuyo: 1081, WLR-716: 811, RPLIDAR-S2: 1972 # Output directory for collected data DATA_PATH = "~/semantic2d_data/2024-04-11-15-24-29" ``` #### 1.2 Configure Line Extraction Launch File The `laser_line_extraction` package extracts line features from LiDAR scans for ICP alignment. **You must configure it for your LiDAR sensor.** Edit `salsa/manually_labeling/semantic_data_collection_ws/src/laser_line_extraction/launch/example.launch`: ```xml ``` **Key parameters to change:** | Parameter | Description | Example Values | |-----------|-------------|----------------| | `frame_id` | TF frame of your LiDAR | `laser`, `base_scan`, `rear_laser`, `rplidar_link` | | `scan_topic` | ROS topic for LiDAR scans | `scan`, `/scan`, `/rplidar/scan` | | `min_range` | Minimum valid range (m) | Hokuyo: 0.1, WLR-716: 0.15, RPLIDAR: 0.2 | | `max_range` | Maximum valid range (m) | Hokuyo: 60.0, WLR-716: 25.0, RPLIDAR: 30.0 | **How to find your frame_id:** ```bash # Method 1: From rostopic rostopic echo /scan --noarr -n 1 | grep frame_id # Method 2: From tf tree rosrun tf view_frames # Creates frames.pdf ``` #### 1.3 Configure ROS Topics for Data Collection The default ROS topic subscriptions in `dataset_collection.py` are: | Topic | Message Type | Description | |-------|--------------|-------------| | `scan` | `sensor_msgs/LaserScan` | LiDAR scan data | | `line_segments` | `laser_line_extraction/LineSegmentList` | Line features | | `bluetooth_teleop/cmd_vel` | `geometry_msgs/Twist` | Velocity commands | | `robot_pose` | `geometry_msgs/PoseStamped` | Robot pose | **To customize for your robot**, modify the subscribers in `dataset_collection.py`: ```python # Example: For Hokuyo UTM-30LX-EW lidar self.scan_sub = rospy.Subscriber("/scan", LaserScan, self.scan_callback) self.dwa_cmd_sub = rospy.Subscriber('/cmd_vel', Twist, self.dwa_cmd_callback) self.robot_pose_pub = rospy.Subscriber('/amcl_pose', PoseWithCovarianceStamped, self.robot_pose_callback) # Example: For custom robot with namespaced topics self.scan_sub = rospy.Subscriber("/my_robot/laser/scan", LaserScan, self.scan_callback) ``` #### 1.4 Run Data Collection ```bash # Terminal 1: Start ROS master roscore # Terminal 2: Compile and launch line extraction cd salsa/manually_labeling/semantic_data_collection_ws catkin_make source devel/setup.bash roslaunch laser_line_extraction example.launch # Terminal 3: Start data collection cd salsa/manually_labeling python dataset_collection.py # Terminal 4: Play rosbag rosbag play your_data.bag ``` #### 1.5 Generate Train/Dev/Test Splits After data collection, generate index files (`train.txt`, `dev.txt`, `test.txt`) that define the dataset splits. **Configure** `salsa/manually_labeling/generateTrainDevSet.py`: ```python ################ CUSTOMIZATION REQUIRED ################ # The path of your dataset folder: train_folder = '~/semantic2d_data/2024-04-11-15-24-29' # Split percentages (must sum to 1.0) TRAIN_RATIO = 0.70 # 70% for training DEV_RATIO = 0.10 # 10% for validation/development TEST_RATIO = 0.20 # 20% for testing ######################################################## ``` **Run the script:** ```bash cd salsa/manually_labeling python generateTrainDevSet.py ``` **Example output:** ``` Dataset folder: /home/user/semantic2d_data/2024-04-11-15-24-29 Total samples: 20427 Split ratios: Train=70%, Dev=10%, Test=20% Split sizes: Train=14298, Dev=2042, Test=4087 Generated split files: - /home/user/.../train.txt - /home/user/.../dev.txt - /home/user/.../test.txt Done! ``` The script automatically: - Counts total samples in the `positions/` folder - Calculates split sizes based on the defined ratios - Shuffles data randomly before splitting - Generates the three `.txt` files with sample filenames --- ### Step 2: Manual Labeling Use LabelMe to manually label the environment map with semantic classes. #### 2.1 Run LabelMe ```bash labelme # Optionally, use RViz to visualize RGB images while labeling: roscore rosbag play your_data.bag rviz # Add Image display for camera topic ``` #### 2.2 Labeling Process 1. **Open** your occupancy grid map image (`.pgm` or `.png`) 2. **Create polygons** around each object (`Ctrl+N`) 3. **Select class** from dropdown: Chair, Door, Elevator, Pillar, Sofa, Table, Trash bin, Wall 4. **DO NOT label people** - they are automatically detected as dynamic objects 5. **Save** as `.json` file **Demo: How to use LabelMe** ![Labelme example](./salsa/manually_labeling/labelme_example.svg "labelme_example") ![Labelme demo](./demos/labelme_demo.gif "How to use LabelMe for semantic labeling") **Keyboard shortcuts:** | Shortcut | Action | |----------|--------| | `Ctrl+N` | Create new polygon | | `Ctrl+S` | Save annotation | | `Ctrl+Z` | Undo | | `Delete` | Delete selected polygon | | `Ctrl+E` | Edit label | | `D` | Next image | | `A` | Previous image | #### 2.3 Export Labeled Map ```bash # Export to label images labelme_export_json your_map.json -o labelme_output # For older labelme versions: # labelme_json_to_dataset your_map.json -o labelme_output ``` **Output structure** (see [labelme_output](./salsa/manually_labeling/labelme_output)): ``` labelme_output/ ├── img.png # Original map image ├── label.png # Semantic label image (class IDs) ├── label_viz.png # Colored visualization └── label_names.txt # Class name list ``` --- ### Step 3: Automatic Labeling Use ICP-based scan matching to automatically transfer labels from the map to each LiDAR scan. #### 3.1 Configure Automatic Labeling Edit `salsa/automatic_labeling/semi_automated_labeling_framework.py`: ```python ################ CUSTOMIZATION REQUIRED ################ # Dataset paths DATASET_ODIR = "/home/user/semantic2d_data/2024-04-04-12-16-41" DATASET_NAME = "train" # Options: train, dev, test # Map parameters (from your_map.yaml file) MAP_ORIGIN = np.array([-21.200000, -34.800000, 0.000000]) # [x, y, theta] MAP_RESOLUTION = 0.025000 # meters per pixel # Labeled map paths (from Step 2) MAP_LABEL_PATH = '../manually_labeling/labelme_output/label.png' MAP_PATH = '../manually_labeling/labelme_output/img.png' # LiDAR sensor parameters (see Customization section below) POINTS = 1081 AGNLE_MIN = -2.356194496154785 # -135 degrees in radians AGNLE_MAX = 2.356194496154785 # +135 degrees in radians RANGE_MAX = 60.0 # URDF: LiDAR to base_link transformation JOINT_XYZ = [-0.12, 0.0, 0.0] # [x, y, z] translation JOINT_RPY = [0.0, 0.0, 0.0] # [roll, pitch, yaw] rotation ``` #### 3.2 Run Automatic Labeling ```bash cd salsa/automatic_labeling python semi_automated_labeling_framework.py ``` **What the algorithm does:** 1. For each LiDAR scan: - Extract line features (stable structures like walls) - Use ICP to refine robot pose alignment with the map - Project LiDAR points to map coordinates - Match each point to semantic labels via pixel lookup - Points in free space → labeled as "Person" (dynamic objects) --- ## Customizing for Different LiDAR Sensors ### Supported Sensor Configurations The code includes pre-configured parameters for three sensors: | Parameter | Hokuyo UTM-30LX-EW | WLR-716 | RPLIDAR-S2 | |-----------|-------------------|---------|------------| | `POINTS` | 1081 | 811 | 1972 | | `ANGLE_MIN` (rad) | -2.356 (-135°) | -2.356 (-135°) | -3.142 (-180°) | | `ANGLE_MAX` (rad) | 2.356 (+135°) | 2.356 (+135°) | 3.142 (+180°) | | `RANGE_MIN` (m) | 0.1 | 0.15 | 0.2 | | `RANGE_MAX` (m) | 60.0 | 25.0 | 30.0 | | `JOINT_XYZ` | [-0.12, 0, 0] | [0.065, 0, 0.182] | [0.065, 0, 0.11] | | `JOINT_RPY` | [0, 0, 0] | [3.14, 0, 0] | [0, 0, 3.14] | | `frame_id` (launch) | `rear_laser` | `wlr716_link` | `rplidar_link` | | `scan_topic` (launch) | `scan` | `/wj716_base/scan` | `/rplidar_base/scan` | ### How to Configure Your Own Sensor #### Method 1: From ROS Topic ```bash # Get sensor parameters from ROS rostopic echo /scan --noarr -n 1 # Output shows: # angle_min: -2.356... # angle_max: 2.356... # angle_increment: 0.00436... # range_max: 60.0 # ranges: ``` #### Method 2: Calculate from Specifications ```python import numpy as np # From your sensor datasheet fov_degrees = 270 # Field of view angular_resolution = 0.25 # Degrees per point # Calculate parameters points = int(fov_degrees / angular_resolution) + 1 # = 1081 angle_min = -np.radians(fov_degrees / 2) # = -2.356 angle_max = np.radians(fov_degrees / 2) # = +2.356 ``` ### URDF Transformation (JOINT_XYZ, JOINT_RPY) Find the LiDAR mounting position from your robot's URDF file: ```xml ``` ```python # Use these values in the config JOINT_XYZ = [-0.12, 0.0, 0.0] # From xyz attribute JOINT_RPY = [0.0, 0.0, 0.0] # From rpy attribute ``` ### Complete Example: Adding a New Sensor In `semi_automated_labeling_framework.py`: ```python ################ CUSTOMIZATION REQUIRED ################ # Comment out existing configuration # # Hokuyo UTM-30LX-EW: # POINTS = 1081 # AGNLE_MIN = -2.356194496154785 # ... # Add YOUR sensor configuration: # SICK TiM561: POINTS = 811 # From rostopic echo /scan AGNLE_MIN = -2.356 # -135 degrees AGNLE_MAX = 2.356 # +135 degrees RANGE_MAX = 10.0 # 10 meters # URDF transformation (from robot model) JOINT_XYZ = [0.15, 0.0, 0.2] # Mounted 15cm forward, 20cm up JOINT_RPY = [0.0, 0.0, 0.0] # No rotation ``` ### Modifying Data Collection for Your Robot In `dataset_collection.py`: ```python ################ CUSTOMIZATION REQUIRED ################ # 1. Set number of points for your sensor POINTS = 811 # Your sensor's point count # 2. Set output directory DATA_PATH = "~/my_robot_data/environment_1" # 3. Modify ROS subscribers for your topics (in __init__): # Original (Jackal robot): self.scan_sub = rospy.Subscriber("scan", LaserScan, self.scan_callback) self.dwa_cmd_sub = rospy.Subscriber('bluetooth_teleop/cmd_vel', Twist, self.dwa_cmd_callback) self.robot_pose_pub = rospy.Subscriber('robot_pose', PoseStamped, self.robot_pose_callback) # For YOUR robot (example): self.scan_sub = rospy.Subscriber("/my_robot/scan", LaserScan, self.scan_callback) self.dwa_cmd_sub = rospy.Subscriber('/my_robot/cmd_vel', Twist, self.dwa_cmd_callback) self.robot_pose_pub = rospy.Subscriber('/amcl_pose', PoseWithCovarianceStamped, self.robot_pose_callback) ``` ### Quick Reference: Files to Modify | Task | File | Parameters to Change | |------|------|---------------------| | Line Extraction | `salsa/manually_labeling/semantic_data_collection_ws/src/laser_line_extraction/launch/example.launch` | `frame_id`, `scan_topic`, `min_range`, `max_range` | | Data Collection | `salsa/manually_labeling/dataset_collection.py` | `POINTS`, `DATA_PATH`, ROS topics | | Dataset Splits | `salsa/manually_labeling/generateTrainDevSet.py` | `train_folder`, `TRAIN_RATIO`, `DEV_RATIO`, `TEST_RATIO` | | Manual Labeling | `~/.labelmerc` | Label names/colors (optional) | | Auto Labeling | `salsa/automatic_labeling/semi_automated_labeling_framework.py` | `DATASET_ODIR`, `MAP_*`, `POINTS`, `ANGLE_*`, `JOINT_*` | | Visualization | `salsa/automatic_labeling/draw_semantic_label_sample.py` | `DATASET_ODIR`, `POINTS`, `ANGLE_*` | ### Launch File Configurations for Different Sensors **Hokuyo UTM-30LX-EW (Jackal robot):** ```xml ``` **WLR-716 (Custom robot):** ```xml ``` **RPLIDAR-S2 (Custom robot):** ```xml ``` --- ## Visualization Plot the labeled semantic LiDAR data: ```bash # Configure the same sensor parameters in draw_semantic_label_sample.py cd salsa/automatic_labeling python draw_semantic_label_sample.py ``` ![semantic_lidar_data](demos/semantic_lidar_data_2000.png "semantic_lidar_data_2000") --- ## Citation ```bibtex @article{xie2026semantic2d, title={Semantic2D: Enabling Semantic Scene Understanding with 2D Lidar Alone}, author={Xie, Zhanteng and Pan, Yipeng and Zhang, Yinqiang and Pan, Jia and Dames, Philip}, journal={arXiv preprint arXiv:2409.09899}, year={2026} } ```