File size: 3,805 Bytes
93c6961 d15aea0 93c6961 d15aea0 93c6961 d15aea0 93c6961 d569576 5868c0d | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 | ---
name: VVSim
tags:
- aerial-ground cooperative perception
- autonomous driving
- trajectory prediction
license: cc-by-4.0
task_categories:
- object-detection
- image-segmentation
- other
task_ids:
- semantic-segmentation
- vehicle-detection
---
# VVSim Dataset
**VVSim** is a large-scale dataset created for aerial–ground cooperative perception (AGCP). It integrates synchronized multimodal sensing data and state information collected simultaneously from vehicles and UAVs. The dataset contains **61K** fully annotated frames that cover **19** interaction scenarios (e.g., cut-in and lane change), along with **5** weather conditions (e.g., sunny, foggy, rainy, cloudy, snowy) and **11** scene types such as city, town, university, highway, and mountain environments. Beyond these frames, VVSim provides **255K** LiDAR sweeps and **3.5M** images (e.g., **1.2M** RGB images, **1.2M** semantic segmentation images, and **1.1M** depth images), accompanied by detailed annotations for 2D and 3D bounding boxes, object trajectories, and agent states.
## Data Annotation
VVSim provides comprehensive annotations for both 3D and 2D perception tasks. Each object in the scene is annotated with a **3D bounding box (986K)**, capturing its spatial extent and orientation in the world coordinate system. In addition, **2D bounding boxes** are generated via projection of the corresponding 3D boxes onto the image plane for camera-based tasks.
Beyond geometric information, VVSim records detailed object state attributes, including precise position `(x, y, z)`, rotation angles (roll, pitch, yaw), and velocity vectors. All annotated objects are categorized into one of **7** classes: **car, pedestrian, van, trailer, truck, traffic cone,** and **speed limit sign**, and the detailed size information is shown in the table below. These rich annotations enable comprehensive evaluation of both spatial localization and motion prediction, supporting a wide range of cooperative perception tasks across aerial and ground agents. The class distribution is illustrated in the figure below.
VVSim further provides **30** semantic segmentation categories that span road structures, dynamic agents, traffic infrastructure, vegetation, buildings, and diverse environmental objects.
## Coordinate Systems
VVSim employs several coordinate systems to represent sensor data:
- **LiDAR coordinate system (FLU):** x points forward, y points left, z points upward.
- **Ego-vehicle coordinate system (FRD):** x points forward, y points right, z points downward. All vehicle-mounted sensors are defined relative to this frame.
- **Global/world coordinate system (RFU):** x points right, y points forward, z points upward.
- **Camera coordinate system (FRD):** x points forward, y points right, z points downward.
## Sensor Parameters
In VVSim, both vehicle and aerial platforms are equipped with sensors to support diverse perception tasks. All sensors operate at a frame rate of 25 Hz.
**Vehicle sensors:**
- **Camera:** Four sets (front, rear, left, right), each consisting of RGB, depth, and semantic segmentation cameras. Resolution: 1600 × 900 px, FOV 90°. Approximate translations (m) relative to the vehicle origin: front (0.5, 0.0, -2.0), rear (-0.65, 0.0, -2.0), left (0.0, -0.5, -2.0), right (0.0, 0.5, -2.0). Camera orientations (yaw, pitch, roll, deg): front (0,0,0), rear (180,0,0), left (-90,0,0), right (90,0,0).
- **LiDAR:** One 64-beam LiDAR on top at (0.0, 0.0, -1.0) m, orientation (0,0,0), providing high-density 3D point clouds.
**UAV sensors:**
- **Camera:** Two downward-facing cameras providing RGB and semantic segmentation images. Resolution: 3840 × 2160 px, FOV 90°. Translation relative to drone origin: (0.0, 0.0, 0.2) m, orientation (0, -90, 0) deg (yaw, pitch, roll).
|