|
|
--- |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- drone |
|
|
- drones |
|
|
- event_camera |
|
|
- event |
|
|
- neuromorphic |
|
|
- detection |
|
|
- tracking |
|
|
- forecasting |
|
|
- multimodal |
|
|
pretty_name: 'FRED: Florence RGB-Event Dataset' |
|
|
size_categories: |
|
|
- 1M<n<10M |
|
|
--- |
|
|
|
|
|
<p align="center" style="font-size:40px;"> |
|
|
</b>⭐ FRED: Florence RGB-Event Drone dataset ⭐</b> |
|
|
</p> |
|
|
|
|
|
|
|
|
<div align="center"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/sliders_auto/montage.webp?raw=true" |
|
|
alt="Demo HDR" |
|
|
width=720/> |
|
|
</div> |
|
|
|
|
|
|
|
|
Official repository for the **FRED dataset**, a large-scale multimodal dataset specifically designed for drone detection, tracking, and trajectory forecasting, with spatiotemprally **synchronized RGB and event data**. |
|
|
It includes **train** and **test** splits with zipped subfolders for each sequence. |
|
|
|
|
|
The dataset can also be **downloaded** from [here](https://drive.google.com/drive/folders/1pISIErXOx76xmCqkwhS3-azWOMlTKZMp?usp=share_link). |
|
|
The dataset splits in .txt format, along with the alternative **challenging split**, can be found [here](https://github.com/miccunifi/FRED/tree/main/dataset_splits). |
|
|
|
|
|
**Demos and examples** can be found in the official [website](https://miccunifi.github.io/FRED/). Check it out, it's pretty cool! :) |
|
|
|
|
|
|
|
|
--- |
|
|
|
|
|
## 📂 Dataset Structure |
|
|
|
|
|
``` |
|
|
FRED/ |
|
|
├── train/ |
|
|
│ ├── 0.zip |
|
|
│ ├── 1.zip |
|
|
│ └── ... |
|
|
├── test/ |
|
|
│ ├── 100.zip |
|
|
│ ├── 101.zip |
|
|
│ └── ... |
|
|
``` |
|
|
|
|
|
Each `.zip` file corresponds to one sequence (rgb frames, event data, and annotations). |
|
|
Event data comprends both already extracted event frames and the relative **.hdf5 file**, containing the raw event stream. |
|
|
|
|
|
--- |
|
|
|
|
|
## 📝 Annotation Format |
|
|
|
|
|
Each sequence includes two **`.txt` annotation file** with bounding box and identity information for every frame. |
|
|
Since rgb images are padded to enable a same coordinate space between the two modalities, the event videos have additional boxes corresponding to the padded area in the RGB. |
|
|
To better divide the 2 cases, we divided the annotation into **`coordinates.txt`** that represents the **extended boxes** and the **`coordinates_rgb.txt`** for the boxes **excluding the padding**. |
|
|
|
|
|
We reccomend using the **extended boxes** version since it facilitates the training and the overall cases in which the drone falls into the padding is relatively limited when compared to the number of samples. |
|
|
|
|
|
The format of the annotations is: |
|
|
|
|
|
``` |
|
|
time: x1, y1, x2, y2, id, class |
|
|
``` |
|
|
- **time** → time relative to the start of the recording in 'seconds.microseconds' for the annotation |
|
|
- **x1, y1** → top-left corner of the bounding box |
|
|
- **x2, y2** → bottom-right corner of the bounding box |
|
|
- **id** → unique identifier for the drone, consistent across frames (for tracking) |
|
|
- **class** → drone type label |
|
|
|
|
|
📌 Example: |
|
|
``` |
|
|
1.33332: 490.0, 413.0, 539.0, 448.0, 1, DJI Mini 2 |
|
|
6.33327: 609.0, 280.0, 651.0, 308.0, 2, DarwinFPV cineape20 |
|
|
``` |
|
|
|
|
|
This structure is compatible with standard detection and tracking pipelines, while maintaining instance-level identity across time. |
|
|
|
|
|
--- |
|
|
|
|
|
## 📥 Download |
|
|
|
|
|
### Clone the entire dataset |
|
|
```bash |
|
|
git lfs install |
|
|
git clone https://huggingface.co/datasets/GabrieleMagrini/FRED |
|
|
``` |
|
|
|
|
|
### Download specific sequences |
|
|
```bash |
|
|
wget https://huggingface.co/datasets/GabrieleMagrini/FRED/train/0.zip |
|
|
wget https://huggingface.co/datasets/GabrieleMagrini/FRED/test/100.zip |
|
|
``` |
|
|
|
|
|
### Use with 🤗 Datasets |
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
# Load full dataset |
|
|
ds = load_dataset("GabrieleMagrini/FRED") |
|
|
|
|
|
# Load specific split |
|
|
train_set = load_dataset("GabrieleMagrini/FRED", split="train") |
|
|
test_set = load_dataset("GabrieleMagrini/FRED", split="test") |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## 🖼️ Examples |
|
|
|
|
|
<div align="center"> |
|
|
<h1 style="font-size: 2rem; margin-bottom: 1rem; color: #fff;">Night</h1> |
|
|
</div> |
|
|
<div style="text-align:center;"> |
|
|
<div style="display:inline-block; margin:0 10px;"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/rgb_night.gif?raw=true" width="400"/> |
|
|
</div> |
|
|
<div style="display:inline-block; margin:0 10px;"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/event_night.gif?raw=true" width="400"/> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
|
|
|
<div align="center"> |
|
|
<h1 style="font-size: 2rem; margin-bottom: 1rem; color: #fff;">Raining</h1> |
|
|
</div> |
|
|
<div style="text-align:center;"> |
|
|
<div style="display:inline-block; margin:0 10px;"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/webps/rgb_rain.webp?raw=true" width="400"/> |
|
|
</div> |
|
|
<div style="display:inline-block; margin:0 10px;"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/webps/event_rain.webp?raw=true" width="400"/> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
<div align="center"> |
|
|
<h1 style="font-size: 2rem; margin-bottom: 1rem; color: #fff;">Indoor</h1> |
|
|
</div> |
|
|
<div style="text-align:center;"> |
|
|
<div style="display:inline-block; margin:0 10px;"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/rgb_indoor.gif?raw=true" width="400"/> |
|
|
</div> |
|
|
<div style="display:inline-block; margin:0 10px;"> |
|
|
<img src="https://github.com/miccunifi/FRED/blob/main/static/videos/event_indoor.gif?raw=true" width="400"/> |
|
|
</div> |
|
|
</div> |
|
|
|
|
|
|
|
|
|
|
|
--- |
|
|
|
|
|
## ✨ Citation |
|
|
|
|
|
If you use **FRED** in your research, please cite: |
|
|
|
|
|
``` |
|
|
@inproceedings{magrini2025fred, |
|
|
title={FRED: The Florence RGB-Event Drone Dataset}, |
|
|
author={Magrini, Gabriele and Marini, Niccol{`o} and Becattini, Federico and Berlincioni, Lorenzo and Biondi, Niccol{`o} and Pala, Pietro and Del Bimbo, Alberto}, |
|
|
booktitle={Proceedings of the 33rd ACM International conference on multimedia}, |
|
|
year={2025} |
|
|
} |
|
|
``` |