nielsr's picture
nielsr HF Staff
Improve dataset card: Enhance metadata, add intro, and link CARLARain code
9f1b2ca verified
|
raw
history blame
6.16 kB
---
language:
- en
license: cc-by-sa-4.0
size_categories:
- 1K<n<10K
task_categories:
- image-to-image
pretty_name: 'High-resolution Rainy Image'
tags:
- rain
- autonomous-driving
- driving-simulation
- semantic-segmentation
- synthetic-data
---
The **High-resolution Rainy Image (HRI) dataset** is a synthetic dataset created through a learning-from-rendering approach, detailed in the paper [Learning from Rendering: Realistic and Controllable Extreme Rainy Image Synthesis for Autonomous Driving Simulation](https://huggingface.co/papers/2502.16421). Designed for autonomous driving simulation, HRI provides realistic and controllable extreme rainy images to enhance visual perception models. It comprises 3,200 paired rainy-clean images, along with corresponding depth and rain layer mask images, captured across three diverse scenes (lane, citystreet, and japanesestreet) at a high resolution of 2048x1024. This dataset is particularly valuable for tasks such as semantic segmentation, instance segmentation, depth estimation, and object detection in challenging weather conditions, as demonstrated in the related work with the CARLARain simulator.
# High-resolution Rainy Image (HRI) Dataset
This is the dataset in the paper "[Learning from Rendering: Realistic and Controllable Extreme Rainy Image Synthesis for Autonomous Driving Simulation](https://arxiv.org/abs/2502.16421)".
* Project Page: https://kb824999404.github.io/HRIG/
* Paper: [Learning from Rendering: Realistic and Controllable Extreme Rainy Image Synthesis for Autonomous Driving Simulation](https://arxiv.org/abs/2502.16421)
* Code (Learning-from-Rendering): https://github.com/kb824999404/HRIG
* Code (CARLARain): https://github.com/kb824999404/CARLARain
<table>
<tr>
<td style="padding: 0;width=30%;"><img src="Imgs/lane/lane (1).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/lane/lane (2).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/lane/lane (3).jpg" /></td>
</tr>
<tr>
<td style="padding: 0;width=30%;"><img src="Imgs/lane/lane (4).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/lane/lane (5).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/lane/lane (6).jpg" /></td>
</tr>
<tr>
<td style="padding: 0;width=30%;"><img src="Imgs/citystreet/citystreet (1).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/citystreet/citystreet (2).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/citystreet/citystreet (3).jpg" /></td>
</tr>
<tr>
<td style="padding: 0;width=30%;"><img src="Imgs/citystreet/citystreet (4).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/citystreet/citystreet (5).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/citystreet/citystreet (6).jpg" /></td>
</tr>
<tr>
<td style="padding: 0;width=30%;"><img src="Imgs/japanesestreet/japanese (1).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/japanesestreet/japanese (2).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/japanesestreet/japanese (3).jpg" /></td>
</tr>
<tr>
<td style="padding: 0;width=30%;"><img src="Imgs/japanesestreet/japanese (4).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/japanesestreet/japanese (5).jpg" /></td>
<td style="padding: 0;width=30%;"><img src="Imgs/japanesestreet/japanese (6).jpg" /></td>
</tr>
</table>
## HRI Dataset
The High-resolution Rainy Image (HRI) dataset in the rendering stage.
<table style="text-align: center;">
<tr>
<th>scene</th>
<th>dataset type</th>
<th>resolution</th>
<th>viewpoints</th>
<th>moments</th>
<th>intensities</th>
<th>image pairs</th>
</tr>
<tr>
<td style="vertical-align: middle;" rowspan="2">lane</td>
<td>training set</td>
<td style="vertical-align: middle;" rowspan="2">2048×1024</td>
<td>3</td>
<td style="vertical-align: middle;" rowspan="2">100</td>
<td style="vertical-align: middle;" rowspan="2">4</td>
<td>1200</td>
</tr>
<tr>
<td>test set</td>
<td>1</td>
<td>400</td>
</tr>
<tr>
<td style="vertical-align: middle;" rowspan="2">citystreet</td>
<td>training set</td>
<td style="vertical-align: middle;" rowspan="2">2048×1024</td>
<td>5</td>
<td style="vertical-align: middle;" rowspan="2">25</td>
<td style="vertical-align: middle;" rowspan="2">4</td>
<td>500</td>
</tr>
<tr>
<td>test set</td>
<td>1</td>
<td>100</td>
</tr>
<tr>
<td style="vertical-align: middle;" rowspan="2">japanesestreet</td>
<td>training set</td>
<td style="vertical-align: middle;" rowspan="2">2048×1024</td>
<td>8</td>
<td style="vertical-align: middle;" rowspan="2">25</td>
<td style="vertical-align: middle;" rowspan="2">4</td>
<td>800</td>
</tr>
<tr>
<td>test set</td>
<td>2</td>
<td>200</td>
</tr>
</table>
* `clean`: background RGB images and depth images of all scenes.
* `rainy`: rain layer images, RGB rainy images and depth rainy images of all scenes.
* `trainset.json`: the sample lists of the training set.
* `testset.json`: the sample lists of the test set.
* For each sample in the training set and the test set:
* `scene`: the scene name
* `sequence`: the viewpoint name
* `intensity`: the rain intensity
* `wind`: the wind direction( all zero for the HRI dataset)
* `background`: the path of the background RGB image
* `depth`: the path of the background depth image
* `rain_layer`: the path of the rain layer image
* `rainy_depth`: the path of the rainy depth image
* `rainy_image`: the path of the rainy RGB image
## BlenderFiles
The Blender files for rendering RGB and depth images of all viewpoints are included in the directory of each scene.
## Rain streak database
The Rain streak database from the paper [Rain Rendering for Evaluating and Improving Robustness to Bad Weather](https://github.com/astra-vision/rain-rendering).
## Citation
When using these datasets, please cite our paper:
```
@article{zhou2025high,
title={High-resolution Rainy Image Synthesis: Learning from Rendering},
author={Zhou, Kaibin and Zhao, Shengjie and Deng, Hao and Zhang, Lin},
journal={arXiv preprint arXiv:2502.16421},
year={2025}
}
```