Update README.md
Browse files
README.md
CHANGED
|
@@ -6,4 +6,54 @@ language:
|
|
| 6 |
- en
|
| 7 |
base_model:
|
| 8 |
- facebook/dinov2-large
|
| 9 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
- en
|
| 7 |
base_model:
|
| 8 |
- facebook/dinov2-large
|
| 9 |
+
---
|
| 10 |
+
|
| 11 |
+
# SynRS3D: A Synthetic Dataset for Global 3D Semantic Understanding from Monocular Remote Sensing Imagery
|
| 12 |
+
|
| 13 |
+
**Authors:**
|
| 14 |
+
[Jian Song](https://scholar.google.ch/citations?user=CgcMFJsAAAAJ&hl=zh-CN)<sup>1,2</sup>, [Hongruixuan Chen](https://scholar.google.ch/citations?user=XOk4Cf0AAAAJ&hl=zh-CN&oi=ao)<sup>1</sup>, [Weihao Xuan](https://weihaoxuan.com/)<sup>1,2</sup>, [Junshi Xia](https://scholar.google.com/citations?user=n1aKdTkAAAAJ&hl=en)<sup>2</sup>, [Naoto Yokoya](https://scholar.google.co.jp/citations?user=DJ2KOn8AAAAJ&hl=en)<sup>1,2</sup>
|
| 15 |
+
|
| 16 |
+
<sup>1</sup> The University of Tokyo
|
| 17 |
+
<sup>2</sup> RIKEN AIP
|
| 18 |
+
|
| 19 |
+
**Conference:** Neural Information Processing Systems (Spotlight), 2024
|
| 20 |
+
|
| 21 |
+
For more details, please refer to our [paper](https://arxiv.org/pdf/2406.18151) and visit our GitHub [repository](https://github.com/JTRNEO/SynRS3D).
|
| 22 |
+
|
| 23 |
+
---
|
| 24 |
+
|
| 25 |
+
### Overview
|
| 26 |
+
|
| 27 |
+
**TL;DR:**
|
| 28 |
+
We are excited to release two high-performing models for **height estimation** and **land cover mapping**. These models were trained on the SynRS3D dataset using our novel domain adaptation method, **RS3DAda**.
|
| 29 |
+
|
| 30 |
+
- **Encoder:** Vision Transformer (ViT-L), pretrained with **DINOv2**
|
| 31 |
+
- **Decoder:** [DPT](https://arxiv.org/abs/2103.13413), trained from scratch
|
| 32 |
+
|
| 33 |
+
These models excel in tasks involving large-scale global 3D semantic understanding from high-resolution remote sensing imagery. Feel free to integrate them into your projects for enhanced performance in related applications.
|
| 34 |
+
|
| 35 |
+
---
|
| 36 |
+
|
| 37 |
+
### How to Cite
|
| 38 |
+
|
| 39 |
+
If you find the RS3DAda model useful in your research, please consider citing:
|
| 40 |
+
|
| 41 |
+
```
|
| 42 |
+
@misc{song2024synrs3dsyntheticdatasetglobal,
|
| 43 |
+
title={SynRS3D: A Synthetic Dataset for Global 3D Semantic Understanding from Monocular Remote Sensing Imagery},
|
| 44 |
+
author={Jian Song and Hongruixuan Chen and Weihao Xuan and Junshi Xia and Naoto Yokoya},
|
| 45 |
+
year={2024},
|
| 46 |
+
eprint={2406.18151},
|
| 47 |
+
archivePrefix={arXiv},
|
| 48 |
+
primaryClass={cs.CV},
|
| 49 |
+
url={https://arxiv.org/abs/2406.18151},
|
| 50 |
+
}
|
| 51 |
+
```
|
| 52 |
+
|
| 53 |
+
---
|
| 54 |
+
|
| 55 |
+
### Contact
|
| 56 |
+
|
| 57 |
+
For any questions or feedback, please reach out via email at **song@ms.k.u-tokyo.ac.jp**.
|
| 58 |
+
|
| 59 |
+
We hope you enjoy using the pretrained RS3DAda models!
|