English
JTRNEO commited on
Commit
ffbf527
·
verified ·
1 Parent(s): d9a7e7d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -1
README.md CHANGED
@@ -6,4 +6,54 @@ language:
6
  - en
7
  base_model:
8
  - facebook/dinov2-large
9
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  - en
7
  base_model:
8
  - facebook/dinov2-large
9
+ ---
10
+
11
+ # SynRS3D: A Synthetic Dataset for Global 3D Semantic Understanding from Monocular Remote Sensing Imagery
12
+
13
+ **Authors:**
14
+ [Jian Song](https://scholar.google.ch/citations?user=CgcMFJsAAAAJ&hl=zh-CN)<sup>1,2</sup>, [Hongruixuan Chen](https://scholar.google.ch/citations?user=XOk4Cf0AAAAJ&hl=zh-CN&oi=ao)<sup>1</sup>, [Weihao Xuan](https://weihaoxuan.com/)<sup>1,2</sup>, [Junshi Xia](https://scholar.google.com/citations?user=n1aKdTkAAAAJ&hl=en)<sup>2</sup>, [Naoto Yokoya](https://scholar.google.co.jp/citations?user=DJ2KOn8AAAAJ&hl=en)<sup>1,2</sup>
15
+
16
+ <sup>1</sup> The University of Tokyo
17
+ <sup>2</sup> RIKEN AIP
18
+
19
+ **Conference:** Neural Information Processing Systems (Spotlight), 2024
20
+
21
+ For more details, please refer to our [paper](https://arxiv.org/pdf/2406.18151) and visit our GitHub [repository](https://github.com/JTRNEO/SynRS3D).
22
+
23
+ ---
24
+
25
+ ### Overview
26
+
27
+ **TL;DR:**
28
+ We are excited to release two high-performing models for **height estimation** and **land cover mapping**. These models were trained on the SynRS3D dataset using our novel domain adaptation method, **RS3DAda**.
29
+
30
+ - **Encoder:** Vision Transformer (ViT-L), pretrained with **DINOv2**
31
+ - **Decoder:** [DPT](https://arxiv.org/abs/2103.13413), trained from scratch
32
+
33
+ These models excel in tasks involving large-scale global 3D semantic understanding from high-resolution remote sensing imagery. Feel free to integrate them into your projects for enhanced performance in related applications.
34
+
35
+ ---
36
+
37
+ ### How to Cite
38
+
39
+ If you find the RS3DAda model useful in your research, please consider citing:
40
+
41
+ ```
42
+ @misc{song2024synrs3dsyntheticdatasetglobal,
43
+ title={SynRS3D: A Synthetic Dataset for Global 3D Semantic Understanding from Monocular Remote Sensing Imagery},
44
+ author={Jian Song and Hongruixuan Chen and Weihao Xuan and Junshi Xia and Naoto Yokoya},
45
+ year={2024},
46
+ eprint={2406.18151},
47
+ archivePrefix={arXiv},
48
+ primaryClass={cs.CV},
49
+ url={https://arxiv.org/abs/2406.18151},
50
+ }
51
+ ```
52
+
53
+ ---
54
+
55
+ ### Contact
56
+
57
+ For any questions or feedback, please reach out via email at **song@ms.k.u-tokyo.ac.jp**.
58
+
59
+ We hope you enjoy using the pretrained RS3DAda models!