Update README.md
Browse files
README.md
CHANGED
|
@@ -1,54 +1,28 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
pipeline_tag: image-segmentation
|
| 4 |
-
|
|
|
|
|
|
|
| 5 |
---
|
| 6 |
|
| 7 |
-
# UAGLNet
|
| 8 |
|
| 9 |
-
**
|
| 10 |
|
| 11 |
-
|
| 12 |
|
| 13 |
-
|
|
|
|
|
|
|
| 14 |
|
| 15 |
-
|
|
|
|
|
|
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
## Quick Start
|
| 22 |
-
|
| 23 |
-
### Installation
|
| 24 |
-
|
| 25 |
-
Clone this repository and create the environment.
|
| 26 |
-
```bash
|
| 27 |
-
git git@github.com:Dstate/UAGLNet.git
|
| 28 |
-
cd UAGLNet
|
| 29 |
-
|
| 30 |
-
conda create -n uaglnet python=3.8 -y
|
| 31 |
-
conda activate uaglnet
|
| 32 |
-
conda install pytorch==1.10.0 torchvision==0.11.0 torchaudio==0.10.0 cudatoolkit=11.3 -c pytorch -c conda-forge
|
| 33 |
-
pip install -r requirements.txt
|
| 34 |
-
```
|
| 35 |
-
|
| 36 |
-
### Data Preprocessing
|
| 37 |
-
|
| 38 |
-
We conduct experiments on the Inria, WHU, and Massachusetts datasets. Detailed guidance for dataset preprocessing is provided here: [DATA_PREPARATION.md](https://github.com/Dstate/UAGLNet/blob/main/assets/DATA_PREPARATION.md).
|
| 39 |
-
|
| 40 |
-
### Training & Testing
|
| 41 |
-
|
| 42 |
-
Training and testing examples on the Inria dataset:
|
| 43 |
-
```bash
|
| 44 |
-
# training
|
| 45 |
-
python UAGLNet_train.py -c config/inria/UAGLNet.py
|
| 46 |
-
|
| 47 |
-
# testing
|
| 48 |
-
python UAGLNet_test.py -c config/inria/UAGLNet.py
|
| 49 |
-
```
|
| 50 |
-
|
| 51 |
-
### Main Results
|
| 52 |
|
| 53 |
The following table presents the performance of UAGLNet on building extraction benchmarks.
|
| 54 |
|
|
@@ -58,32 +32,13 @@ The following table presents the performance of UAGLNet on building extraction b
|
|
| 58 |
| Mass | 76.97 | 86.99 | 88.28 | 85.73 | [UAGLNet_Mass](https://huggingface.co/ldxxx/UAGLNet_Massachusetts) |
|
| 59 |
| WHU | 92.07 | 95.87 | 96.21 | 95.54 | [UAGLNet_WHU](https://huggingface.co/ldxxx/UAGLNet_WHU) |
|
| 60 |
|
| 61 |
-
You can quickly reproduce these results by running `Reproduce.py`, which will load the pretrained checkpoints from Hugging Face and perform inference.
|
| 62 |
-
|
| 63 |
-
```bash
|
| 64 |
-
# Inria
|
| 65 |
-
python Reproduce.py -d Inria
|
| 66 |
-
|
| 67 |
-
# Massachusetts
|
| 68 |
-
python Reproduce.py -d Mass
|
| 69 |
-
|
| 70 |
-
# WHU
|
| 71 |
-
python Reproduce.py -d WHU
|
| 72 |
-
```
|
| 73 |
-
|
| 74 |
## Citation
|
| 75 |
If you find this project useful in your research, please cite it as:
|
| 76 |
-
```
|
| 77 |
@article{UAGLNet,
|
| 78 |
title = {UAGLNet: Uncertainty-Aggregated Global-Local Fusion Network with Cooperative CNN-Transformer for Building Extraction},
|
| 79 |
author = {Siyuan Yao and Dongxiu Liu and Taotao Li and Shengjie Li and Wenqi Ren and Xiaochun Cao},
|
| 80 |
journal = {arXiv preprint arXiv:2512.12941},
|
| 81 |
year = {2025}
|
| 82 |
}
|
| 83 |
-
```
|
| 84 |
-
|
| 85 |
-
## Acknowledgement
|
| 86 |
-
This work is built upon [BuildingExtraction](https://github.com/stdcoutzrh/BuildingExtraction), [GeoSeg](https://github.com/WangLibo1995/GeoSeg/tree/main) and [SMT](https://github.com/AFeng-x/SMT). We sincerely appreciate their contributions which provide a clear pipeline and well-organized code.
|
| 87 |
-
|
| 88 |
-
## License
|
| 89 |
-
This project is licensed under the [Apache License 2.0](https://github.com/Dstate/UAGLNet/blob/main/LICENSE).
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
pipeline_tag: image-segmentation
|
| 4 |
+
tags:
|
| 5 |
+
- building-extraction
|
| 6 |
+
- remote-sensing
|
| 7 |
---
|
| 8 |
|
| 9 |
+
# UAGLNet: Uncertainty-Aggregated Global-Local Fusion Network with Cooperative CNN-Transformer for Building Extraction
|
| 10 |
|
| 11 |
+
This repository contains the official implementation of **UAGLNet**, a model for building extraction from remote sensing images, as presented in the paper *"UAGLNet: Uncertainty-Aggregated Global-Local Fusion Network with Cooperative CNN-Transformer for Building Extraction"*.
|
| 12 |
|
| 13 |
+
UAGLNet addresses the challenges of building extraction from remote sensing images due to complex structure variations. It proposes an Uncertainty-Aggregated Global-Local Fusion Network capable of exploiting high-quality global-local visual semantics under the guidance of uncertainty modeling. Specifically, it features a novel cooperative encoder with hybrid CNN and transformer layers, an intermediate cooperative interaction block (CIB) to narrow feature gaps, and a Global-Local Fusion (GLF) module. Additionally, an Uncertainty-Aggregated Decoder (UAD) is introduced to explicitly estimate pixel-wise uncertainty and mitigate segmentation ambiguity in uncertain regions.
|
| 14 |
|
| 15 |
+
## Paper
|
| 16 |
+
* **ArXiv:** [2512.12941](https://arxiv.org/abs/2512.12941)
|
| 17 |
+
* **Hugging Face Papers:** [2512.12941](https://huggingface.co/papers/2512.12941)
|
| 18 |
|
| 19 |
+
## Code
|
| 20 |
+
* **GitHub Repository:** [Dstate/UAGLNet](https://github.com/Dstate/UAGLNet)
|
| 21 |
+
* **Hugging Face Collection:** [ldxxx/uaglnet](https://huggingface.co/collections/ldxxx/uaglnet)
|
| 22 |
|
| 23 |
+
<img width="1000" src="https://github.com/Dstate/UAGLNet/raw/main/assets/architecture2.png">
|
| 24 |
|
| 25 |
+
## Main Results
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
The following table presents the performance of UAGLNet on building extraction benchmarks.
|
| 28 |
|
|
|
|
| 32 |
| Mass | 76.97 | 86.99 | 88.28 | 85.73 | [UAGLNet_Mass](https://huggingface.co/ldxxx/UAGLNet_Massachusetts) |
|
| 33 |
| WHU | 92.07 | 95.87 | 96.21 | 95.54 | [UAGLNet_WHU](https://huggingface.co/ldxxx/UAGLNet_WHU) |
|
| 34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
## Citation
|
| 36 |
If you find this project useful in your research, please cite it as:
|
| 37 |
+
```bibtex
|
| 38 |
@article{UAGLNet,
|
| 39 |
title = {UAGLNet: Uncertainty-Aggregated Global-Local Fusion Network with Cooperative CNN-Transformer for Building Extraction},
|
| 40 |
author = {Siyuan Yao and Dongxiu Liu and Taotao Li and Shengjie Li and Wenqi Ren and Xiaochun Cao},
|
| 41 |
journal = {arXiv preprint arXiv:2512.12941},
|
| 42 |
year = {2025}
|
| 43 |
}
|
| 44 |
+
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|