Add pipeline tag, library name, and paper link
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,19 +1,22 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
| 3 |
datasets:
|
| 4 |
- heig-vd-geo/GridNet-HD
|
| 5 |
language:
|
| 6 |
- en
|
|
|
|
| 7 |
metrics:
|
| 8 |
- mean_iou
|
| 9 |
-
|
| 10 |
-
|
| 11 |
---
|
|
|
|
| 12 |
# GridNet-HD Baseline: Image semantic segmentation and LiDAR projection framework
|
| 13 |
|
| 14 |
## Overview
|
| 15 |
|
| 16 |
-
This repository provides a reproducible implementation of a semantic segmentation pipeline and 3D projection baseline used in
|
| 17 |
|
| 18 |
* A semantic segmentation pipeline transformer-based with `UperNetForSemanticSegmentation` (via HuggingFace Transformers).
|
| 19 |
* Support for high-resolution aerial imagery using random crop during training and sliding window inference at test time.
|
|
@@ -334,7 +337,4 @@ If you use this repo in research, please cite:
|
|
| 334 |
eprint={2601.13052},
|
| 335 |
url={https://arxiv.org/abs/2601.13052},
|
| 336 |
}
|
| 337 |
-
```
|
| 338 |
-
|
| 339 |
-
|
| 340 |
-
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- openmmlab/upernet-swin-tiny
|
| 4 |
datasets:
|
| 5 |
- heig-vd-geo/GridNet-HD
|
| 6 |
language:
|
| 7 |
- en
|
| 8 |
+
license: mit
|
| 9 |
metrics:
|
| 10 |
- mean_iou
|
| 11 |
+
library_name: transformers
|
| 12 |
+
pipeline_tag: image-segmentation
|
| 13 |
---
|
| 14 |
+
|
| 15 |
# GridNet-HD Baseline: Image semantic segmentation and LiDAR projection framework
|
| 16 |
|
| 17 |
## Overview
|
| 18 |
|
| 19 |
+
This repository provides a reproducible implementation of a semantic segmentation pipeline and 3D projection baseline used in the paper [GridNet-HD: A High-Resolution Multi-Modal Dataset for LiDAR-Image Fusion on Power Line Infrastructure](https://huggingface.co/papers/2601.13052). The framework includes:
|
| 20 |
|
| 21 |
* A semantic segmentation pipeline transformer-based with `UperNetForSemanticSegmentation` (via HuggingFace Transformers).
|
| 22 |
* Support for high-resolution aerial imagery using random crop during training and sliding window inference at test time.
|
|
|
|
| 337 |
eprint={2601.13052},
|
| 338 |
url={https://arxiv.org/abs/2601.13052},
|
| 339 |
}
|
| 340 |
+
```
|
|
|
|
|
|
|
|
|