Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,12 @@
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: mit
|
| 3 |
---
|
| 4 |
+
# SegFormer model fine-tuned on AROI
|
| 5 |
+
|
| 6 |
+
SegFormer model fine-tuned on AROI dataset [AROI: Annotated Retinal OCT Images Database](https://ieeexplore.ieee.org/abstract/document/9596934).
|
| 7 |
+
|
| 8 |
+
Disclaimer: The team releasing SegFormer did not write a model card for this model so this model card has been written by the Hugging Face team.
|
| 9 |
+
|
| 10 |
+
## Model description
|
| 11 |
+
|
| 12 |
+
SegFormer consists of a hierarchical Transformer encoder and a lightweight all-MLP decode head to achieve great results on semantic segmentation benchmarks such as ADE20K and Cityscapes. The hierarchical Transformer is first pre-trained on ImageNet-1k, after which a decode head is added and fine-tuned altogether on a downstream dataset.
|