Add pipeline_tag and improve model card metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +7 -5
README.md CHANGED
@@ -1,12 +1,10 @@
1
  ---
2
- license: cc-by-nc-4.0
3
- task_categories:
4
- - image-segmentation
5
  language:
6
  - en
 
 
7
  tags:
8
  - medical-imaging
9
- - image-segmentation
10
  - vision-language-models
11
  - clip
12
  - unimedclip
@@ -23,7 +21,11 @@ tags:
23
  <a href="https://huggingface.co/TahaKoleilat/MedCLIPSeg" target="_blank"><img alt="HuggingFace Models" src="https://img.shields.io/badge/Models-Reproduce-2ea44f?logo=huggingface&logoColor=white" height="25"/></a>
24
  <a href="#citation"><img alt="Citation" src="https://img.shields.io/badge/Citation-BibTeX-6C63FF?logo=bookstack&logoColor=white" height="25"/></a>
25
 
26
- This repository hosts the **official trained model checkpoints** for **MedCLIPSeg**, a vision–language framework for **medical image segmentation** built on top of **CLIP**.
 
 
 
 
27
 
28
  The released checkpoints correspond exactly to the experiments reported in our paper and are provided **for evaluation and reproducibility purposes only**.
29
 
 
1
  ---
 
 
 
2
  language:
3
  - en
4
+ license: cc-by-nc-4.0
5
+ pipeline_tag: image-segmentation
6
  tags:
7
  - medical-imaging
 
8
  - vision-language-models
9
  - clip
10
  - unimedclip
 
21
  <a href="https://huggingface.co/TahaKoleilat/MedCLIPSeg" target="_blank"><img alt="HuggingFace Models" src="https://img.shields.io/badge/Models-Reproduce-2ea44f?logo=huggingface&logoColor=white" height="25"/></a>
22
  <a href="#citation"><img alt="Citation" src="https://img.shields.io/badge/Citation-BibTeX-6C63FF?logo=bookstack&logoColor=white" height="25"/></a>
23
 
24
+ This repository hosts the **official trained model checkpoints** for **MedCLIPSeg**, presented in the paper [MedCLIPSeg: Probabilistic Vision-Language Adaptation for Data-Efficient and Generalizable Medical Image Segmentation](https://huggingface.co/papers/2602.20423).
25
+
26
+ **Authors:** Taha Koleilat, Hojat Asgariandehkordi, Omid Nejati Manzari, Berardino Barile, Yiming Xiao, Hassan Rivaz.
27
+
28
+ MedCLIPSeg is a vision–language framework for **medical image segmentation** built on top of **CLIP**. It adapts CLIP for robust, data-efficient, and uncertainty-aware segmentation through probabilistic cross-modal attention and bidirectional interaction between image and text tokens.
29
 
30
  The released checkpoints correspond exactly to the experiments reported in our paper and are provided **for evaluation and reproducibility purposes only**.
31