Add model card and metadata

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +27 -1
README.md CHANGED
@@ -1,5 +1,31 @@
1
  ---
2
  license: mit
 
3
  ---
4
 
5
- This repository will be populated very soon.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ pipeline_tag: image-text-to-text
4
  ---
5
 
6
+ # COLIPRI: Comprehensive Language-Image Pre-training for 3D Medical Image Understanding
7
+
8
+ COLIPRI (Comprehensive Language-Image Pre-training) is a family of vision-language encoders designed for 3D medical image understanding.
9
+
10
+ The model was introduced in the paper: [Comprehensive language-image pre-training for 3D medical image understanding](https://huggingface.co/papers/2510.15042).
11
+
12
+ ## Description
13
+
14
+ COLIPRI aligns 3D medical images with paired text using a multi-objective pre-training strategy. It combines vision-language alignment with a report generation objective and vision-only pre-training, allowing the model to benefit from both image-only and paired image-text 3D datasets.
15
+
16
+ The COLIPRI encoders achieve state-of-the-art performance across various tasks:
17
+ - **Radiological Report Generation**: Generating text descriptions from 3D scans.
18
+ - **Semantic Segmentation**: Downstream adaptation for anatomical and pathological segmentation.
19
+ - **Zero-shot Classification**: Predicting labels without task-specific fine-tuning.
20
+ - **Classification Probing**: Providing strong feature representations for medical diagnostics.
21
+
22
+ ## Citation
23
+
24
+ ```bibtex
25
+ @article{wald2025comprehensive,
26
+ title={Comprehensive language-image pre-training for 3D medical image understanding},
27
+ author={Wald, Tassilo and Hamamci, Ibrahim Ethem and Gao, Yuan and Bond-Taylor, Sam and Sharma, Harshita and Ilse, Maximilian and Lo, Cynthia and Melnichenko, Olesya and Codella, Noel C. F. and Wetscherek, Maria Teodora and others},
28
+ journal={arXiv preprint arXiv:2510.15042},
29
+ year={2025}
30
+ }
31
+ ```