cwangrun commited on
Commit
83af716
·
verified ·
1 Parent(s): 0d50af0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -16
README.md CHANGED
@@ -8,23 +8,39 @@ base_model:
8
  pipeline_tag: zero-shot-image-classification
9
  tags:
10
  - medical
 
 
 
 
 
 
 
 
 
11
  ---
12
  # CheXficient
13
 
14
- CheXficient is a vision-language foundation model for chest X-ray (CXR) interpretation, developed to enhance both data- and computation-efficiency.
15
- It enables joint image-text representation learning and supports prompt-based zero-shot classification.
 
 
 
 
 
 
16
 
17
- This repository provides a Hugging Face-compatible implementation for seamless integration into research workflows.
18
 
19
  ------------------------------------------------------------------------
20
 
21
  ## Model Overview
22
 
23
- - Architecture: Vision-Language dual encoder
24
- - Input: Chest X-ray image + text prompts
25
- - Output: Image-text similarity logits and embeddings
26
- - Framework: PyTorch + Hugging Face Transformers
27
- - Intended Use: Research in medical AI and multimodal learning
 
 
28
 
29
  ------------------------------------------------------------------------
30
 
@@ -88,21 +104,15 @@ print(probs)
88
  ```
89
 
90
 
91
- ## Intended Use
92
-
93
- - Zero-shot CXR findings classification
94
- - Prompt-based disease detection
95
-
96
-
97
  ------------------------------------------------------------------------
98
 
99
  ## Citation
100
 
101
  ``` bibtex
102
  @article{chexficient2024,
103
- title={CheXficient: Efficient Vision-Language Learning for Chest X-ray Understanding},
104
  author={...},
105
  journal={...},
106
- year={2024}
107
  }
108
  ```
 
8
  pipeline_tag: zero-shot-image-classification
9
  tags:
10
  - medical
11
+ datasets:
12
+ - simwit/mimic-cxr
13
+ - danjacobellis/chexpert
14
+ - rajpurkarlab/ReXGradient-160K
15
+ - BahaaEldin0/NIH-Chest-Xray-14
16
+ - SampadKar/vindr-cxr
17
+ metrics:
18
+ - accuracy
19
+ - bleu
20
  ---
21
  # CheXficient
22
 
23
+ [Paper](https://arxiv.org/abs/2602.22843) | [GitHub](https://github.com/cwangrun/CheXficient)
24
+
25
+ CheXficient is a vision-language foundation model for chest X-ray (CXR) interpretation, designed to improve both **data efficiency** and **computational efficiency** during pretraining.
26
+
27
+ Instead of scaling indiscriminately to ever-larger datasets, CheXficient adopts a principled data curation strategy to selectively prioritize informative training samples.
28
+ This approach demonstrates that active, structured data selection can serve as a cost-effective alternative to brute-force dataset enlargement.
29
+
30
+ The model follows a dual-encoder architecture and supports prompt-based zero-shot classification via joint image-text representation learning.
31
 
 
32
 
33
  ------------------------------------------------------------------------
34
 
35
  ## Model Overview
36
 
37
+ - **Architecture:** Vision-language dual encoder
38
+ - **Image Backbone:** DINOv2 (base)
39
+ - **Text Backbone:** BioClinicalBERT
40
+ - **Input:** Chest X-ray image + text prompts
41
+ - **Output:** Image-text similarity logits and embeddings
42
+ - **Framework:** PyTorch + Hugging Face Transformers
43
+ - **Intended Use:** Research in medical AI and multimodal learning
44
 
45
  ------------------------------------------------------------------------
46
 
 
104
  ```
105
 
106
 
 
 
 
 
 
 
107
  ------------------------------------------------------------------------
108
 
109
  ## Citation
110
 
111
  ``` bibtex
112
  @article{chexficient2024,
113
+ title={A data- and compute-efficient chest X-ray foundation model beyond aggressive scaling},
114
  author={...},
115
  journal={...},
116
+ year={2026}
117
  }
118
  ```