Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,26 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: apache-2.0
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
+
language:
|
| 4 |
+
- en
|
| 5 |
+
pipeline_tag: image-to-text
|
| 6 |
+
tags:
|
| 7 |
+
- medical
|
| 8 |
+
---
|
| 9 |
+
|
| 10 |
+
# Model Card for PathBLIP-2
|
| 11 |
+
|
| 12 |
+
A vision-language model built upon the BLIP-2 framework in combination with BioGPT and HIPT for pathology report generation and cross-modal retrieval of melanocytic lesions.
|
| 13 |
+
|
| 14 |
+
## Model Details
|
| 15 |
+
|
| 16 |
+
This repository contains multiple checkpoints for the models which were used for the experiments in the paper.
|
| 17 |
+
The models were trained on a dataset of 19,636 melanocytic lesion cases, consisting of one or more whole slide images (WSIs) and a pathology report, using different training configurations.
|
| 18 |
+
The supporting code is available from the corresponding GitHub repository.
|
| 19 |
+
We refer to the paper for more information regarding the dataset, finetuning, evaluation, and limitations.
|
| 20 |
+
|
| 21 |
+
- **Paper: *["On the Importance of Text Preprocessing for Multimodal Representation Learning and Pathology Report Generation"](https://arxiv.org/abs/2502.19285)***
|
| 22 |
+
- **Repository:** [GitHub](https://github.com/RTLucassen/report_preprocessing)
|
| 23 |
+
- **Framework:** [BLIP-2](https://huggingface.co/Salesforce/blip2-opt-2.7b)
|
| 24 |
+
- **Base language model:** [BioGPT](https://huggingface.co/microsoft/biogpt)
|
| 25 |
+
- **WSI feature extractor:** [HIPT](https://github.com/mahmoodlab/HIPT)
|
| 26 |
+
- **License:** Apache-2.0
|