Venn2024 commited on
Commit
486b424
·
verified ·
1 Parent(s): 2eabe7b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -32
README.md CHANGED
@@ -16,15 +16,13 @@ tags:
16
  * **Model Name:** Med-GLIP
17
  * **Paper Title:** Med-GLIP: Advancing Medical Language-Image Pre-training with Large-scale Grounded Dataset
18
  * **Authors:** Ziye Deng, Ruihan He, Jiaxiang Liu, Yuan Wang, Zijie Meng, Songtao Jiang, Yong Xie, Zuozhu Liu
19
- * **Affiliations:** (Not explicitly mentioned in the abstract, but affiliations of the authors would be listed here)
20
  * **Version:** v1
21
  * **Date:** (Presumed August 2025)
22
  * **Model Type:** Medical Language-Image Pre-training Model with Visual Grounding capabilities.
23
  * **Relevant Links:**
24
  * arXiv Page: [https://arxiv.org/abs/2508.10528v1](https://arxiv.org/abs/2508.10528v1)
25
  * DOI: [https://doi.org/10.48550/arXiv.2508.10528](https://doi.org/10.48550/arXiv.2508.10528)
26
- * Code Repository: (Add link if available)
27
- * Model Weights: (Add link if available)
28
  * **License:** Creative Commons Attribution 4.0 International (CC BY 4.0)
29
  * **Citation:**
30
 
@@ -104,32 +102,3 @@ The model aims to overcome the limitations of existing methods in **fine-grained
104
  * **Clinical Risk:** The model is an AI research tool and **must not** be used for primary clinical diagnosis or patient care without explicit, strict clinical validation and regulatory approval. Misinterpretation of results could lead to patient harm.
105
  * **Interpretability:** While the grounding feature aids in interpretability, the overall decision-making process is complex, and failures should be treated with caution.
106
  * (Refer to the full paper for a detailed discussion of ethical and societal implications.)
107
-
108
- ## How to Get Started with the Model
109
-
110
- (If the model code and weights are released, this section provides usage instructions)
111
-
112
- ```python
113
- # Example: Loading the model and processor (assuming compatibility with a library like Hugging Face's transformers)
114
-
115
- # from transformers import AutoProcessor, AutoModelForVisualGrounding
116
- # import torch
117
- # from PIL import Image
118
-
119
- # # Load processor and model
120
- # processor = AutoProcessor.from_pretrained("your-org/med-glip")
121
- # model = AutoModelForVisualGrounding.from_pretrained("your-org/med-glip")
122
-
123
- # # Prepare input
124
- # image = Image.open("path/to/medical_image.jpg")
125
- # text_query = "evidence of right lower lobe consolidation"
126
-
127
- # inputs = processor(images=image, text=text_query, return_tensors="pt")
128
-
129
- # # Perform inference
130
- # with torch.no_grad():
131
- # outputs = model(**inputs)
132
-
133
- # # Process the output to get bounding boxes (implementation details vary)
134
- # predicted_boxes = outputs.logits
135
- # ...
 
16
  * **Model Name:** Med-GLIP
17
  * **Paper Title:** Med-GLIP: Advancing Medical Language-Image Pre-training with Large-scale Grounded Dataset
18
  * **Authors:** Ziye Deng, Ruihan He, Jiaxiang Liu, Yuan Wang, Zijie Meng, Songtao Jiang, Yong Xie, Zuozhu Liu
19
+ * **Affiliations:** Zhejiang University
20
  * **Version:** v1
21
  * **Date:** (Presumed August 2025)
22
  * **Model Type:** Medical Language-Image Pre-training Model with Visual Grounding capabilities.
23
  * **Relevant Links:**
24
  * arXiv Page: [https://arxiv.org/abs/2508.10528v1](https://arxiv.org/abs/2508.10528v1)
25
  * DOI: [https://doi.org/10.48550/arXiv.2508.10528](https://doi.org/10.48550/arXiv.2508.10528)
 
 
26
  * **License:** Creative Commons Attribution 4.0 International (CC BY 4.0)
27
  * **Citation:**
28
 
 
102
  * **Clinical Risk:** The model is an AI research tool and **must not** be used for primary clinical diagnosis or patient care without explicit, strict clinical validation and regulatory approval. Misinterpretation of results could lead to patient harm.
103
  * **Interpretability:** While the grounding feature aids in interpretability, the overall decision-making process is complex, and failures should be treated with caution.
104
  * (Refer to the full paper for a detailed discussion of ethical and societal implications.)