Upload MedVisionAI best model (epoch 100) with benchmark results
Browse files- README.md +44 -55
- config.json +1 -1
- figures/architecture.png +1 -0
- figures/badge.png +1 -0
- figures/performance.png +1 -0
- pytorch_model.bin +0 -0
README.md
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
-
library_name:
|
| 4 |
---
|
| 5 |
# MedVisionAI
|
| 6 |
<!-- markdownlint-disable first-line-h1 -->
|
|
@@ -8,27 +8,27 @@ library_name: timm
|
|
| 8 |
<!-- markdownlint-disable no-duplicate-header -->
|
| 9 |
|
| 10 |
<div align="center">
|
| 11 |
-
<img src="figures/
|
| 12 |
</div>
|
| 13 |
<hr>
|
| 14 |
|
| 15 |
<div align="center" style="line-height: 1;">
|
| 16 |
<a href="LICENSE" style="margin: 2px;">
|
| 17 |
-
<img alt="License" src="figures/
|
| 18 |
</a>
|
| 19 |
</div>
|
| 20 |
|
| 21 |
## 1. Introduction
|
| 22 |
|
| 23 |
-
MedVisionAI
|
| 24 |
|
| 25 |
<p align="center">
|
| 26 |
-
<img width="80%" src="figures/
|
| 27 |
</p>
|
| 28 |
|
| 29 |
-
|
| 30 |
|
| 31 |
-
|
| 32 |
|
| 33 |
## 2. Evaluation Results
|
| 34 |
|
|
@@ -36,70 +36,59 @@ Beyond diagnostic accuracy improvements, this version also offers enhanced expla
|
|
| 36 |
|
| 37 |
<div align="center">
|
| 38 |
|
| 39 |
-
| | Benchmark |
|
| 40 |
|---|---|---|---|---|---|
|
| 41 |
-
| **
|
| 42 |
-
| |
|
| 43 |
-
| |
|
| 44 |
-
| **
|
| 45 |
-
| |
|
| 46 |
-
| |
|
| 47 |
-
| |
|
| 48 |
-
|
|
| 49 |
-
| |
|
| 50 |
-
|
|
| 51 |
-
| |
|
| 52 |
-
|
|
| 53 |
-
| | Brain Anomaly | 0.681 | 0.697 | 0.709 | 0.665 |
|
| 54 |
-
| | Lung Nodule | 0.745 | 0.761 | 0.773 | 0.782 |
|
| 55 |
-
| | Safety Compliance | 0.812 | 0.801 | 0.825 | 0.800 |
|
| 56 |
|
| 57 |
</div>
|
| 58 |
|
| 59 |
### Overall Performance Summary
|
| 60 |
-
MedVisionAI demonstrates strong
|
| 61 |
|
| 62 |
-
## 3. Clinical Integration & API
|
| 63 |
-
We
|
| 64 |
|
| 65 |
## 4. How to Run Locally
|
| 66 |
|
| 67 |
-
Please refer to our
|
| 68 |
|
| 69 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 70 |
|
| 71 |
-
|
| 72 |
-
2. Multi-GPU inference is available for high-resolution 3D volume processing.
|
| 73 |
-
|
| 74 |
-
The model architecture of MedVisionAI-Lite is optimized for edge deployment, sharing the same preprocessing pipeline as the main MedVisionAI model.
|
| 75 |
-
|
| 76 |
-
### Preprocessing Configuration
|
| 77 |
-
We recommend using the following preprocessing configuration for optimal results.
|
| 78 |
```python
|
| 79 |
-
|
| 80 |
-
"target_spacing": [1.0, 1.0, 1.0], # mm
|
| 81 |
-
"intensity_normalization": "z-score",
|
| 82 |
-
"window_center": 40,
|
| 83 |
-
"window_width": 400
|
| 84 |
-
}
|
| 85 |
-
```
|
| 86 |
|
| 87 |
-
|
| 88 |
-
|
| 89 |
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
|
| 93 |
-
Required DICOM Tags:
|
| 94 |
-
- PatientID (anonymized)
|
| 95 |
-
- StudyDate
|
| 96 |
-
- Modality
|
| 97 |
-
- PixelSpacing
|
| 98 |
-
- SliceThickness (for 3D volumes)
|
| 99 |
```
|
| 100 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 101 |
## 5. License
|
| 102 |
-
This
|
| 103 |
|
| 104 |
## 6. Contact
|
| 105 |
-
|
|
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
library_name: transformers
|
| 4 |
---
|
| 5 |
# MedVisionAI
|
| 6 |
<!-- markdownlint-disable first-line-h1 -->
|
|
|
|
| 8 |
<!-- markdownlint-disable no-duplicate-header -->
|
| 9 |
|
| 10 |
<div align="center">
|
| 11 |
+
<img src="figures/architecture.png" width="60%" alt="MedVisionAI" />
|
| 12 |
</div>
|
| 13 |
<hr>
|
| 14 |
|
| 15 |
<div align="center" style="line-height: 1;">
|
| 16 |
<a href="LICENSE" style="margin: 2px;">
|
| 17 |
+
<img alt="License" src="figures/badge.png" style="display: inline-block; vertical-align: middle;"/>
|
| 18 |
</a>
|
| 19 |
</div>
|
| 20 |
|
| 21 |
## 1. Introduction
|
| 22 |
|
| 23 |
+
MedVisionAI is a state-of-the-art medical imaging foundation model designed for multi-modal diagnostic analysis. In this latest release, MedVisionAI has been enhanced with advanced attention mechanisms and domain-specific pre-training on over 5 million medical images spanning radiology, pathology, and ophthalmology datasets.
|
| 24 |
|
| 25 |
<p align="center">
|
| 26 |
+
<img width="80%" src="figures/performance.png">
|
| 27 |
</p>
|
| 28 |
|
| 29 |
+
The model demonstrates exceptional performance across diverse medical imaging tasks. On the CheXpert benchmark, MedVisionAI achieves an AUC of 0.932, significantly outperforming previous approaches. For MRI brain tumor segmentation, the model attains a Dice score of 0.891.
|
| 30 |
|
| 31 |
+
MedVisionAI is designed for research and clinical decision support. It should always be used under medical professional supervision and is not intended as a standalone diagnostic tool.
|
| 32 |
|
| 33 |
## 2. Evaluation Results
|
| 34 |
|
|
|
|
| 36 |
|
| 37 |
<div align="center">
|
| 38 |
|
| 39 |
+
| | Benchmark | ResNet-Medical | DenseNet-Med | ViT-Medical | MedVisionAI |
|
| 40 |
|---|---|---|---|---|---|
|
| 41 |
+
| **Radiology** | X-Ray Classification | 0.845 | 0.862 | 0.878 | 0.906 |
|
| 42 |
+
| | CT Detection | 0.812 | 0.831 | 0.849 | 0.896 |
|
| 43 |
+
| | MRI Segmentation | 0.756 | 0.778 | 0.801 | 0.890 |
|
| 44 |
+
| **Pathology** | Pathology Analysis | 0.823 | 0.841 | 0.859 | 0.875 |
|
| 45 |
+
| | Tumor Localization | 0.789 | 0.804 | 0.821 | 0.861 |
|
| 46 |
+
| | Abnormality Detection | 0.834 | 0.852 | 0.867 | 0.884 |
|
| 47 |
+
| **Screening** | Mammography Screening | 0.867 | 0.881 | 0.894 | 0.913 |
|
| 48 |
+
| | Skin Lesion Classification | 0.801 | 0.819 | 0.837 | 0.864 |
|
| 49 |
+
| | Retinal Diagnosis | 0.778 | 0.795 | 0.812 | 0.852 |
|
| 50 |
+
| **Specialized**| Ultrasound Detection | 0.745 | 0.762 | 0.779 | 0.820 |
|
| 51 |
+
| | Bone Fracture Detection | 0.856 | 0.871 | 0.886 | 0.895 |
|
| 52 |
+
| | Cardiac Imaging | 0.723 | 0.741 | 0.759 | 0.828 |
|
|
|
|
|
|
|
|
|
|
| 53 |
|
| 54 |
</div>
|
| 55 |
|
| 56 |
### Overall Performance Summary
|
| 57 |
+
MedVisionAI demonstrates strong diagnostic accuracy across all evaluated medical imaging modalities, with particularly notable results in radiology and screening applications.
|
| 58 |
|
| 59 |
+
## 3. Clinical Integration & API
|
| 60 |
+
We provide a HIPAA-compliant API for clinical integration. Please contact our team for enterprise deployment options.
|
| 61 |
|
| 62 |
## 4. How to Run Locally
|
| 63 |
|
| 64 |
+
Please refer to our documentation for running MedVisionAI in your local environment.
|
| 65 |
|
| 66 |
+
### Input Format
|
| 67 |
+
The model accepts DICOM, PNG, and JPEG image formats. For optimal performance, ensure images are:
|
| 68 |
+
- Resolution: minimum 224x224 pixels
|
| 69 |
+
- Bit depth: 8-bit or 16-bit grayscale for radiology
|
| 70 |
+
- Color space: RGB for pathology and dermoscopy
|
| 71 |
|
| 72 |
+
### Model Configuration
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 73 |
```python
|
| 74 |
+
from transformers import AutoModel, AutoProcessor
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 75 |
|
| 76 |
+
model = AutoModel.from_pretrained("MedVisionAI")
|
| 77 |
+
processor = AutoProcessor.from_pretrained("MedVisionAI")
|
| 78 |
|
| 79 |
+
# For X-ray classification
|
| 80 |
+
inputs = processor(images=xray_image, return_tensors="pt")
|
| 81 |
+
outputs = model(**inputs)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 82 |
```
|
| 83 |
|
| 84 |
+
### Recommended Settings
|
| 85 |
+
- Batch size: 16 for inference
|
| 86 |
+
- Mixed precision: FP16 recommended
|
| 87 |
+
- Input normalization: ImageNet statistics
|
| 88 |
+
|
| 89 |
## 5. License
|
| 90 |
+
This model is licensed under the [Apache 2.0 License](LICENSE). Use in clinical settings requires appropriate regulatory approval in your jurisdiction.
|
| 91 |
|
| 92 |
## 6. Contact
|
| 93 |
+
For research collaborations: research@medvisionai.health
|
| 94 |
+
For clinical inquiries: clinical@medvisionai.health
|
config.json
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
{
|
| 2 |
"model_type": "vit",
|
| 3 |
"architectures": ["ViTForImageClassification"]
|
| 4 |
-
|
|
|
|
| 1 |
{
|
| 2 |
"model_type": "vit",
|
| 3 |
"architectures": ["ViTForImageClassification"]
|
| 4 |
+
}
|
figures/architecture.png
ADDED
|
figures/badge.png
ADDED
|
figures/performance.png
ADDED
|
pytorch_model.bin
CHANGED
|
Binary files a/pytorch_model.bin and b/pytorch_model.bin differ
|
|
|