toolevalxm commited on
Commit
0fce5fb
·
verified ·
1 Parent(s): bb4ee03

Upload folder using huggingface_hub

Browse files
Files changed (6) hide show
  1. README.md +43 -34
  2. config.json +5 -3
  3. figures/fig1.png +0 -0
  4. figures/fig2.png +0 -0
  5. figures/fig3.png +0 -0
  6. pytorch_model.bin +2 -2
README.md CHANGED
@@ -8,27 +8,27 @@ library_name: transformers
8
  <!-- markdownlint-disable no-duplicate-header -->
9
 
10
  <div align="center">
11
- <img src="figures/architecture.png" width="60%" alt="MedVisionNet" />
12
  </div>
13
  <hr>
14
 
15
  <div align="center" style="line-height: 1;">
16
  <a href="LICENSE" style="margin: 2px;">
17
- <img alt="License" src="figures/badge.png" style="display: inline-block; vertical-align: middle;"/>
18
  </a>
19
  </div>
20
 
21
  ## 1. Introduction
22
 
23
- MedVisionNet is a state-of-the-art medical imaging AI model designed for clinical diagnosis assistance. The model has been trained on over 2 million anonymized medical images from diverse sources including CT scans, MRI, X-rays, and ultrasound imaging. MedVisionNet achieves exceptional performance across multiple medical imaging tasks including tumor detection, organ segmentation, and fracture identification.
24
 
25
  <p align="center">
26
- <img width="80%" src="figures/performance.png">
27
  </p>
28
 
29
- The model leverages a novel attention mechanism specifically designed for medical image analysis, allowing it to focus on clinically relevant regions while maintaining spatial context. This architecture enables the model to provide accurate predictions with explainable attention maps that can assist radiologists in their diagnostic workflow.
30
 
31
- MedVisionNet has been validated against multiple clinical datasets and has demonstrated performance comparable to or exceeding that of experienced radiologists in specific tasks.
32
 
33
  ## 2. Evaluation Results
34
 
@@ -36,46 +36,55 @@ MedVisionNet has been validated against multiple clinical datasets and has demon
36
 
37
  <div align="center">
38
 
39
- | Benchmark | BaselineNet | RadioNet | MedViT | MedVisionNet |
40
- |---|---|---|---|---|
41
- | Tumor Detection (AUC) | 0.871 | 0.889 | 0.902 | 0.871 |
42
- | Organ Segmentation (Dice) | 0.823 | 0.841 | 0.856 | 0.863 |
43
- | Fracture Detection (Sensitivity) | 0.792 | 0.815 | 0.834 | 0.857 |
44
- | Nodule Classification (F1) | 0.756 | 0.778 | 0.801 | 0.794 |
45
- | Vessel Segmentation (IoU) | 0.689 | 0.721 | 0.745 | 0.739 |
 
 
 
 
 
 
 
 
 
 
46
 
47
  </div>
48
 
49
  ### Overall Performance Summary
50
- MedVisionNet demonstrates superior performance across all evaluated medical imaging benchmarks, with particularly strong results in tumor detection and organ segmentation tasks.
51
 
52
- ## 3. Clinical Validation
53
- The model has undergone rigorous clinical validation with board-certified radiologists. Multi-center trials have shown consistent performance across different scanner types and patient demographics.
54
 
55
- ## 4. How to Use
56
 
57
- ```python
58
- from transformers import AutoModel, AutoImageProcessor
59
 
60
- processor = AutoImageProcessor.from_pretrained("hospital-ai/MedVisionNet")
61
- model = AutoModel.from_pretrained("hospital-ai/MedVisionNet")
 
 
 
 
 
62
 
63
- # Load your medical image
64
- import PIL.Image
65
- image = PIL.Image.open("chest_xray.png")
 
66
 
67
- inputs = processor(images=image, return_tensors="pt")
68
- outputs = model(**inputs)
69
  ```
70
 
71
- ### Recommended Settings
72
- - Image resolution: 512x512 for optimal performance
73
- - Preprocessing: DICOM standardization recommended
74
- - Inference: Batch size of 1 for production use
75
-
76
  ## 5. License
77
- This model is released under the Apache 2.0 License. Clinical deployment requires appropriate regulatory approval.
78
 
79
  ## 6. Contact
80
- For research collaborations: research@medvisionnet.ai
81
- For clinical inquiries: clinical@medvisionnet.ai
 
8
  <!-- markdownlint-disable no-duplicate-header -->
9
 
10
  <div align="center">
11
+ <img src="figures/fig1.png" width="60%" alt="MedVisionNet" />
12
  </div>
13
  <hr>
14
 
15
  <div align="center" style="line-height: 1;">
16
  <a href="LICENSE" style="margin: 2px;">
17
+ <img alt="License" src="figures/fig2.png" style="display: inline-block; vertical-align: middle;"/>
18
  </a>
19
  </div>
20
 
21
  ## 1. Introduction
22
 
23
+ MedVisionNet represents a breakthrough in medical imaging analysis. This latest version incorporates advanced convolutional architectures with attention mechanisms, enabling superior performance in detecting and classifying medical conditions from radiological scans. The model has been extensively validated across multiple medical imaging modalities including X-rays, CT scans, and MRI.
24
 
25
  <p align="center">
26
+ <img width="80%" src="figures/fig3.png">
27
  </p>
28
 
29
+ Compared to the previous version, MedVisionNet shows remarkable improvements in sensitivity and specificity metrics. For instance, in the ChestX-ray14 benchmark, the model's AUC-ROC has increased from 0.82 in the previous version to 0.94 in the current version. This improvement stems from enhanced feature extraction capabilities during training.
30
 
31
+ Beyond its improved diagnostic capabilities, this version also offers reduced false positive rates and enhanced support for multi-label classification.
32
 
33
  ## 2. Evaluation Results
34
 
 
36
 
37
  <div align="center">
38
 
39
+ | | Benchmark | Baseline | ResNet50 | DenseNet | MedVisionNet |
40
+ |---|---|---|---|---|---|
41
+ | **Radiology Tasks** | Chest X-ray Classification | 0.820 | 0.845 | 0.861 | 0.941 |
42
+ | | Lung Nodule Detection | 0.765 | 0.788 | 0.801 | 0.883 |
43
+ | | Pneumonia Detection | 0.801 | 0.823 | 0.835 | 0.927 |
44
+ | **CT Analysis** | Liver Segmentation | 0.712 | 0.745 | 0.758 | 0.858 |
45
+ | | Tumor Detection | 0.689 | 0.721 | 0.739 | 0.865 |
46
+ | | Organ Boundary | 0.756 | 0.778 | 0.791 | 0.873 |
47
+ | | Lesion Localization | 0.698 | 0.725 | 0.742 | 0.865 |
48
+ | **MRI Tasks** | Brain Tumor Segmentation | 0.801 | 0.832 | 0.847 | 0.912 |
49
+ | | White Matter Analysis | 0.734 | 0.761 | 0.778 | 0.888 |
50
+ | | Cardiac Function | 0.721 | 0.748 | 0.763 | 0.867 |
51
+ | | Spine Assessment | 0.689 | 0.712 | 0.728 | 0.867 |
52
+ | **Ultrasound**| Fetal Measurement | 0.778 | 0.801 | 0.815 | 0.903 |
53
+ | | Thyroid Nodule | 0.701 | 0.728 | 0.745 | 0.877 |
54
+ | | Cardiac Echo | 0.745 | 0.771 | 0.788 | 0.880 |
55
+ | | Abdominal Scan | 0.712 | 0.738 | 0.752 | 0.878 |
56
 
57
  </div>
58
 
59
  ### Overall Performance Summary
60
+ MedVisionNet demonstrates superior performance across all evaluated medical imaging categories, with particularly notable results in radiology and MRI analysis tasks.
61
 
62
+ ## 3. Clinical Integration
63
+ We offer DICOM integration and HL7 FHIR compatibility for seamless hospital workflow integration. Please check our official documentation for deployment guidelines.
64
 
65
+ ## 4. How to Run Locally
66
 
67
+ Please refer to our code repository for more information about running MedVisionNet locally.
 
68
 
69
+ ### Recommended Configuration
70
+ We recommend using the following settings for optimal performance:
71
+ ```
72
+ batch_size: 8
73
+ image_size: 512x512
74
+ preprocessing: histogram_equalization
75
+ ```
76
 
77
+ ### Input Format
78
+ For medical image analysis, please follow the template:
79
+ ```python
80
+ from medvisionnet import MedVisionNet
81
 
82
+ model = MedVisionNet.from_pretrained("medvisionnet-v2")
83
+ prediction = model.analyze(dicom_path="path/to/image.dcm")
84
  ```
85
 
 
 
 
 
 
86
  ## 5. License
87
+ This code repository is licensed under the [Apache 2.0 License](LICENSE). The use of MedVisionNet models is also subject to the [Apache 2.0 License](LICENSE).
88
 
89
  ## 6. Contact
90
+ If you have any questions, please raise an issue on our GitHub repository or contact us at support@medvisionnet.health.
 
config.json CHANGED
@@ -1,6 +1,8 @@
1
  {
2
  "model_type": "vit",
3
- "architectures": ["ViTForImageClassification"],
4
- "num_labels": 5,
5
- "image_size": 512
 
 
6
  }
 
1
  {
2
  "model_type": "vit",
3
+ "architectures": ["VisionTransformer"],
4
+ "image_size": 512,
5
+ "patch_size": 16,
6
+ "num_channels": 3,
7
+ "hidden_size": 768
8
  }
figures/fig1.png CHANGED
figures/fig2.png CHANGED
figures/fig3.png CHANGED
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52b33af4763decfef25880d740367f1c500bc2aeb669dcaea7e9776da2a85b27
3
- size 123
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55ba003bb441780b6490fb6d29261615d4a04ff03907e08ece393240c3d853bf
3
+ size 1049