toolevalxm commited on
Commit
430ae43
·
verified ·
1 Parent(s): abe1f9b

Upload folder using huggingface_hub

Browse files
Files changed (6) hide show
  1. README.md +52 -57
  2. config.json +8 -17
  3. figures/fig1.png +0 -0
  4. figures/fig2.png +0 -0
  5. figures/fig3.png +0 -0
  6. pytorch_model.bin +2 -2
README.md CHANGED
@@ -20,15 +20,15 @@ library_name: transformers
20
 
21
  ## 1. Introduction
22
 
23
- MedVisionNet represents a breakthrough in medical image analysis using vision transformers. This latest version incorporates advanced attention mechanisms specifically designed for radiological image interpretation. The model demonstrates state-of-the-art performance across multiple medical imaging modalities including X-ray, CT, MRI, and ultrasound.
24
 
25
  <p align="center">
26
  <img width="80%" src="figures/fig3.png">
27
  </p>
28
 
29
- Compared to the previous version, MedVisionNet v2 shows remarkable improvements in segmentation tasks. In the BraTS 2024 challenge, the model's mean Dice score improved from 0.82 in the previous version to 0.91 in the current version. This advancement stems from the multi-scale feature pyramid network architecture: the previous model processed images at a single resolution, whereas the new version analyzes at 4 different scales simultaneously.
30
 
31
- Beyond its improved segmentation capabilities, this version also provides uncertainty quantification and enhanced interpretability through attention map visualization.
32
 
33
  ## 2. Evaluation Results
34
 
@@ -36,81 +36,76 @@ Beyond its improved segmentation capabilities, this version also provides uncert
36
 
37
  <div align="center">
38
 
39
- | | Benchmark | ModelA | ModelB | ModelA-v2 | MedVisionNet |
40
  |---|---|---|---|---|---|
41
- | **Segmentation Tasks** | Tumor Segmentation | 0.823 | 0.841 | 0.855 | 0.856 |
42
- | | Organ Detection | 0.891 | 0.902 | 0.911 | 0.909 |
43
- | | Brain MRI Segmentation | 0.812 | 0.825 | 0.838 | 0.836 |
44
- | **Classification Tasks** | Lesion Classification | 0.756 | 0.771 | 0.785 | 0.779 |
45
- | | Skin Lesion Classification | 0.834 | 0.848 | 0.862 | 0.856 |
46
- | | Pathology Classification | 0.789 | 0.802 | 0.815 | 0.809 |
47
- | | Chest X-ray Analysis | 0.867 | 0.879 | 0.891 | 0.886 |
48
- | **Detection Tasks** | Bone Fracture Detection | 0.723 | 0.738 | 0.752 | 0.749 |
49
- | | Mammography Detection | 0.801 | 0.815 | 0.828 | 0.829 |
50
- | | Dental X-ray Analysis | 0.778 | 0.792 | 0.805 | 0.799 |
51
- | | Cardiac Imaging | 0.845 | 0.858 | 0.871 | 0.869 |
52
- | **Advanced Analysis** | CT Scan Analysis | 0.856 | 0.869 | 0.882 | 0.879 |
53
- | | Ultrasound Segmentation | 0.734 | 0.748 | 0.761 | 0.756 |
54
- | | Retinal Screening | 0.812 | 0.826 | 0.839 | 0.836 |
55
- | | Spine Alignment | 0.767 | 0.781 | 0.794 | 0.789 |
56
 
57
  </div>
58
 
59
  ### Overall Performance Summary
60
- MedVisionNet demonstrates exceptional performance across all evaluated medical imaging benchmarks, with particularly strong results in segmentation and classification tasks.
61
 
62
- ## 3. Clinical Integration & API Platform
63
- We provide a HIPAA-compliant API for clinical integration. Please contact our enterprise team for deployment options.
64
 
65
  ## 4. How to Run Locally
66
 
67
- Please refer to our code repository for more information about running MedVisionNet locally.
68
 
69
- Key usage recommendations for MedVisionNet:
70
 
71
- 1. Image preprocessing with DICOM standardization is supported.
72
- 2. Batch inference is optimized for throughput in clinical workflows.
73
 
74
- The model architecture of MedVisionNet-Lite is identical to its base model, optimized for edge deployment in medical devices.
75
 
76
- ### Input Specifications
77
- Images should be preprocessed to the following specifications:
78
- ```
79
- input_size = (512, 512)
80
- normalization = "medical_standard" # Uses Hounsfield units for CT
81
- channels = 1 # grayscale for most modalities
82
- ```
83
-
84
- ### Inference Configuration
85
- We recommend the following configuration for optimal results:
86
  ```python
87
- config = {
88
- "confidence_threshold": 0.85,
89
- "nms_threshold": 0.5,
90
- "use_tta": True, # Test-time augmentation
91
- "ensemble_mode": "weighted_average"
92
  }
93
  ```
94
 
95
- ### DICOM Integration
96
- For DICOM file processing, use our provided utilities:
97
- ```python
98
- from medvisionnet import DicomProcessor
99
-
100
- processor = DicomProcessor()
101
- image = processor.load_dicom("path/to/study.dcm")
102
- prediction = model.predict(image)
103
- ```
104
 
105
- For multi-slice CT analysis, we recommend the following approach:
 
106
  ```python
107
- ct_processor = DicomProcessor(modality="CT")
108
- volume = ct_processor.load_series("path/to/ct_series/")
109
- predictions = model.predict_volume(volume, slice_thickness=1.0)
 
 
 
 
 
 
 
 
 
110
  ```
111
 
112
  ## 5. License
113
- This code repository is licensed under the [Apache 2.0 License](LICENSE). The use of MedVisionNet models is also subject to the [Apache 2.0 License](LICENSE). The model is approved for research use; clinical deployment requires additional validation.
114
 
115
  ## 6. Contact
116
- If you have any questions, please raise an issue on our GitHub repository or contact us at medical-ai@medvisionnet.ai.
 
20
 
21
  ## 1. Introduction
22
 
23
+ MedVisionNet is a state-of-the-art medical imaging AI model designed for clinical diagnostics and research applications. In the latest release, MedVisionNet has achieved significant improvements in detection accuracy and segmentation precision through advanced training techniques and multi-modal learning capabilities. The model demonstrates exceptional performance across various medical imaging benchmarks, including radiology, pathology, and nuclear medicine applications.
24
 
25
  <p align="center">
26
  <img width="80%" src="figures/fig3.png">
27
  </p>
28
 
29
+ Compared to the previous version, MedVisionNet shows remarkable improvements in handling complex diagnostic scenarios. For instance, in the RSNA 2024 chest X-ray challenge, the model's sensitivity increased from 82.3% in the previous version to 91.7% in the current release. This advancement stems from enhanced feature extraction during the encoding process: the previous model processed images at 512x512 resolution, whereas the new version operates at 1024x1024 with hierarchical attention mechanisms.
30
 
31
+ Beyond its improved diagnostic capabilities, this version also offers reduced false positive rates and enhanced support for multi-organ analysis.
32
 
33
  ## 2. Evaluation Results
34
 
 
36
 
37
  <div align="center">
38
 
39
+ | | Benchmark | RadNet-v1 | ScanAI | DiagnosticPro | MedVisionNet |
40
  |---|---|---|---|---|---|
41
+ | **Detection Tasks** | Tumor Detection | 0.845 | 0.862 | 0.871 | 0.864 |
42
+ | | Nodule Detection | 0.812 | 0.834 | 0.841 | 0.846 |
43
+ | | Bone Fracture | 0.789 | 0.803 | 0.815 | 0.853 |
44
+ | **Segmentation Tasks** | Organ Segmentation | 0.891 | 0.902 | 0.911 | 0.908 |
45
+ | | Vessel Segmentation | 0.823 | 0.841 | 0.852 | 0.789 |
46
+ | | Lesion Classification | 0.856 | 0.871 | 0.879 | 0.884 |
47
+ | | Tissue Analysis | 0.801 | 0.819 | 0.828 | 0.800 |
48
+ | **Advanced Analysis** | Anomaly Detection | 0.778 | 0.792 | 0.805 | 0.740 |
49
+ | | Pathology Grading | 0.834 | 0.848 | 0.859 | 0.829 |
50
+ | | 3D Reconstruction | 0.756 | 0.778 | 0.791 | 0.817 |
51
+ | | Registration Accuracy | 0.812 | 0.829 | 0.838 | 0.826 |
52
+ | **Image Processing** | Image Quality | 0.867 | 0.881 | 0.889 | 0.918 |
53
+ | | Contrast Enhancement | 0.723 | 0.745 | 0.759 | 0.733 |
54
+ | | Artifact Removal | 0.698 | 0.721 | 0.734 | 0.655 |
55
+ | | Dose Estimation | 0.845 | 0.859 | 0.867 | 0.843 |
56
 
57
  </div>
58
 
59
  ### Overall Performance Summary
60
+ MedVisionNet demonstrates superior performance across all evaluated medical imaging benchmark categories, with particularly notable results in detection and segmentation tasks.
61
 
62
+ ## 3. Clinical Platform & API Access
63
+ We offer a clinical validation interface and secure API for healthcare providers to integrate MedVisionNet. Please check our official website for HIPAA-compliant deployment options.
64
 
65
  ## 4. How to Run Locally
66
 
67
+ Please refer to our clinical deployment guide for more information about running MedVisionNet in clinical environments.
68
 
69
+ Compared to previous versions, the deployment recommendations for MedVisionNet have the following changes:
70
 
71
+ 1. GPU memory requirement increased to 16GB for full resolution inference.
72
+ 2. DICOM preprocessing pipeline is now integrated into the model.
73
 
74
+ The model architecture of MedVisionNet-Lite is optimized for edge deployment while maintaining diagnostic accuracy suitable for screening applications.
75
 
76
+ ### Input Preprocessing
77
+ We recommend using the following preprocessing pipeline:
 
 
 
 
 
 
 
 
78
  ```python
79
+ preprocessing_config = {
80
+ "target_spacing": [1.0, 1.0, 1.0], # mm
81
+ "intensity_normalization": "z-score",
82
+ "window_center": 40,
83
+ "window_width": 400
84
  }
85
  ```
86
 
87
+ ### Inference Configuration
88
+ We recommend setting the confidence threshold $T_{conf}$ to 0.75 for clinical screening.
 
 
 
 
 
 
 
89
 
90
+ ### Multi-Modal Input Template
91
+ For combined imaging studies, use the following input structure:
92
  ```python
93
+ study_template = {
94
+ "modality": "{modality_type}",
95
+ "series": [
96
+ {"path": "{dicom_path_1}", "description": "{series_desc_1}"},
97
+ {"path": "{dicom_path_2}", "description": "{series_desc_2}"}
98
+ ],
99
+ "patient_info": {
100
+ "age": "{age}",
101
+ "gender": "{gender}",
102
+ "clinical_history": "{history}"
103
+ }
104
+ }
105
  ```
106
 
107
  ## 5. License
108
+ This model is licensed under the [Apache 2.0 License](LICENSE). Clinical deployment requires additional validation per local regulatory requirements. The model supports research use and clinical validation studies.
109
 
110
  ## 6. Contact
111
+ If you have any questions, please raise an issue on our GitHub repository or contact us at support@medvisionnet.ai.
config.json CHANGED
@@ -1,19 +1,10 @@
1
  {
2
- "model_type": "vit",
3
- "architectures": [
4
- "ViTForImageClassification"
5
- ],
6
- "hidden_size": 768,
7
- "num_hidden_layers": 12,
8
- "num_attention_heads": 12,
9
- "intermediate_size": 3072,
10
- "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.0,
12
- "attention_probs_dropout_prob": 0.0,
13
- "initializer_range": 0.02,
14
- "layer_norm_eps": 1e-12,
15
- "image_size": 512,
16
- "patch_size": 16,
17
- "num_channels": 1,
18
- "qkv_bias": true
19
  }
 
1
  {
2
+ "model_type": "vit",
3
+ "architectures": ["ViTForImageClassification"],
4
+ "image_size": 512,
5
+ "patch_size": 16,
6
+ "num_labels": 15,
7
+ "hidden_size": 768,
8
+ "num_attention_heads": 12,
9
+ "num_hidden_layers": 12
 
 
 
 
 
 
 
 
 
10
  }
figures/fig1.png CHANGED
figures/fig2.png CHANGED
figures/fig3.png CHANGED
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cb65aecc1a4620d34d4069cc12aa569580dc1564151c3a97edd2060cc795b56
3
- size 39
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0727b93b529e6a279064c8d3b57fe11f1c4193ad9873cb0dadf029fab6d8b23
3
+ size 42