xueh commited on
Commit
fca9dbe
·
verified ·
1 Parent(s): a081ba5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -182
README.md CHANGED
@@ -1,182 +0,0 @@
1
- ---
2
- license: mit
3
- ---
4
-
5
- # Model Card for SNRAware
6
-
7
- <!-- Provide a quick summary of what the model is/does. -->
8
-
9
- A deep learning imaging AI model with imaging transformer, for MR denoising.
10
-
11
- ![image](https://cdn-uploads.huggingface.co/production/uploads/690cf6928b7eacea549fd405/ohWF8O1xe0rOJWvhXmxLf.png)
12
-
13
- ## Model Details
14
-
15
- SNRAware is an imaging transformer model trained to denoise complex MR image data.
16
- Imaging transformers use attention modules to capture local, global, and inter-frame
17
- signal and noise characteristics. Denoising training used the SNRAware method,
18
- generating MR-realistic noises on the fly to create low SNR samples with unitary noise scaling.
19
- Model received low SNR complex images and g-factor maps as input, producing high SNR complex images as output.
20
-
21
- Please refer to the publication for technical details.
22
-
23
- ### Model Description
24
-
25
- <!-- Provide a longer summary of what this model is. -->
26
-
27
- The published model was instantiated with a high-res net (HRnet) backbone and consists of multiple imaging attention modules.
28
- Two models were published:
29
- - SNRAware-small: a 27.7 million parameter model
30
- - SNRAware-medium: a 55.1 million parameter model
31
- - SNRAware-large: a 109 million parameter model
32
-
33
- Input to model is 5D tensor [B, C, T/F, H, W] for batch, channel, time/frame, height and width. Output tensor is in the shape of
34
- [B, C-1, T/F, H, W]. The last channel in input is the g-factor map.
35
-
36
- - **Developed by:** Microsoft Research, Health Futures
37
- - **Model type:** Imaging Transformer
38
- - **License:** MIT
39
-
40
- ### Model Sources
41
-
42
- <!-- Provide the basic links for the model. -->
43
-
44
- - **Repository:** https://github.com/microsoft/SNRAware
45
- - **Paper:** https://pubs.rsna.org/doi/10.1148/ryai.250227
46
-
47
- ## Uses
48
-
49
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
50
-
51
- The model should accept the reconstructed MR complex images and g-factor maps and produce the denoising images.
52
- The model expects the unitary noise level in the input images. This can be achieved by reconstruction images with
53
- Gadgetron framework.
54
-
55
- ### Direct Use
56
-
57
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
58
-
59
- Denoise the complex MR images.
60
-
61
- ### Out-of-Scope Use
62
-
63
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
64
-
65
- No out-of-scope use should be attempted. Input to the model should be SNRUnit reconstructed.
66
-
67
- ## Bias, Risks, and Limitations
68
-
69
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
70
-
71
- No bias and risks are associated with this model. Only limitation is input data should have unitary noise scaling.
72
-
73
- ### Recommendations
74
-
75
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
76
-
77
- None
78
-
79
- ## How to Get Started with the Model
80
-
81
- Please refer to the documentation in https://github.com/microsoft/SNRAware to get started.
82
-
83
- ## Training Details
84
-
85
- ### Training Data
86
-
87
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
-
89
- Training data is not shared.
90
-
91
- ### Training Procedure
92
-
93
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
- Please refer to the publication for training details.
95
-
96
- #### Training Hyperparameters
97
-
98
- - **Training regime:** fp32 <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
99
-
100
- #### Speeds, Sizes, Times
101
-
102
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
103
-
104
- For a typical MR image data, the inference time on H100 is ~3-7s for SNRAware-small and ~5-11s for SNRAware-medium.
105
-
106
- ## Evaluation
107
-
108
- Both in-distribution and generalization tests were performed. For the in-distribution tests, the MR noise was generated and added to clean images to lower its SNR.
109
- Model was applied to restore the image quality. PSNR and SSIM were computed against the ground-truth. For the generalization tests, new data outside the training cohort
110
- were acquired for different imaging sequences, field strength, anatomies and resolution. Model outputs were compared to raw inputs and scored by clinicians to judge quality.
111
-
112
- Please refer to the publication for evaluation details.
113
-
114
- ### Results
115
-
116
- Imaging transformer model outperformed competing model architectures in the in-distribution tests.
117
- The SNRAware training also enabled imaging transformer models to generalize well to unseen applications without further training.
118
-
119
- #### Summary
120
-
121
-
122
- ## Environmental Impact
123
-
124
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
125
-
126
- - **Hardware Type:** B200
127
- - **Hours used:** 100
128
- - **Cloud Provider:** Azure
129
- - **Compute Region:** westus2
130
- - **Carbon Emitted:** 7.5
131
-
132
- ### Model Architecture and Objective
133
-
134
- Imaging Transformer with a high-res net backbone
135
-
136
- ### Compute Infrastructure
137
-
138
- Azure GPU VMs
139
-
140
- #### Hardware
141
-
142
- NVIDIA B200 x16
143
-
144
- #### Software
145
-
146
- Pytorch 2.8.0+cu128
147
-
148
- ## Citation
149
-
150
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
151
-
152
- **BibTeX:**
153
-
154
- ```latex
155
- @article{
156
- doi:10.1148/ryai.250227,
157
- author = {Xue, Hui and Hooper, Sarah M. and Pierce, Iain and Davies, Rhodri H. and Stairs, John and Naegele, Joseph and Campbell-Washburn, Adrienne E. and Manisty, Charlotte and Moon, James C. and Treibel, Thomas A. and Hansen, Michael S. and Kellman, Peter},
158
- title = {SNRAware: Improved Deep Learning MRI Denoising with Signal-to-noise Ratio Unit Training and G-factor Map Augmentation},
159
- journal = {Radiology: Artificial Intelligence},
160
- volume = {0},
161
- number = {ja},
162
- pages = {e250227},
163
- year = {0},
164
- doi = {10.1148/ryai.250227},
165
- note ={PMID: 41123451},
166
- URL = {https://doi.org/10.1148/ryai.250227}
167
- }
168
-
169
- @misc{xue2024imagingtransformermridenoising,
170
- title={Imaging transformer for MRI denoising with the SNR unit training: enabling generalization across field-strengths, imaging contrasts, and anatomy},
171
- author={Hui Xue and Sarah Hooper and Azaan Rehman and Iain Pierce and Thomas Treibel and Rhodri Davies and W Patricia Bandettini and Rajiv Ramasawmy and Ahsan Javed and Zheren Zhu and Yang Yang and James Moon and Adrienne Campbell and Peter Kellman},
172
- year={2024},
173
- eprint={2404.02382},
174
- archivePrefix={arXiv},
175
- primaryClass={eess.IV},
176
- url={https://arxiv.org/abs/2404.02382},
177
- }
178
- ```
179
-
180
- ## Model Card Contact
181
-
182
- Hui Xue, xueh@microsoft.com