AI-vBRAIN commited on
Commit
5bba72d
·
verified ·
1 Parent(s): 3a3458b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +44 -5
README.md CHANGED
@@ -9,8 +9,6 @@ framework: keras
9
  task: image-translation
10
  ---
11
 
12
-
13
-
14
  # pyMEAL: Multi-Encoder-Augmentation-Aware-Learning
15
 
16
  pyMEAL is a multi-encoder framework for augmentation-aware learning that accurately performs CT-to-T1-weighted MRI translation under diverse augmentations. It utilizes four dedicated encoders and three fusion strategies, concatenation (CC), fusion layer (FL), and controller block (BD), to capture augmentation-specific features. MEAL-BD outperforms conventional augmentation methods, achieving SSIM > 0.83 and PSNR > 25 dB in CT-to-T1w translation.
@@ -27,8 +25,49 @@ scipy
27
 
28
  antspyx
29
 
30
- ## Tutorials
31
- To use the different modules of pyMEAL, please refer to the tutorial section in our GitHub repository (https://github.com/ai-vbrain/pyMEAL)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
  ## How to get support?
34
- Just write to amoilyas@hkcoche.org or amaradesa@hkcoche.org
 
9
  task: image-translation
10
  ---
11
 
 
 
12
  # pyMEAL: Multi-Encoder-Augmentation-Aware-Learning
13
 
14
  pyMEAL is a multi-encoder framework for augmentation-aware learning that accurately performs CT-to-T1-weighted MRI translation under diverse augmentations. It utilizes four dedicated encoders and three fusion strategies, concatenation (CC), fusion layer (FL), and controller block (BD), to capture augmentation-specific features. MEAL-BD outperforms conventional augmentation methods, achieving SSIM > 0.83 and PSNR > 25 dB in CT-to-T1w translation.
 
25
 
26
  antspyx
27
 
28
+
29
+ ---
30
+
31
+ ## Available Models
32
+
33
+ | Model ID | File Name | Description |
34
+ |----------|------------------------------------------------|---------------------------------------------|
35
+ | BD | `builder1_mode1l1abW512_1_11211z1p1rt_.h5` | Builder-based architecture model |
36
+ | CC | `best_moderRl_RHID2_1mo.h5` | Encoder-concatenation-based configuration |
37
+ | FL | `bestac22_mode3l_512m2_m21.h5` | Feature-level fusion-based model |
38
+ | NA | `direct7_11ag23f11.h5` | Direct training baseline model |
39
+ | TA | `best_modelaf2ndab7_221ag12g11.h5` | traditional augmentation configuration model|
40
+
41
+ ---
42
+
43
+ ## Download Model Files
44
+
45
+ You can download any `.h5` file directly:
46
+
47
+ - [Download builder1_mode1l1abW512_1_11211z1p1rt_.h5](https://huggingface.co/AI-vBRAIN/pyMEAL/resolve/main/builder1_mode1l1abW512_1_11211z1p1rt_.h5)
48
+ - [Download best_moderRl_RHID2_1mo.h5](https://huggingface.co/AI-vBRAIN/pyMEAL/resolve/main/best_moderRl_RHID2_1mo.h5)
49
+ - [Download bestac22_mode3l_512m2_m21.h5](https://huggingface.co/AI-vBRAIN/pyMEAL/resolve/main/bestac22_mode3l_512m2_m21.h5)
50
+ - [Download direct7_11ag23f11.h5](https://huggingface.co/AI-vBRAIN/pyMEAL/resolve/main/direct7_11ag23f11.h5)
51
+ - [Download best_modelaf2ndab7_221ag12g11.h5](https://huggingface.co/AI-vBRAIN/pyMEAL/resolve/main/best_modelaf2ndab7_221ag12g11.h5)
52
+
53
+ ---
54
+
55
+ ## How to Use
56
+
57
+ ### Load a Model (Basic)
58
+
59
+ ```python
60
+ import tensorflow as tf
61
+
62
+ # Load the model
63
+ model = tf.keras.models.load_model("model.h5", compile=False)
64
+
65
+ # Run inference
66
+ output = model.predict(input_data)
67
+
68
+ Here, input_data refers to a CT image, and the corresponding T1-weighted (T1w) image is produced as the output.
69
+
70
+ For detailed instructions on how to use each module of the pyMEAL software, please refer to the tutorial section of our GitHub repository.
71
 
72
  ## How to get support?
73
+ Just write to Dr. Ilyas (amoilyas@hkcoche.org) or Dr. Maradesa (amaradesa@hkcoche.org)