whilethis commited on
Commit
1c09125
·
verified ·
1 Parent(s): 085f07a

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +78 -0
README.md ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ tags:
4
+ - medical-imaging
5
+ - self-supervised-learning
6
+ - masked-autoencoder
7
+ - 3d-ct
8
+ - pretraining
9
+ ---
10
+
11
+ # NEMESIS
12
+
13
+ **Superpatch-based 3D Medical Image Self-Supervised Pretraining via Noise-Enhanced Dual-Masking**
14
+
15
+ > IEEE AICAS 2026
16
+
17
+ ## Overview
18
+
19
+ NEMESIS is a self-supervised pretraining framework for 3D CT volumes using:
20
+ - **Superpatch processing** (128³ sub-volumes) — memory-efficient ViT pretraining
21
+ - **Dual-masking (MATB)** — plane-wise (xy) + axis-wise (z) masking, exploiting CT anisotropy
22
+ - **NEMESIS Tokens (NTs)** — learnable tokens summarising visible patches via cross-attention
23
+ - **Noise-enhanced reconstruction** — Gaussian noise injection for regularisation
24
+
25
+ ### Key result (BTCV organ classification, frozen linear probe)
26
+
27
+ | Method | AUROC |
28
+ |---|---|
29
+ | **NEMESIS (frozen)** | **0.9633** |
30
+ | SuPreM (fine-tuned) | 0.9493 |
31
+ | VoCo (fine-tuned) | 0.9387 |
32
+
33
+ ## Checkpoints
34
+
35
+ | File | embed_dim | depth | mask_ratio |
36
+ |---|---|---|---|
37
+ | `MAE_768_0.5.pt` | 768 | 6 | 0.5 | ← **main model (paper)** |
38
+ | `MAE_768_0.25.pt` | 768 | 6 | 0.25 | ablation |
39
+ | `MAE_768_0.75.pt` | 768 | 6 | 0.75 | ablation |
40
+ | `MAE_576_0.5.pt` | 576 | 6 | 0.5 | ablation |
41
+ | `MAE_384_0.5.pt` | 384 | 6 | 0.5 | ablation |
42
+ | (others) | | | | embed_dim × mask_ratio ablations |
43
+
44
+ ## Usage
45
+
46
+ ```bash
47
+ pip install huggingface_hub
48
+ huggingface-cli download whilethis/NEMESIS MAE_768_0.5.pt --local-dir pretrained/
49
+ ```
50
+
51
+ ```python
52
+ import torch
53
+ from nemesis.models.mae import MAEgic3DMAE
54
+
55
+ ckpt = torch.load("pretrained/MAE_768_0.5.pt", map_location="cpu")
56
+ model = MAEgic3DMAE(
57
+ embed_dim=768, depth=6, num_heads=8,
58
+ decoder_embed_dim=128, decoder_depth=3,
59
+ num_maegic_tokens=8,
60
+ )
61
+ model.load_state_dict(ckpt["model_state_dict"])
62
+ encoder = model.encoder
63
+ ```
64
+
65
+ ## Code
66
+
67
+ [https://github.com/whilethis00/NEMESIS-public](https://github.com/whilethis00/NEMESIS-public)
68
+
69
+ ## Citation
70
+
71
+ ```bibtex
72
+ @inproceedings{jung2026nemesis,
73
+ title = {{NEMESIS}: Superpatch-based 3{D} Medical Image Self-Supervised Pretraining via Noise-Enhanced Dual-Masking},
74
+ author = {Jung, Hyeonseok and others},
75
+ booktitle = {IEEE International Conference on Artificial Intelligence Circuits and Systems (AICAS)},
76
+ year = {2026},
77
+ }
78
+ ```