Link to Primus paper and update pipeline tag

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +33 -3
README.md CHANGED
@@ -1,12 +1,22 @@
1
  ---
2
- license: cc-by-4.0
3
  datasets:
4
  - AnonRes/OpenMind
5
- pipeline_tag: image-feature-extraction
 
6
  tags:
7
  - medical
8
  ---
9
 
 
 
 
 
 
 
 
 
 
 
10
  # OpenMind Benchmark 3D SSL Models
11
 
12
  > **Model from the paper**: [An OpenMind for 3D medical vision self-supervised learning](https://arxiv.org/abs/2412.17041)
@@ -36,7 +46,7 @@ Each model was pre-trained using a particular SSL method on the [OpenMind Datase
36
  We release SSL checkpoints for two backbone architectures:
37
 
38
  - **ResEnc-L**: A CNN-based encoder [[a](https://arxiv.org/abs/2410.23132), [b](https://arxiv.org/abs/2404.09556)]
39
- - **Primus-M**: A transformer-based encoder [[Primus paper](https://arxiv.org/abs/2503.01835)]
40
 
41
  Each encoder has been pre-trained using one of the following SSL techniques:
42
 
@@ -50,3 +60,23 @@ Each encoder has been pre-trained using one of the following SSL techniques:
50
  | [SimMIM](https://openaccess.thecvf.com/content/CVPR2022/html/Xie_SimMIM_A_Simple_Framework_for_Masked_Image_Modeling_CVPR_2022_paper.html) | Simple masked reconstruction based pretraining method (TR only) |
51
  | [SwinUNETR SSL](https://arxiv.org/abs/2111.14791) | Rotation, Contrastive and Reconstruction based pre-training method. |
52
  | [SimCLR](https://arxiv.org/abs/2002.05709) | Transfer of 2D Contrastive learning baseline method to 3D |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  datasets:
3
  - AnonRes/OpenMind
4
+ license: cc-by-4.0
5
+ pipeline_tag: image-segmentation
6
  tags:
7
  - medical
8
  ---
9
 
10
+ # Primus: Enforcing Attention Usage for 3D Medical Image Segmentation
11
+
12
+ This repository contains pre-trained checkpoints for **Primus**, a Transformer-centric architecture for 3D medical image segmentation.
13
+
14
+ - **Paper**: [Primus: Enforcing Attention Usage for 3D Medical Image Segmentation](https://huggingface.co/papers/2503.01835)
15
+ - **Code**: [MIC-DKFZ/nnUNet](https://github.com/MIC-DKFZ/nnUNet/blob/master/documentation/primus.md)
16
+ - **Pre-training codebase**: [MIC-DKFZ/nnssl](https://github.com/MIC-DKFZ/nnssl)
17
+
18
+ ---
19
+
20
  # OpenMind Benchmark 3D SSL Models
21
 
22
  > **Model from the paper**: [An OpenMind for 3D medical vision self-supervised learning](https://arxiv.org/abs/2412.17041)
 
46
  We release SSL checkpoints for two backbone architectures:
47
 
48
  - **ResEnc-L**: A CNN-based encoder [[a](https://arxiv.org/abs/2410.23132), [b](https://arxiv.org/abs/2404.09556)]
49
+ - **Primus-M**: A transformer-based encoder introduced in the [Primus paper](https://huggingface.co/papers/2503.01835)
50
 
51
  Each encoder has been pre-trained using one of the following SSL techniques:
52
 
 
60
  | [SimMIM](https://openaccess.thecvf.com/content/CVPR2022/html/Xie_SimMIM_A_Simple_Framework_for_Masked_Image_Modeling_CVPR_2022_paper.html) | Simple masked reconstruction based pretraining method (TR only) |
61
  | [SwinUNETR SSL](https://arxiv.org/abs/2111.14791) | Rotation, Contrastive and Reconstruction based pre-training method. |
62
  | [SimCLR](https://arxiv.org/abs/2002.05709) | Transfer of 2D Contrastive learning baseline method to 3D |
63
+
64
+ ## Citation
65
+
66
+ If you find this model or the OpenMind benchmark useful, please cite:
67
+
68
+ ```text
69
+ @article{wald2025primus,
70
+ title={Primus: Enforcing Attention Usage for 3D Medical Image Segmentation},
71
+ author={Wald, Tassilo and Roy, Saikat and Isensee, Fabian and Ulrich, Constantin and Ziegler, Sebastian and Trofimova, Dasha and Stock, Raphael and Baumgartner, Michael and K{\"o}hler, Gregor and Maier-Hein, Klaus},
72
+ journal={arXiv preprint arXiv:2503.01835},
73
+ year={2025}
74
+ }
75
+
76
+ @article{wald2024openmind,
77
+ title={An OpenMind for 3D medical vision self-supervised learning},
78
+ author={Wald, Tassilo and Ulrich, Constantin and Suprijadi, J. and Ziegler, Sebastian and Nohel, M. and Peretzke, R. and others},
79
+ journal={arXiv preprint arXiv:2412.17041},
80
+ year={2024}
81
+ }
82
+ ```