Improve model card with Primus paper info and segmentation tag

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +43 -26
README.md CHANGED
@@ -1,18 +1,23 @@
1
  ---
2
- license: cc-by-4.0
3
  datasets:
4
  - AnonRes/OpenMind
5
- pipeline_tag: image-feature-extraction
 
6
  tags:
7
  - medical
8
  ---
9
 
10
- # OpenMind Benchmark 3D SSL Models
 
 
11
 
12
- > **Model from the paper**: [An OpenMind for 3D medical vision self-supervised learning](https://arxiv.org/abs/2412.17041)
13
- > **Pre-training codebase used to create checkpoint**: [MIC-DKFZ/nnssl](https://github.com/MIC-DKFZ/nnssl)
14
- > **Dataset**: [AnonRes/OpenMind](https://huggingface.co/datasets/AnonRes/OpenMind)
15
- > **Downstream (segmentation) fine-tuning**: [TaWald/nnUNet](https://github.com/TaWald/nnUNet)
 
 
 
16
 
17
  ---
18
 
@@ -20,16 +25,13 @@ tags:
20
 
21
  ## Overview
22
 
23
- This repository hosts pre-trained checkpoints from the **OpenMind** benchmark:
24
- 📄 **An OpenMind for 3D medical vision self-supervised learning** (Wald, T., Ulrich, C., Suprijadi, J., Ziegler, S., Nohel, M., Peretzke, R., ... & Maier-Hein, K. H. (2024).)
25
- ([arXiv:2412.17041](https://arxiv.org/abs/2412.17041)) — the first extensive benchmark study for **self-supervised learning (SSL)** on **3D medical imaging** data.
26
 
27
- Each model was pre-trained using a particular SSL method on the [OpenMind Dataset](https://huggingface.co/datasets/AnonRes/OpenMind), a large-scale, standardized collection of public brain MRI datasets.
28
 
29
- **These models are not recommended to be used as-is for feature extraction.** Instead we recommend using the downstream fine-tuning frameworks for **segmentation** and **classification** adaptation, available in the [adaptation repository](https://github.com/TaWald/nnUNet).
30
- *While manual download is possible, we recommend using the auto-download feature of the fine-tuning repository by providing the repository URL on Hugging Face instead of a local checkpoint path.*
31
 
32
- ---
33
 
34
  ## Model Variants
35
 
@@ -38,15 +40,30 @@ We release SSL checkpoints for two backbone architectures:
38
  - **ResEnc-L**: A CNN-based encoder [[a](https://arxiv.org/abs/2410.23132), [b](https://arxiv.org/abs/2404.09556)]
39
  - **Primus-M**: A transformer-based encoder [[Primus paper](https://arxiv.org/abs/2503.01835)]
40
 
41
- Each encoder has been pre-trained using one of the following SSL techniques:
42
-
43
- | Method | Description |
44
- |---------------|-------------|
45
- | [Volume Contrastive (VoCo)](https://arxiv.org/abs/2402.17300) | Contrastive pretraining method for 3D volumes |
46
- | [VolumeFusion (VF)](https://arxiv.org/abs/2306.16925) | Spatial volume fusion-based segmentation SSL method |
47
- | [Models Genesis (MG)](https://www.sciencedirect.com/science/article/pii/S1361841520302048) | Reconstruction and denoising based pretraining method |
48
- | [Masked Autoencoders (MAE)](https://openaccess.thecvf.com/content/CVPR2022/html/He_Masked_Autoencoders_Are_Scalable_Vision_Learners_CVPR_2022_paper) | Default reconstruction based pretraining method |
49
- | [Spark 3D (S3D)](https://arxiv.org/abs/2410.23132) | Sparse reconstruction based pretraining mehtod (CNN only) |
50
- | [SimMIM](https://openaccess.thecvf.com/content/CVPR2022/html/Xie_SimMIM_A_Simple_Framework_for_Masked_Image_Modeling_CVPR_2022_paper.html) | Simple masked reconstruction based pretraining method (TR only) |
51
- | [SwinUNETR SSL](https://arxiv.org/abs/2111.14791) | Rotation, Contrastive and Reconstruction based pre-training method. |
52
- | [SimCLR](https://arxiv.org/abs/2002.05709) | Transfer of 2D Contrastive learning baseline method to 3D |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  datasets:
3
  - AnonRes/OpenMind
4
+ license: cc-by-4.0
5
+ pipeline_tag: image-segmentation
6
  tags:
7
  - medical
8
  ---
9
 
10
+ # OpenMind Benchmark 3D SSL Models: Primus-M
11
+
12
+ This repository hosts pre-trained checkpoints for the **Primus-M** backbone architecture, as introduced in the paper:
13
 
14
+ 📄 **Primus: Enforcing Attention Usage for 3D Medical Image Segmentation**
15
+ (Wald, T., Roy, S., Isensee, F., Ulrich, C., Ziegler, S., Trofimova, D., Stock, R., Baumgartner, M., Köhler, G., & Maier-Hein, K.)
16
+ [[Hugging Face Paper](https://huggingface.co/papers/2503.01835)] [[arXiv:2503.01835](https://arxiv.org/abs/2503.01835)]
17
+
18
+ These weights were also utilized in the **OpenMind** benchmark study:
19
+ 📄 **An OpenMind for 3D medical vision self-supervised learning**
20
+ [[arXiv:2412.17041](https://arxiv.org/abs/2412.17041)]
21
 
22
  ---
23
 
 
25
 
26
  ## Overview
27
 
28
+ This repository hosts pre-trained checkpoints from the **OpenMind** benchmark, the first extensive benchmark study for **self-supervised learning (SSL)** on **3D medical imaging** data.
 
 
29
 
30
+ Each model was pre-trained using a particular SSL method on the [OpenMind Dataset](https://huggingface.co/datasets/AnonRes/OpenMind), a large-scale, standardized collection of public brain MRI datasets.
31
 
32
+ This specific checkpoint uses the **Primus-M** transformer-centric encoder pre-trained with the **VolumeFusion (VF)** self-supervised method. Primus addresses the over-reliance on convolutional blocks in existing architectures, leveraging high-resolution tokens and advanced block design to achieve state-of-the-art performance in 3D medical image segmentation.
 
33
 
34
+ **These models are not recommended to be used as-is for feature extraction.** Instead, we recommend using the downstream fine-tuning frameworks for **segmentation** and **classification** adaptation, available in the [adaptation repository](https://github.com/TaWald/nnUNet) and the main [nnU-Net](https://github.com/MIC-DKFZ/nnUNet) framework.
35
 
36
  ## Model Variants
37
 
 
40
  - **ResEnc-L**: A CNN-based encoder [[a](https://arxiv.org/abs/2410.23132), [b](https://arxiv.org/abs/2404.09556)]
41
  - **Primus-M**: A transformer-based encoder [[Primus paper](https://arxiv.org/abs/2503.01835)]
42
 
43
+ Each encoder has been pre-trained using one of the following SSL techniques: Volume Contrastive (VoCo), VolumeFusion (VF), Models Genesis (MG), Masked Autoencoders (MAE), Spark 3D (S3D), SimMIM, SwinUNETR SSL, or SimCLR.
44
+
45
+ ## Implementation
46
+
47
+ The code for Primus and the fine-tuning pipeline is available in the official nnU-Net repository:
48
+ - **Code**: [MIC-DKFZ/nnUNet](https://github.com/MIC-DKFZ/nnUNet/blob/master/documentation/primus.md)
49
+ - **Pre-training codebase**: [MIC-DKFZ/nnssl](https://github.com/MIC-DKFZ/nnssl)
50
+
51
+ ## Citation
52
+
53
+ If you use this model, please cite the following:
54
+
55
+ ```bibtex
56
+ @article{wald2025primus,
57
+ title={Primus: Enforcing Attention Usage for 3D Medical Image Segmentation},
58
+ author={Wald, Tassilo and Roy, Saikat and Isensee, Fabian and Ulrich, Constantin and Ziegler, Sebastian and Trofimova, Dasha and Stock, Raphael and Baumgartner, Michael and Köhler, Gregor and Maier-Hein, Klaus},
59
+ journal={arXiv preprint arXiv:2503.01835},
60
+ year={2025}
61
+ }
62
+
63
+ @article{wald2024openmind,
64
+ title={An OpenMind for 3D medical vision self-supervised learning},
65
+ author={Wald, Tassilo and Ulrich, Constantin and Suprijadi, J. and Ziegler, Sebastian and Nohel, M. and Peretzke, R. and others},
66
+ journal={arXiv preprint arXiv:2412.17041},
67
+ year={2024}
68
+ }
69
+ ```