datasets:
- AnonRes/OpenMind
license: cc-by-4.0
pipeline_tag: image-segmentation
tags:
- medical
OpenMind Benchmark 3D SSL Models
Models from the papers:
- An OpenMind for 3D medical vision self-supervised learning
- Primus: Enforcing Attention Usage for 3D Medical Image Segmentation Pre-training codebase used to create checkpoint: MIC-DKFZ/nnssl
Dataset: AnonRes/OpenMind
Downstream (segmentation) fine-tuning: TaWald/nnUNet Official Code Documentation: Primus in nnU-Net
Overview
This repository hosts pre-trained checkpoints from the OpenMind benchmark:
📄 An OpenMind for 3D medical vision self-supervised learning (Wald, T., Ulrich, C., Suprijadi, J., Ziegler, S., Nohel, M., Peretzke, R., ... & Maier-Hein, K. H. (2024).)
(arXiv:2412.17041) — the first extensive benchmark study for self-supervised learning (SSL) on 3D medical imaging data.
It also features the Primus architecture:
📄 Primus: Enforcing Attention Usage for 3D Medical Image Segmentation (Wald, T., Roy, S., Isensee, F., Ulrich, C., Ziegler, S., Trofimova, D., ... & Maier-Hein, K. H. (2025).)
(Hugging Face Papers) — introduction of Transformer-centric segmentation architectures that achieve state-of-the-art results.
Each model was pre-trained using a particular SSL method on the OpenMind Dataset, a large-scale, standardized collection of public brain MRI datasets.
These models are not recommended to be used as-is for feature extraction. Instead we recommend using the downstream fine-tuning frameworks for segmentation and classification adaptation, available in the adaptation repository. While manual download is possible, we recommend using the auto-download feature of the fine-tuning repository by providing the repository URL on Hugging Face instead of a local checkpoint path.
Model Variants
We release SSL checkpoints for two backbone architectures:
- ResEnc-L: A CNN-based encoder [a, b]
- Primus-M: A transformer-based encoder [Primus paper]
Each encoder has been pre-trained using one of the following SSL techniques:
| Method | Description |
|---|---|
| Volume Contrastive (VoCo) | Contrastive pretraining method for 3D volumes |
| VolumeFusion (VF) | Spatial volume fusion-based segmentation SSL method |
| Models Genesis (MG) | Reconstruction and denoising based pretraining method |
| Masked Autoencoders (MAE) | Default reconstruction based pretraining method |
| Spark 3D (S3D) | Sparse reconstruction based pretraining mehtod (CNN only) |
| SimMIM | Simple masked reconstruction based pretraining method (TR only) |
| SwinUNETR SSL | Rotation, Contrastive and Reconstruction based pre-training method. |
| SimCLR | Transfer of 2D Contrastive learning baseline method to 3D |
Citation
If you use these models or the Primus architecture, please cite:
@article{wald2025primus,
title={Primus: Enforcing Attention Usage for 3D Medical Image Segmentation},
author={Wald, Tassilo and Roy, Saikat and Isensee, Fabian and Ulrich, Constantin and Ziegler, Sebastian and Trofimova, Dasha and Stock, Raphael and Baumgartner, Michael and Köhler, Gregor and Maier-Hein, Klaus},
journal={arXiv preprint arXiv:2503.01835},
year={2025}
}
@article{wald2024openmind,
title={An OpenMind for 3D medical vision self-supervised learning},
author={Wald, Tassilo and Ulrich, Constantin and Suprijadi, J and Ziegler, Sebastian and Nohel, M and Peretzke, R and others},
journal={arXiv preprint arXiv:2412.17041},
year={2024}
}
