Delete readme.md with huggingface_hub
Browse files
readme.md
DELETED
|
@@ -1,52 +0,0 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: cc-by-4.0
|
| 3 |
-
datasets:
|
| 4 |
-
- AnonRes/OpenMind
|
| 5 |
-
pipeline_tag: image-feature-extraction
|
| 6 |
-
tags:
|
| 7 |
-
- medical
|
| 8 |
-
---
|
| 9 |
-
|
| 10 |
-
# OpenMind Benchmark 3D SSL Models
|
| 11 |
-
|
| 12 |
-
> **Model from the paper**: [An OpenMind for 3D medical vision self-supervised learning](https://arxiv.org/abs/2412.17041)
|
| 13 |
-
> **Pre-training codebase used to create checkpoint**: [MIC-DKFZ/nnssl](https://github.com/MIC-DKFZ/nnssl)
|
| 14 |
-
> **Dataset**: [AnonRes/OpenMind](https://huggingface.co/datasets/AnonRes/OpenMind)
|
| 15 |
-
> **Downstream (segmentation) fine-tuning**: [TaWald/nnUNet](https://github.com/TaWald/nnUNet)
|
| 16 |
-
|
| 17 |
-
---
|
| 18 |
-
|
| 19 |
-

|
| 20 |
-
|
| 21 |
-
## 🔍 Overview
|
| 22 |
-
|
| 23 |
-
This repository hosts pre-trained checkpoints from the **OpenMind** benchmark:
|
| 24 |
-
📄 **"An OpenMind for 3D medical vision self-supervised learning"**
|
| 25 |
-
([arXiv:2412.17041](https://arxiv.org/abs/2412.17041)) — the first extensive benchmark study for **self-supervised learning (SSL)** on **3D medical imaging** data.
|
| 26 |
-
|
| 27 |
-
The models were pre-trained using various SSL methods on the [OpenMind Dataset](https://huggingface.co/datasets/AnonRes/OpenMind), a large-scale, standardized collection of public brain MRI datasets.
|
| 28 |
-
|
| 29 |
-
**These models are not recommended to be used as-is.** Instead we recommend using the downstream fine-tuning pipelines for **segmentation** and **classification**, available in the [adaptation repository](https://github.com/TaWald/nnUNet).
|
| 30 |
-
*While direct download is possible, we recommend using the auto-download of the respective fine-tuning repositories.*
|
| 31 |
-
|
| 32 |
-
---
|
| 33 |
-
|
| 34 |
-
## 🧠 Model Variants
|
| 35 |
-
|
| 36 |
-
We release SSL checkpoints for two backbone architectures:
|
| 37 |
-
|
| 38 |
-
- **ResEnc-L**: A CNN-based encoder [[link1](https://arxiv.org/abs/2410.23132), [link2](https://arxiv.org/abs/2404.09556)]
|
| 39 |
-
- **Primus-M**: A transformer-based encoder [[Primus paper](https://arxiv.org/abs/2503.01835)]
|
| 40 |
-
|
| 41 |
-
Each encoder has been pre-trained using the following SSL techniques:
|
| 42 |
-
|
| 43 |
-
| Method | Description |
|
| 44 |
-
|---------------|-------------|
|
| 45 |
-
| [Volume Contrastive (VoCo)](https://arxiv.org/abs/2402.17300) | Global contrastive learning in 3D volumes |
|
| 46 |
-
| [VolumeFusion (VF)](https://arxiv.org/abs/2306.16925) | Spatial fusion-based SSL |
|
| 47 |
-
| [Models Genesis (MG)](https://www.sciencedirect.com/science/article/pii/S1361841520302048) | Classic 3D self-reconstruction |
|
| 48 |
-
| [Masked Autoencoders (MAE)](https://openaccess.thecvf.com/content/CVPR2022/html/He_Masked_Autoencoders_Are_Scalable_Vision_Learners_CVPR_2022_paper) | Patch masking and reconstruction |
|
| 49 |
-
| [Spark 3D (S3D)](https://arxiv.org/abs/2410.23132) | 3D adaptation of Spark framework |
|
| 50 |
-
| [SimMIM](https://openaccess.thecvf.com/content/CVPR2022/html/Xie_SimMIM_A_Simple_Framework_for_Masked_Image_Modeling_CVPR_2022_paper.html) | Simple masked reconstruction |
|
| 51 |
-
| [SwinUNETR SSL](https://arxiv.org/abs/2111.14791) | Transformer-based pre-training |
|
| 52 |
-
| [SimCLR](https://arxiv.org/abs/2002.05709) | Contrastive learning baseline |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|