Add pipeline tag and improve model card
#1
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,30 +1,29 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
library_name: diffusers
|
|
|
|
|
|
|
| 4 |
tags:
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
---
|
| 12 |
|
| 13 |
# Pixel Diffusion UNet β LoDoChallenge (DM4CT)
|
| 14 |
|
| 15 |
-
This repository contains the pretrained **pixel-space diffusion UNet**
|
| 16 |
-
**DM4CT: Benchmarking Diffusion Models for CT Reconstruction (ICLR 2026)** benchmark.
|
| 17 |
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
|
| 22 |
---
|
| 23 |
|
| 24 |
## π¬ Model Overview
|
| 25 |
|
| 26 |
-
This model learns a **prior over CT reconstruction images** using a denoising diffusion probabilistic model (DDPM).
|
| 27 |
-
It operates directly in **pixel space** (not latent space).
|
| 28 |
|
| 29 |
- **Architecture**: 2D UNet (Diffusers `UNet2DModel`)
|
| 30 |
- **Input resolution**: 512 Γ 512
|
|
@@ -34,42 +33,58 @@ It operates directly in **pixel space** (not latent space).
|
|
| 34 |
- **Training dataset**: Low Dose Grand Challenge (LoDoChallenge)
|
| 35 |
- **Intensity normalization**: Rescaled to (-1, 1)
|
| 36 |
|
| 37 |
-
This model is intended to be combined with data-consistency correction for CT reconstruction.
|
| 38 |
|
| 39 |
---
|
| 40 |
|
| 41 |
## π Dataset: Low Dose Grand Challenge
|
| 42 |
|
| 43 |
-
Source:
|
| 44 |
-
https://www.aapm.org/grandchallenge/lowdosect/
|
| 45 |
|
| 46 |
Preprocessing steps:
|
| 47 |
-
- Train/test split
|
| 48 |
-
- Rescale reconstructed slices to (-1, 1)
|
| 49 |
-
- No geometry information is embedded in the model
|
| 50 |
|
| 51 |
-
The model learns an unconditional image prior over CT slices.
|
| 52 |
|
| 53 |
---
|
| 54 |
|
| 55 |
## π§ Training Details
|
| 56 |
|
| 57 |
-
- Optimizer
|
| 58 |
-
- Learning rate
|
| 59 |
-
-
|
| 60 |
-
- Training
|
| 61 |
-
- Hardware: NVIDIA A100 GPU
|
| 62 |
-
|
| 63 |
-
Training script:
|
| 64 |
-
https://github.com/DM4CT/DM4CT/blob/main/train_pixel.py
|
| 65 |
|
| 66 |
---
|
| 67 |
|
| 68 |
## π Usage
|
| 69 |
|
|
|
|
|
|
|
| 70 |
```python
|
| 71 |
from diffusers import DDPMPipeline
|
|
|
|
|
|
|
| 72 |
pipeline = DDPMPipeline.from_pretrained("jiayangshi/lodochallenge_pixel_diffusion")
|
| 73 |
-
)
|
| 74 |
|
| 75 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
library_name: diffusers
|
| 3 |
+
license: mit
|
| 4 |
+
pipeline_tag: image-to-image
|
| 5 |
tags:
|
| 6 |
+
- computed-tomography
|
| 7 |
+
- ct-reconstruction
|
| 8 |
+
- diffusion-model
|
| 9 |
+
- inverse-problems
|
| 10 |
+
- dm4ct
|
| 11 |
+
- sparse-view-ct
|
| 12 |
---
|
| 13 |
|
| 14 |
# Pixel Diffusion UNet β LoDoChallenge (DM4CT)
|
| 15 |
|
| 16 |
+
This repository contains the pretrained **pixel-space diffusion UNet** presented in the paper [DM4CT: Benchmarking Diffusion Models for Computed Tomography Reconstruction](https://huggingface.co/papers/2602.18589) (ICLR 2026).
|
|
|
|
| 17 |
|
| 18 |
+
- **Project Page:** [https://dm4ct.github.io/DM4CT/](https://dm4ct.github.io/DM4CT/)
|
| 19 |
+
- **Codebase:** [https://github.com/DM4CT/DM4CT](https://github.com/DM4CT/DM4CT)
|
| 20 |
+
- **Paper:** [DM4CT: Benchmarking Diffusion Models for Computed Tomography Reconstruction](https://huggingface.co/papers/2602.18589)
|
| 21 |
|
| 22 |
---
|
| 23 |
|
| 24 |
## π¬ Model Overview
|
| 25 |
|
| 26 |
+
This model learns a **prior over CT reconstruction images** using a denoising diffusion probabilistic model (DDPM). It operates directly in **pixel space** (not latent space).
|
|
|
|
| 27 |
|
| 28 |
- **Architecture**: 2D UNet (Diffusers `UNet2DModel`)
|
| 29 |
- **Input resolution**: 512 Γ 512
|
|
|
|
| 33 |
- **Training dataset**: Low Dose Grand Challenge (LoDoChallenge)
|
| 34 |
- **Intensity normalization**: Rescaled to (-1, 1)
|
| 35 |
|
| 36 |
+
This model is intended to be combined with data-consistency correction for CT reconstruction tasks as detailed in the DM4CT benchmark.
|
| 37 |
|
| 38 |
---
|
| 39 |
|
| 40 |
## π Dataset: Low Dose Grand Challenge
|
| 41 |
|
| 42 |
+
Source: [AAPM Low Dose CT Grand Challenge](https://www.aapm.org/grandchallenge/lowdosect/)
|
|
|
|
| 43 |
|
| 44 |
Preprocessing steps:
|
| 45 |
+
- Train/test split.
|
| 46 |
+
- Rescale reconstructed slices to (-1, 1).
|
| 47 |
+
- No geometry information is embedded in the model.
|
| 48 |
|
| 49 |
+
The model learns an unconditional image prior over medical CT slices.
|
| 50 |
|
| 51 |
---
|
| 52 |
|
| 53 |
## π§ Training Details
|
| 54 |
|
| 55 |
+
- **Optimizer**: AdamW
|
| 56 |
+
- **Learning rate**: 1e-4
|
| 57 |
+
- **Hardware**: NVIDIA A100 GPU
|
| 58 |
+
- **Training script**: [train_pixel.py](https://github.com/DM4CT/DM4CT/blob/main/train_pixel.py)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 59 |
|
| 60 |
---
|
| 61 |
|
| 62 |
## π Usage
|
| 63 |
|
| 64 |
+
You can use this model with the `diffusers` library:
|
| 65 |
+
|
| 66 |
```python
|
| 67 |
from diffusers import DDPMPipeline
|
| 68 |
+
|
| 69 |
+
# Load the pretrained pipeline
|
| 70 |
pipeline = DDPMPipeline.from_pretrained("jiayangshi/lodochallenge_pixel_diffusion")
|
|
|
|
| 71 |
|
| 72 |
+
# Generate a sample (unconditional CT slice prior)
|
| 73 |
+
image = pipeline().images[0]
|
| 74 |
+
image.save("generated_ct_slice.png")
|
| 75 |
+
```
|
| 76 |
+
|
| 77 |
+
---
|
| 78 |
+
|
| 79 |
+
## Citation
|
| 80 |
+
|
| 81 |
+
```bibtex
|
| 82 |
+
@inproceedings{
|
| 83 |
+
shi2026dmct,
|
| 84 |
+
title={{DM}4{CT}: Benchmarking Diffusion Models for Computed Tomography Reconstruction},
|
| 85 |
+
author={Shi, Jiayang and Pelt, Dani{\"e}l M and Batenburg, K Joost},
|
| 86 |
+
booktitle={The Fourteenth International Conference on Learning Representations},
|
| 87 |
+
year={2026},
|
| 88 |
+
url={https://openreview.net/forum?id=YE5scJekg5}
|
| 89 |
+
}
|
| 90 |
+
```
|