Improve model card metadata and fix usage snippet
Browse filesHi! I'm Niels from the community team at Hugging Face.
I've noticed this model is part of the DM4CT benchmark for CT reconstruction. This PR improves the model card by:
- Adding the `pipeline_tag: image-to-image` to improve discoverability.
- Fixing a syntax error in the Python usage snippet.
- Linking the model to the associated paper page on Hugging Face.
README.md
CHANGED
|
@@ -1,23 +1,23 @@
|
|
| 1 |
---
|
| 2 |
-
license: mit
|
| 3 |
library_name: diffusers
|
|
|
|
|
|
|
| 4 |
tags:
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
|
| 8 |
-
|
| 9 |
-
|
| 10 |
-
|
| 11 |
---
|
| 12 |
|
| 13 |
# Pixel Diffusion UNet β Real-world Synchrotron Dataset (DM4CT)
|
| 14 |
|
| 15 |
-
This repository contains the pretrained **pixel-space diffusion UNet**
|
| 16 |
-
**DM4CT: Benchmarking Diffusion Models for CT Reconstruction (ICLR 2026)** benchmark.
|
| 17 |
|
| 18 |
-
π
|
| 19 |
-
π Arxiv
|
| 20 |
-
π Codebase
|
| 21 |
|
| 22 |
---
|
| 23 |
|
|
@@ -31,7 +31,7 @@ It operates directly in **pixel space** (not latent space).
|
|
| 31 |
- **Channels**: 1 (grayscale CT slice)
|
| 32 |
- **Training objective**: Ξ΅-prediction (standard DDPM formulation)
|
| 33 |
- **Noise schedule**: Linear beta schedule
|
| 34 |
-
- **Training dataset**: Synchrotron Dataset of rocks
|
| 35 |
- **Intensity normalization**: Rescaled to (-1, 1)
|
| 36 |
|
| 37 |
This model is intended to be combined with data-consistency correction for CT reconstruction.
|
|
@@ -40,8 +40,7 @@ This model is intended to be combined with data-consistency correction for CT re
|
|
| 40 |
|
| 41 |
## π Dataset: Real-world Synchrotron Dataset
|
| 42 |
|
| 43 |
-
Source:
|
| 44 |
-
https://zenodo.org/records/15420527
|
| 45 |
|
| 46 |
Preprocessing steps:
|
| 47 |
- Train/test split
|
|
@@ -54,22 +53,39 @@ The model learns an unconditional image prior over CT slices.
|
|
| 54 |
|
| 55 |
## π§ Training Details
|
| 56 |
|
| 57 |
-
- Optimizer
|
| 58 |
-
- Learning rate
|
| 59 |
-
-
|
| 60 |
-
- Training
|
| 61 |
-
- Hardware: NVIDIA A100 GPU
|
| 62 |
-
|
| 63 |
-
Training script:
|
| 64 |
-
https://github.com/DM4CT/DM4CT/blob/main/train_pixel.py
|
| 65 |
|
| 66 |
---
|
| 67 |
|
| 68 |
## π Usage
|
| 69 |
|
|
|
|
|
|
|
| 70 |
```python
|
| 71 |
from diffusers import DDPMPipeline
|
|
|
|
|
|
|
| 72 |
pipeline = DDPMPipeline.from_pretrained("jiayangshi/synchrotron_pixel_diffusion")
|
| 73 |
-
)
|
| 74 |
|
| 75 |
-
model
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
---
|
|
|
|
| 2 |
library_name: diffusers
|
| 3 |
+
license: mit
|
| 4 |
+
pipeline_tag: image-to-image
|
| 5 |
tags:
|
| 6 |
+
- computed-tomography
|
| 7 |
+
- ct-reconstruction
|
| 8 |
+
- diffusion-model
|
| 9 |
+
- inverse-problems
|
| 10 |
+
- dm4ct
|
| 11 |
+
- sparse-view-ct
|
| 12 |
---
|
| 13 |
|
| 14 |
# Pixel Diffusion UNet β Real-world Synchrotron Dataset (DM4CT)
|
| 15 |
|
| 16 |
+
This repository contains the pretrained **pixel-space diffusion UNet** presented in the paper [DM4CT: Benchmarking Diffusion Models for Computed Tomography Reconstruction](https://huggingface.co/papers/2602.18589).
|
|
|
|
| 17 |
|
| 18 |
+
π **Project Page:** [https://dm4ct.github.io/DM4CT/](https://dm4ct.github.io/DM4CT/)
|
| 19 |
+
π **Arxiv:** [https://arxiv.org/abs/2602.18589](https://arxiv.org/abs/2602.18589)
|
| 20 |
+
π **Codebase:** [https://github.com/DM4CT/DM4CT](https://github.com/DM4CT/DM4CT)
|
| 21 |
|
| 22 |
---
|
| 23 |
|
|
|
|
| 31 |
- **Channels**: 1 (grayscale CT slice)
|
| 32 |
- **Training objective**: Ξ΅-prediction (standard DDPM formulation)
|
| 33 |
- **Noise schedule**: Linear beta schedule
|
| 34 |
+
- **Training dataset**: Real-world Synchrotron Dataset of rocks
|
| 35 |
- **Intensity normalization**: Rescaled to (-1, 1)
|
| 36 |
|
| 37 |
This model is intended to be combined with data-consistency correction for CT reconstruction.
|
|
|
|
| 40 |
|
| 41 |
## π Dataset: Real-world Synchrotron Dataset
|
| 42 |
|
| 43 |
+
Source: [Zenodo](https://zenodo.org/records/15420527)
|
|
|
|
| 44 |
|
| 45 |
Preprocessing steps:
|
| 46 |
- Train/test split
|
|
|
|
| 53 |
|
| 54 |
## π§ Training Details
|
| 55 |
|
| 56 |
+
- **Optimizer**: AdamW
|
| 57 |
+
- **Learning rate**: 1e-4
|
| 58 |
+
- **Hardware**: NVIDIA A100 GPU
|
| 59 |
+
- **Training script**: [train_pixel.py](https://github.com/DM4CT/DM4CT/blob/main/train_pixel.py)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 60 |
|
| 61 |
---
|
| 62 |
|
| 63 |
## π Usage
|
| 64 |
|
| 65 |
+
You can use this model with the `diffusers` library as follows:
|
| 66 |
+
|
| 67 |
```python
|
| 68 |
from diffusers import DDPMPipeline
|
| 69 |
+
|
| 70 |
+
# Load the pipeline
|
| 71 |
pipeline = DDPMPipeline.from_pretrained("jiayangshi/synchrotron_pixel_diffusion")
|
|
|
|
| 72 |
|
| 73 |
+
# Access the UNet model
|
| 74 |
+
model = pipeline.unet
|
| 75 |
+
model.eval()
|
| 76 |
+
```
|
| 77 |
+
|
| 78 |
+
---
|
| 79 |
+
|
| 80 |
+
## Citation
|
| 81 |
+
|
| 82 |
+
```bibtex
|
| 83 |
+
@inproceedings{
|
| 84 |
+
shi2026dmct,
|
| 85 |
+
title={{DM}4{CT}: Benchmarking Diffusion Models for Computed Tomography Reconstruction},
|
| 86 |
+
author={Shi, Jiayang and Pelt, Dani{\"e}l M and Batenburg, K Joost},
|
| 87 |
+
booktitle={The Fourteenth International Conference on Learning Representations},
|
| 88 |
+
year={2026},
|
| 89 |
+
url={https://openreview.net/forum?id=YE5scJekg5}
|
| 90 |
+
}
|
| 91 |
+
```
|