Pixel Diffusion UNet β LoDoInd (DM4CT)
This repository contains the pretrained pixel-space diffusion UNet used in the
DM4CT: Benchmarking Diffusion Models for CT Reconstruction (ICLR 2026) benchmark.
π Paper: https://openreview.net/forum?id=YE5scJekg5
π Arxiv: https://arxiv.org/abs/2602.18589
π Codebase: https://github.com/DM4CT/DM4CT
π¬ Model Overview
This model learns a prior over CT reconstruction images using a denoising diffusion probabilistic model (DDPM).
It operates directly in pixel space (not latent space).
- Architecture: 2D UNet (Diffusers
UNet2DModel) - Input resolution: 512 Γ 512
- Channels: 1 (grayscale CT slice)
- Training objective: Ξ΅-prediction (standard DDPM formulation)
- Noise schedule: Linear beta schedule
- Training dataset: Industry CT dataset (LoDoInd)
- Intensity normalization: Rescaled to (-1, 1)
This model is intended to be combined with data-consistency correction for CT reconstruction.
π Dataset: LoDoInd
Source: https://www.aapm.org/grandchallenge/lowdosect/
Preprocessing steps:
- Train/test split
- Rescale reconstructed slices to (-1, 1)
- No geometry information is embedded in the model
The model learns an unconditional image prior over CT slices.
π§ Training Details
- Optimizer: AdamW
- Learning rate: 1e-4
- Batch size: (insert your batch size)
- Training steps: (insert number of steps)
- Hardware: NVIDIA A100 GPU
Training script: https://github.com/DM4CT/DM4CT/blob/main/train_pixel.py
π Usage
from diffusers import DDPMPipeline
pipeline = DDPMPipeline.from_pretrained("jiayangshi/lodoind_pixel_diffusion")
)
model.eval()
- Downloads last month
- -