Pixel Diffusion UNet – LoDoInd (DM4CT)

This repository contains the pretrained pixel-space diffusion UNet used in the
DM4CT: Benchmarking Diffusion Models for CT Reconstruction (ICLR 2026) benchmark.

πŸ”— Paper: https://openreview.net/forum?id=YE5scJekg5
πŸ”— Arxiv: https://arxiv.org/abs/2602.18589
πŸ”— Codebase: https://github.com/DM4CT/DM4CT


πŸ”¬ Model Overview

This model learns a prior over CT reconstruction images using a denoising diffusion probabilistic model (DDPM).
It operates directly in pixel space (not latent space).

  • Architecture: 2D UNet (Diffusers UNet2DModel)
  • Input resolution: 512 Γ— 512
  • Channels: 1 (grayscale CT slice)
  • Training objective: Ξ΅-prediction (standard DDPM formulation)
  • Noise schedule: Linear beta schedule
  • Training dataset: Industry CT dataset (LoDoInd)
  • Intensity normalization: Rescaled to (-1, 1)

This model is intended to be combined with data-consistency correction for CT reconstruction.


πŸ“Š Dataset: LoDoInd

Source: https://www.aapm.org/grandchallenge/lowdosect/

Preprocessing steps:

  • Train/test split
  • Rescale reconstructed slices to (-1, 1)
  • No geometry information is embedded in the model

The model learns an unconditional image prior over CT slices.


🧠 Training Details

  • Optimizer: AdamW
  • Learning rate: 1e-4
  • Batch size: (insert your batch size)
  • Training steps: (insert number of steps)
  • Hardware: NVIDIA A100 GPU

Training script: https://github.com/DM4CT/DM4CT/blob/main/train_pixel.py


πŸš€ Usage

from diffusers import DDPMPipeline
pipeline = DDPMPipeline.from_pretrained("jiayangshi/lodoind_pixel_diffusion")
)

model.eval()
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Paper for jiayangshi/lodoind_pixel_diffusion