Instructions to use chen-hao-chao/mdm-prime with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use chen-hao-chao/mdm-prime with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="chen-hao-chao/mdm-prime")# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("chen-hao-chao/mdm-prime", dtype="auto") - Notebooks
- Google Colab
- Kaggle
- Local Apps
- vLLM
How to use chen-hao-chao/mdm-prime with vLLM:
Install from pip and serve model
# Install vLLM from pip: pip install vllm # Start the vLLM server: vllm serve "chen-hao-chao/mdm-prime" # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:8000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "chen-hao-chao/mdm-prime", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker
docker model run hf.co/chen-hao-chao/mdm-prime
- SGLang
How to use chen-hao-chao/mdm-prime with SGLang:
Install from pip and serve model
# Install SGLang from pip: pip install sglang # Start the SGLang server: python3 -m sglang.launch_server \ --model-path "chen-hao-chao/mdm-prime" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "chen-hao-chao/mdm-prime", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }'Use Docker images
docker run --gpus all \ --shm-size 32g \ -p 30000:30000 \ -v ~/.cache/huggingface:/root/.cache/huggingface \ --env "HF_TOKEN=<secret>" \ --ipc=host \ lmsysorg/sglang:latest \ python3 -m sglang.launch_server \ --model-path "chen-hao-chao/mdm-prime" \ --host 0.0.0.0 \ --port 30000 # Call the server using curl (OpenAI-compatible API): curl -X POST "http://localhost:30000/v1/completions" \ -H "Content-Type: application/json" \ --data '{ "model": "chen-hao-chao/mdm-prime", "prompt": "Once upon a time,", "max_tokens": 512, "temperature": 0.5 }' - Docker Model Runner
How to use chen-hao-chao/mdm-prime with Docker Model Runner:
docker model run hf.co/chen-hao-chao/mdm-prime
MDM-Prime
MDM-Prime is a discrete diffusion model enhanced with the Partial masking scheme (Prime). It enables fine-grained denoising and improves generation quality across both image and text domains. This model was proposed in our paper Beyond Masked and Unmasked: Discrete Diffusion Models via Partial Masking.
Model Details
- Text Generation
- Dataset: openwebtext (OWT)
- Model Size: 92M, 286M, 375M, 860M
- Context Length: 1,024
- Image Synthesis
- Dataset: CIFAR-10, ImageNet-32
- Model Size: 114M
- Context Length: 32x32x3
How to Use
To download the weights, one can download the huggingface_hub library via pip install -U huggingface_hub and perform the following python code:
from huggingface_hub import hf_hub_download
path = hf_hub_download(
repo_id="chen-hao-chao/mdm-prime",
filename="${checkpoint_name}.pth"
)
Replace ${checkpoint_name}.pth with ${task}/${dataset}/${setup}/${checkpoint_name}.pth (e.g., image/imagenet32/results_prime_l8_imagenet32/checkpoint-599.pth). This repository is organized as follows:
mdm-prime/
βββ README.md
βββ image/
| βββ cifar10/
| βββ imagenet/
| βββ results_mdm_imagenet32/
| βββ results_prime_supertoken_imagenet32/
| βββ results_prime_l2_imagenet32/
| βββ results_prime_l3_imagenet32/
| βββ results_prime_l4_imagenet32/
| βββ results_prime_l6_imagenet32/
| βββ results_prime_l8_imagenet32/
| βββ checkpoint-599.pth
βββ text/
βββ owt/
βββ results_prime_l2_owt/
βββ results_prime_l2_co_owt/
βββ results_prime_l3_owt/
βββ results_prime_l3_co_owt/
βββ results_prime_l4_owt/
βββ results_prime_l4_co_owt/
βββ results_prime_l6_owt/
βββ results_prime_l6_co_owt/
βββ results_prime_l8_owt/
βββ results_prime_l8_co_owt/
βββ checkpoint.ckpt
For more details regarding the training and inference processes, please refer to our github repository: chen-hao-chao/mdm-prime.
Citing MDM-Prime
If you find this code implementation useful, please consider citing our paper.
@inproceedings{chao2025mdmprime,
title = {{Beyond Masked and Unmasked: Discrete Diffusion Models via Partial Masking}},
author = {Chen-Hao Chao, Wei-Fang Sun, Hanwen Liang, Chun-Yi Lee, Rahul G. Krishnan},
booktitle = {Proceedings of the Conference on Neural Information Processing Systems (NeurIPS)},
year = {2025},
}
Model tree for chen-hao-chao/mdm-prime
Base model
kuleshov-group/mdlm-owt