How to use from
SGLang
Install from pip and serve model
# Install SGLang from pip:
pip install sglang
# Start the SGLang server:
python3 -m sglang.launch_server \
    --model-path "d3LLM/d3LLM_Dream_Coder" \
    --host 0.0.0.0 \
    --port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "d3LLM/d3LLM_Dream_Coder",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Use Docker images
docker run --gpus all \
    --shm-size 32g \
    -p 30000:30000 \
    -v ~/.cache/huggingface:/root/.cache/huggingface \
    --env "HF_TOKEN=<secret>" \
    --ipc=host \
    lmsysorg/sglang:latest \
    python3 -m sglang.launch_server \
        --model-path "d3LLM/d3LLM_Dream_Coder" \
        --host 0.0.0.0 \
        --port 30000
# Call the server using curl (OpenAI-compatible API):
curl -X POST "http://localhost:30000/v1/chat/completions" \
	-H "Content-Type: application/json" \
	--data '{
		"model": "d3LLM/d3LLM_Dream_Coder",
		"messages": [
			{
				"role": "user",
				"content": "What is the capital of France?"
			}
		]
	}'
Quick Links

d3LLM: Ultra-Fast Diffusion LLM using Pseudo-Trajectory Distillation 🚀

d3LLM-Dream-Coder is an ultra-fast diffusion language model introduced in the paper d3LLM: Ultra-Fast Diffusion LLM using Pseudo-Trajectory Distillation. It is built on Dream-org/Dream-Coder-v0-Instruct-7B.

Model Description

d3LLM (pseuDo-Distilled Diffusion Large Language Model) is a framework designed to strike a balance between accuracy and parallelism in diffusion LLMs. It achieves up to 10× speedup over vanilla diffusion models like LLaDA/Dream and 5× speedup over autoregressive (AR) models.

The model utilizes two primary innovations:

  • Pseudo-Trajectory Distillation: A training method that teaches the model which tokens can be decoded confidently at early steps.
  • Entropy-Based Multi-Block Decoding: An inference strategy using a KV-cache refresh mechanism to maintain accuracy while maximizing parallelism.

Resources

Usage

For detailed usage instructions, evaluation scripts, and training code, please refer to the official GitHub repository. Since the model uses a custom architecture, ensure you have transformers==4.49.0 installed and use trust_remote_code=True when loading the model.

Citation

@article{arxiv'26:d3llm,
  title   = {d3LLM: Ultra-Fast Diffusion LLM using Pseudo-Trajectory Distillation},
  author  = {Yu-Yang Qian and Junda Su and Lanxiang Hu and Peiyuan Zhang and Zhijie Deng and Peng Zhao and Hao Zhang},
  journal = {ArXiv preprint},
  volume  = {arXiv:2601.07568},
  year    = {2026}
}
Downloads last month
146
Safetensors
Model size
8B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for d3LLM/d3LLM_Dream_Coder

Finetuned
(1)
this model

Dataset used to train d3LLM/d3LLM_Dream_Coder

Paper for d3LLM/d3LLM_Dream_Coder