File size: 1,150 Bytes
5c882e1
 
70730fe
5c882e1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5beaf72
5c882e1
 
 
 
 
661003b
5c882e1
283ff04
661003b
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
---
datasets:
- d3LLM/trajectory_data_llada_32
tags:
- diffusion
- text-generation
- fast-inference
- d3llm
pipeline_tag: text-generation
---


# d3LLM: Ultra-Fast Diffusion LLM using Pseudo-Trajectory Distillation πŸš€

## Model Description

**d3LLM-LLaDA** is an ultra-fast diffusion language model that achieves high generation speed while maintaining competitive performance. Built on the Dream architecture.

## Key Features

- πŸš€ High throughput: **5.0Γ— faster** than autoregressive models (Qwen-2.5-7B-it) on H100 GPU, **3.5Γ— faster** on A100 GPU. Achieves **288.73 tokens/s** on H100 (vs 57.32 for AR baseline) on GSM8K-CoT Dataset.
- πŸ“Š High AUP (Accuracy Under Parallelism) scores across benchmarks
- πŸ”§ Optimized for coding and math reasoning tasks

## Usage

For detailed usage instructions, evaluation scripts, training datasets, and training code, please refer to the official GitHub repository and our blog:

- πŸ‘‰ Code repo: **[https://github.com/hao-ai-lab/d3LLM](https://github.com/hao-ai-lab/d3LLM)**
- 🌐 Blog: **[https://hao-ai-lab.github.io/blogs/text-diffusion/](https://hao-ai-lab.github.io/blogs/text-diffusion/)**