d3LLM-model commited on
Commit
5c882e1
Β·
verified Β·
1 Parent(s): 3993cf9

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - d3LLM-Data-LLaDA/trajectory_data_llada_32
4
+ tags:
5
+ - diffusion
6
+ - text-generation
7
+ - fast-inference
8
+ - d3llm
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+
13
+ # d3LLM: Ultra-Fast Diffusion LLM using Pseudo-Trajectory Distillation πŸš€
14
+
15
+ ## Model Description
16
+
17
+ **d3LLM-LLaDA** is an ultra-fast diffusion language model that achieves high generation speed while maintaining competitive performance. Built on the Dream architecture.
18
+
19
+ ## Key Features
20
+
21
+ - πŸš€ **4.9Γ— faster** than autoregressive models (Qwen-2.5-7B-it) on H100 GPU
22
+ - 🎯 **3.5Γ— faster** on A100 GPU
23
+ - ⚑ **280.97 tokens/s** on H100 (vs 57.32 for AR baseline)
24
+ - πŸ“Š High AUP (Accuracy Under Parallelism) scores across benchmarks
25
+ - πŸ”§ Optimized for coding and math reasoning tasks
26
+
27
+ ## Usage
28
+
29
+ For detailed usage instructions, evaluation scripts, and training code, please refer to the official GitHub repository:
30
+
31
+ πŸ‘‰ **[https://github.com/hao-ai-lab/text-diffusion](https://github.com/hao-ai-lab/text-diffusion)**