File size: 1,994 Bytes
4371625
 
c8d3a40
4371625
 
 
c8d3a40
 
 
 
 
4371625
 
 
c8d3a40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
---
library_name: transformers
license: mit
pipeline_tag: text-generation
---

# LLaDA 1.5: Variance-Reduced Preference Optimization for Large Language Diffusion Models

We introduce LLaDA 1.5, a competitive large diffusion language model, trained by variance-reduced preference optimization (VRPO), as presented in the paper [LLaDA 1.5: Variance-Reduced Preference Optimization for Large Language Diffusion Models](https://huggingface.co/papers/2505.19223).

Compared with LLaDA-8B-Instruct, LLaDA 1.5 achieves better performance on a wide range of tasks, including Math, Code, and Alignment tasks.

[Project Page](https://ml-gsai.github.io/LLaDA-1.5-Demo/)

[Code](https://github.com/ML-GSAI/LLaDA-1.5)

<div style="display: flex; justify-content: center; align-items: center; width: 100%; margin: 0 auto;">
    <img src="https://github.com/ML-GSAI/LLaDA-1.5/raw/main/assets/llada_1_5.png" style="width: 50%; display: block; margin: 0 auto;" />
</div>

## Inference

The LLaDA 1.5 model is available on [Huggingface](https://huggingface.co/GSAI-ML/LLaDA-1.5). Please employ the [transformers](https://huggingface.co/docs/transformers/index) to load.

```python
from transformers import AutoModel, AutoTokenizer
import torch

tokenizer = AutoTokenizer.from_pretrained('GSAI-ML/LLaDA-1.5', trust_remote_code=True)
model = AutoModel.from_pretrained('GSAI-ML/LLaDA-1.5', trust_remote_code=True, torch_dtype=torch.bfloat16)
```

The model is based on LLaDA-8B-Instruct, you can use the code for [LLaDA-8B-Instruct](https://github.com/ML-GSAI/LLaDA/blob/main/generate.py) to inference.

## Citation

Please consider cite:

```bibtex
@article{zhu2025llada,
  title={LLaDA 1.5: Variance-Reduced Preference Optimization for Large Language Diffusion Models},
  author={Zhu, Fengqi and Wang, Rongzhen and Nie, Shen and Zhang, Xiaolu and Wu, Chunwei and Hu, Jun and Zhou, Jun and Chen, Jianfei and Lin, Yankai and Wen, Ji-Rong and others},
  journal={arXiv preprint arXiv:2505.19223},
  year={2025}
}
```