--- license: bsd-2-clause language: - en base_model: - GSAI-ML/LLaDA-8B-Instruct library_name: transformers tags: - DLM - EvoToken - lora - text-generation --- # EvoTokenDLM LoRA adapter training from pretrained weights LLaDA-8B-Instruct Starting from the original MDLM (Masked Discrete Diffusion Language Model) LLaDA-8B-Instruct, we trained the EvoTokenDLM LoRA adapter using the **Continuous Trajectory Supervision** method. Our implementation replaces traditional hard binary masks with evolving soft token distributions. This allows EvoTokenDLM to facilitate a progressive transition from masked states to discrete outputs, effectively supporting revisable decoding. The method and its results are detailed in the paper: [Beyond Hard Masks: Progressive Token Evolution for Diffusion Language Models](https://arxiv.org/abs/2601.07351). ## How to Use ⚠️ **Important:** This is a LoRA adapter and requires the official EvoTokenDLM codebase for inference. For detailed instructions and code, please refer to the official GitHub repository: [EvoTokenDLM GitHub Repository](https://github.com/aim-uofa/EvoTokenDLM) ## Citation If you find this work helpful for your research, please cite: ```BibTeX @article{zhong2026beyond, title={Beyond Hard Masks: Progressive Token Evolution for Diffusion Language Models}, author={Zhong, Linhao and Wu, Linyu and Fang, Bozhen and Feng, Tianjian and Jing, Chenchen and Wang, Wen and Zhang, Jiaheng and Chen, Hao and Shen, Chunhua}, journal={arXiv preprint arXiv:2601.07351}, year={2026} } ```