Beyond Hard Masks: Progressive Token Evolution for Diffusion Language Models
Abstract
EvoToken-DLM introduces a diffusion-based language modeling approach that uses soft token distributions and continuous trajectory supervision to enable revisable decoding and outperforms existing baselines.
Diffusion Language Models (DLMs) offer a promising alternative for language modeling by enabling parallel decoding through iterative refinement. However, most DLMs rely on hard binary masking and discrete token assignments, which hinder the revision of early decisions and underutilize intermediate probabilistic representations. In this paper, we propose EvoToken-DLM, a novel diffusion-based language modeling approach that replaces hard binary masks with evolving soft token distributions. EvoToken-DLM enables a progressive transition from masked states to discrete outputs, supporting revisable decoding. To effectively support this evolution, we introduce continuous trajectory supervision, which aligns training objectives with iterative probabilistic updates. Extensive experiments across multiple benchmarks show that EvoToken-DLM consistently achieves superior performance, outperforming strong diffusion-based and masked DLM baselines. Project webpage: https://aim-uofa.github.io/EvoTokenDLM.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- CDLM: Consistency Diffusion Language Models For Faster Sampling (2025)
- Deferred Commitment Decoding for Diffusion Language Models with Confidence-Aware Sliding Windows (2026)
- CD4LM: Consistency Distillation and aDaptive Decoding for Diffusion Language Models (2026)
- Context-Aware Initialization for Reducing Generative Path Length in Diffusion Language Models (2025)
- WeDLM: Reconciling Diffusion Language Models with Standard Causal Attention for Fast Inference (2025)
- DIP: Dynamic In-Context Planner For Diffusion Language Models (2026)
- Breaking the Bottleneck with DiffuApriel: High-Throughput Diffusion LMs with Mamba Backbone (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper