Hengchang-Liu commited on
Commit
ab5ac6b
·
verified ·
1 Parent(s): aa61815

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -6
README.md CHANGED
@@ -35,13 +35,11 @@ Hengchang Liu, Zhao Yang, Bing Su
35
 
36
  ---
37
 
38
- ## 🧠 Method Overview
39
 
40
- CAL is a training-free framework that:
41
 
42
- * characterizes the *Oracle Peak* phenomenon in diffusion language models,
43
- * calibrates model confidence with a length bias function,
44
- * adaptively discovers near-optimal infilling lengths before decoding.
45
 
46
  The method is evaluated on multiple infilling benchmarks including:
47
 
@@ -66,6 +64,6 @@ If you find this work useful, please cite:
66
  eprint={2602.00476},
67
  archivePrefix={arXiv},
68
  primaryClass={cs.LG},
69
- url={https://arxiv.org/abs/2602.00476}
70
  }
71
  ```
 
35
 
36
  ---
37
 
38
+ ## 📝 Abstract (from the paper)
39
 
40
+ Diffusion language models (DLMs) provide a bidirectional generation framework naturally suited for infilling, yet their performance is constrained by the pre-specified infilling length. In this paper, we reveal that DLMs possess an inherent ability to discover the correct infilling length. We identify two key statistical phenomena in the first-step denoising confidence: a local *Oracle Peak* that emerges near the ground-truth length and a systematic *Length Bias* that often obscures this signal. By leveraging this signal and calibrating the bias, our training-free method **CAL** (**C**alibrated **A**daptive **L**ength) enables DLMs to approximate the optimal length through an efficient search before formal decoding. Empirical evaluations demonstrate that CAL improves Pass@1 by up to 47.7\% over fixed-length baselines and 40.5% over chat-based adaptive methods in code infilling, while boosting BLEU-2 and ROUGE-L by up to 8.5% and 9.9% in text infilling. These results demonstrate that CAL paves the way for robust DLM infilling without requiring any specialized training.
41
 
42
+ ## 🧠 Evaluation
 
 
43
 
44
  The method is evaluated on multiple infilling benchmarks including:
45
 
 
64
  eprint={2602.00476},
65
  archivePrefix={arXiv},
66
  primaryClass={cs.LG},
67
+ url={https://arxiv.org/abs/2602.00476},
68
  }
69
  ```