菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-03
📄 Abstract - Efficient Self-Evaluation for Diffusion Language Models via Sequence Regeneration

Diffusion large language models (dLLMs) have recently attracted significant attention for their ability to enhance diversity, controllability, and parallelism. However, their non-sequential, bidirectionally masked generation makes quality assessment difficult, underscoring the need for effective self-evaluation. In this work, we propose DiSE, a simple yet effective self-evaluation confidence quantification method for dLLMs. DiSE quantifies confidence by computing the probability of regenerating the tokens in the entire generated sequence, given the full context. This method enables more efficient and reliable quality assessment by leveraging token regeneration probabilities, facilitating both likelihood estimation and robust uncertainty quantification. Building upon DiSE, we further introduce a flexible-length generation framework, which adaptively controls the sequence length based on the model's self-assessment of its own output. We analyze and validate the feasibility of DiSE from the perspective of dLLM generalization, and empirically demonstrate that DiSE is positively correlated with both semantic coherence and answer accuracy. Extensive experiments on likelihood evaluation, uncertainty quantification, and flexible-length generation further confirm the effectiveness of the proposed DiSE.

顶级标签: llm model evaluation natural language processing
详细标签: diffusion language models self-evaluation uncertainty quantification sequence regeneration flexible-length generation 或 搜索:

通过序列再生实现扩散语言模型的高效自我评估 / Efficient Self-Evaluation for Diffusion Language Models via Sequence Regeneration


1️⃣ 一句话总结

这篇论文提出了一种名为DiSE的新方法,它能让扩散大语言模型通过计算完整序列的再生概率来评估自身生成内容的质量和可信度,从而更高效地判断答案好坏并灵活控制生成长度。

源自 arXiv: 2603.02760