菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-04
📄 Abstract - CoRe: Context-Robust Remasking for Diffusion Language Models

Standard decoding in Masked Diffusion Models (MDMs) is hindered by context rigidity: tokens are retained based on transient high confidence, often ignoring that early predictions lack full context. This creates cascade effects where initial inconsistencies misguide the remaining generation. Existing revision strategies attempt to mitigate this by relying on static confidence scores, but these signals are inherently myopic; inconsistent tokens can appear confident to the model itself. We propose Context-Robust Remasking (CoRe), a training-free framework for inference-time revision. Rather than trusting static token probabilities, CoRe identifies context-brittle tokens by probing their sensitivity to targeted masked-context perturbations. We formalize revision as a robust optimization objective over context shifts and efficiently approximate this objective to prioritize unstable tokens for revision. On LLaDA-8B-Base, CoRe delivers consistent improvements across reasoning and code benchmarks, outperforming compute-matched baselines and improving MBPP by up to 9.2 percentage points.

顶级标签: natural language processing model training model evaluation
详细标签: diffusion language models inference-time revision context robustness decoding strategy masked diffusion 或 搜索:

CoRe:面向扩散语言模型的上下文鲁棒性重掩码方法 / CoRe: Context-Robust Remasking for Diffusion Language Models


1️⃣ 一句话总结

本文提出了一种无需额外训练、在推理时动态识别并修正因上下文信息不足而生成错误的文本片段的方法,显著提升了扩散语言模型在代码生成和推理任务上的表现。

源自 arXiv: 2602.04096