📄 论文总结
离散扩散模型的漏洞规避:确定性绕过采样壁垒 / Loopholing Discrete Diffusion: Deterministic Bypass of the Sampling Wall
1️⃣ 一句话总结
这篇论文提出了一种名为‘漏洞规避’的新方法,通过在离散扩散模型中引入确定性潜变量路径来保留分布信息,从而显著提升了文本生成的质量和连贯性,并在推理任务中取得了更好的表现。
Discrete diffusion models offer a promising alternative to autoregressive generation through parallel decoding, but they suffer from a sampling wall: once categorical sampling occurs, rich distributional information collapses into one-hot vectors and cannot be propagated across steps, forcing subsequent steps to operate with limited information. To mitigate this problem, we introduce Loopholing, a novel and simple mechanism that preserves this information via a deterministic latent pathway, leading to Loopholing Discrete Diffusion Models (LDDMs). Trained efficiently with a self-conditioning strategy, LDDMs achieve substantial gains-reducing generative perplexity by up to 61% over prior baselines, closing (and in some cases surpassing) the gap with autoregressive models, and producing more coherent text. Applied to reasoning tasks, LDDMs also improve performance on arithmetic benchmarks such as Countdown and Game of 24. These results also indicate that loopholing mitigates idle steps and oscillations, providing a scalable path toward high-quality non-autoregressive text generation.
离散扩散模型的漏洞规避:确定性绕过采样壁垒 / Loopholing Discrete Diffusion: Deterministic Bypass of the Sampling Wall
这篇论文提出了一种名为‘漏洞规避’的新方法,通过在离散扩散模型中引入确定性潜变量路径来保留分布信息,从而显著提升了文本生成的质量和连贯性,并在推理任务中取得了更好的表现。