菜单

🤖 系统
📄 Abstract - KLASS: KL-Guided Fast Inference in Masked Diffusion Models

Masked diffusion models have demonstrated competitive results on various tasks including language generation. However, due to its iterative refinement process, the inference is often bottlenecked by slow and static sampling speed. To overcome this problem, we introduce `KL-Adaptive Stability Sampling' (KLASS), a fast yet effective sampling method that exploits token-level KL divergence to identify stable, high-confidence predictions. By unmasking multiple tokens in each iteration without any additional model training, our approach speeds up generation significantly while maintaining sample quality. On reasoning benchmarks, KLASS achieves up to $2.78\times$ wall-clock speedups while improving performance over standard greedy decoding, attaining state-of-the-art results among diffusion-based samplers. We further validate KLASS across diverse domains, including text, image, and molecular generation, showing its effectiveness as a broadly applicable sampler across different models.

顶级标签: model training model evaluation natural language processing
详细标签: diffusion models inference acceleration sampling methods kl divergence generation efficiency 或 搜索:

📄 论文总结

KLASS:基于KL引导的掩码扩散模型快速推理方法 / KLASS: KL-Guided Fast Inference in Masked Diffusion Models


1️⃣ 一句话总结

这篇论文提出了一种名为KLASS的快速采样方法,通过利用KL散度识别稳定预测,在不额外训练模型的情况下大幅加速掩码扩散模型的生成过程,并在文本、图像和分子生成等多个领域保持甚至提升了生成质量。


📄 打开原文 PDF