菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-02
📄 Abstract - AR-MAP: Are Autoregressive Large Language Models Implicit Teachers for Diffusion Large Language Models?

Diffusion Large Language Models (DLLMs) have emerged as a powerful alternative to autoregressive models, enabling parallel token generation across multiple positions. However, preference alignment of DLLMs remains challenging due to high variance introduced by Evidence Lower Bound (ELBO)-based likelihood estimation. In this work, we propose AR-MAP, a novel transfer learning framework that leverages preference-aligned autoregressive LLMs (AR-LLMs) as implicit teachers for DLLM alignment. We reveal that DLLMs can effectively absorb alignment knowledge from AR-LLMs through simple weight scaling, exploiting the shared architectural structure between these divergent generation paradigms. Crucially, our approach circumvents the high variance and computational overhead of direct DLLM alignment and comprehensive experiments across diverse preference alignment tasks demonstrate that AR-MAP achieves competitive or superior performance compared to existing DLLM-specific alignment methods, achieving 69.08\% average score across all tasks and models. Our Code is available at this https URL.

顶级标签: llm model training natural language processing
详细标签: diffusion models preference alignment knowledge transfer autoregressive models weight scaling 或 搜索:

AR-MAP:自回归大语言模型是扩散大语言模型的隐式教师吗? / AR-MAP: Are Autoregressive Large Language Models Implicit Teachers for Diffusion Large Language Models?


1️⃣ 一句话总结

这篇论文提出了一种名为AR-MAP的新方法,它巧妙地利用已经训练好的、能更好理解人类偏好的自回归大语言模型作为‘老师’,通过简单的权重调整,来快速提升另一种并行生成模型(扩散大语言模型)的性能,从而高效解决了后者直接训练困难的问题。

源自 arXiv: 2602.02178