任务感知提升大语言模型的生成质量与不确定性估计 / Task-Awareness Improves LLM Generations and Uncertainty
1️⃣ 一句话总结
这篇论文提出了一种新方法,通过将大语言模型的输出映射到特定任务的结构化空间(如标签、数值或图表),并在此空间内进行最优合成与不确定性评估,从而显著提升了模型回答的准确性和可靠性。
In many applications of LLMs, natural language responses often have an underlying structure such as representing discrete labels, numerical values, or graphs. Yet, existing decoding and uncertainty estimation methods operate only in language space and largely disregard structural information. We address this by modeling LLM outputs directly in a task-dependent latent structure. By equipping this structure with a dissimilarity measure, we can compute Bayes-optimal responses. These are not selected from sampled generations but are newly synthesized by combining individual responses in the latent space. Across different tasks, Bayes-optimal responses consistently outperform standard decoding methods like beam search. Moreover, quantifying uncertainty via the induced Bayesian risk captures variations in terms of the latent structure and improves alignment with output quality and correctness. Our decision-theoretic framework is applicable to any problem that admits a latent response structure and enables reliable task-aware LLM predictions.
任务感知提升大语言模型的生成质量与不确定性估计 / Task-Awareness Improves LLM Generations and Uncertainty
这篇论文提出了一种新方法,通过将大语言模型的输出映射到特定任务的结构化空间(如标签、数值或图表),并在此空间内进行最优合成与不确定性评估,从而显著提升了模型回答的准确性和可靠性。
源自 arXiv: 2601.21500