菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-04
📄 Abstract - Decomposed Prompting Does Not Fix Knowledge Gaps, But Helps Models Say "I Don't Know"

Large language models often struggle to recognize their knowledge limits in closed-book question answering, leading to confident hallucinations. While decomposed prompting is typically used to improve accuracy, we investigate its impact on reliability. We evaluate three task-equivalent prompting regimes: Direct, Assistive, and Incremental, across different model scales and multi-hop QA benchmarks. We find that although accuracy gains from decomposition diminish in frontier models, disagreements between prompting regimes remain highly indicative of potential errors. Because factual knowledge is stable while hallucinations are stochastic, cross-regime agreement provides a precise signal of internal uncertainty. We leverage this signal to implement a training-free abstention policy that requires no retrieval or fine-tuning. Our results show that disagreement-based abstention outperforms standard uncertainty baselines as an error detector, improving both F1 and AUROC across settings. This demonstrates that decomposition-based prompting can serve as a practical diagnostic probe for model reliability in closed-book QA.

顶级标签: llm model evaluation natural language processing
详细标签: decomposed prompting uncertainty estimation closed-book qa hallucination detection abstention policy 或 搜索:

分解式提示并不能弥补知识缺口,但能帮助模型说“我不知道” / Decomposed Prompting Does Not Fix Knowledge Gaps, But Helps Models Say "I Don't Know"


1️⃣ 一句话总结

这篇论文研究发现,通过比较大语言模型在直接、辅助和渐进式等不同分解提示策略下的回答差异,可以精准探测模型的内在不确定性,从而无需额外训练或检索就能有效让模型在闭卷问答中识别并承认自己不知道的问题,减少错误回答。

源自 arXiv: 2602.04853