菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-13
📄 Abstract - Reducing Hallucination in Enterprise AI Workflows via Hybrid Utility Minimum Bayes Risk (HUMBR)

Although LLMs drive automation, it is critical to ensure immense consideration for high-stakes enterprise workflows such as those involving legal matters, risk management, and privacy compliance. For Meta, and other organizations like ours, a single hallucinated clause in such high stakes workflows risks material consequences. We show that by framing hallucination mitigation as a Minimum Bayes Risk (MBR) problem, we can dramatically reduce this risk. Specifically, we introduce a Hybrid Utility MBR (HUMBR) framework that synthesizes semantic embedding similarity with lexical precision to identify consensus without ground-truth references, for which we derive rigorous error bounds. We complement this theoretical analysis with a comprehensive empirical evaluation on widely-used public benchmark suites (TruthfulQA and LegalBench) and also real world data from Meta production deployment. The results from our empirical study show that MBR significantly outperforms standard Universal Self-Consistency. Notably, 81% of the pipeline's suggestions were preferred over human-crafted ground truth, and critical recall failures were virtually eliminated.

顶级标签: llm model evaluation systems
详细标签: hallucination mitigation minimum bayes risk enterprise workflows error bounds utility function 或 搜索:

通过混合效用最小贝叶斯风险(HUMBR)减少企业AI工作流中的幻觉 / Reducing Hallucination in Enterprise AI Workflows via Hybrid Utility Minimum Bayes Risk (HUMBR)


1️⃣ 一句话总结

这篇论文提出了一种名为HUMBR的新方法,通过结合语义相似性和词汇精确度来识别AI生成内容中的共识,从而显著减少大语言模型在企业高风险工作流程(如法律和合规)中产生错误或虚构信息的风险,实验证明其效果优于现有标准方法。

源自 arXiv: 2604.11141