菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-01-13
📄 Abstract - Entropy Sentinel: Continuous LLM Accuracy Monitoring from Decoding Entropy Traces in STEM

Deploying LLMs raises two coupled challenges: (1) monitoring - estimating where a model underperforms as traffic and domains drift - and (2) improvement - prioritizing data acquisition to close the largest performance gaps. We test whether an inference-time signal can estimate slice-level accuracy under domain shift. For each response, we compute an output-entropy profile from final-layer next-token probabilities (from top-k logprobs) and summarize it with eleven statistics. A lightweight classifier predicts instance correctness, and averaging predicted probabilities yields a domain-level accuracy estimate. We evaluate on ten STEM reasoning benchmarks with exhaustive train/test compositions (k in {1,2,3,4}; all "10 choose k" combinations), across nine LLMs from six families (3B-20B). Estimates often track held-out benchmark accuracy, and several models show near-monotonic ordering of domains. Output-entropy profiles are thus an accessible signal for scalable monitoring and for targeting data acquisition.

顶级标签: llm model evaluation machine learning
详细标签: accuracy monitoring decoding entropy domain shift stem reasoning performance estimation 或 搜索:

熵哨兵:基于解码熵迹对STEM领域大语言模型进行持续准确性监控 / Entropy Sentinel: Continuous LLM Accuracy Monitoring from Decoding Entropy Traces in STEM


1️⃣ 一句话总结

这篇论文提出了一种名为‘熵哨兵’的新方法,通过分析大语言模型生成答案时的不确定性(即输出熵),就能低成本、大规模地监控模型在不同科学领域的表现好坏,并指导我们优先收集哪些数据来提升模型性能。

源自 arXiv: 2601.09001