菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-10
📄 Abstract - Breaking the Pre-Sampling Barrier: Activation-Informed Difficulty-Aware Self-Consistency

Self-Consistency (SC) is an effective decoding strategy that improves the reasoning performance of Large Language Models (LLMs) by generating multiple chain-of-thought reasoning paths and selecting the final answer via majority voting. However, it suffers from substantial inference costs because it requires a large number of samples. To mitigate this issue, Difficulty-Adaptive Self-Consistency (DSC) was proposed to reduce unnecessary token usage for easy problems by adjusting the number of samples according to problem difficulty. However, DSC requires additional model calls and pre-sampling to estimate difficulty, and this process is repeated when applying to each dataset, leading to significant computational overhead. In this work, we propose Activation-Informed Difficulty-Aware Self-Consistency (ACTSC) to address these limitations. ACTSC leverages internal difficulty signals reflected in the feed-forward network neuron activations to construct a lightweight difficulty estimation probe, without any additional token generation or model calls. The probe dynamically adjusts the number of samples for SC and can be applied to new datasets without requiring pre-sampling for difficulty estimation. To validate its effectiveness, we conduct experiments on five benchmarks. Experimental results show that ACTSC effectively reduces inference costs while maintaining accuracy relative to existing methods.

顶级标签: llm model evaluation natural language processing
详细标签: self-consistency efficient inference difficulty estimation activation analysis decoding strategy 或 搜索:

突破预采样障碍:基于激活信息感知难度的自一致性方法 / Breaking the Pre-Sampling Barrier: Activation-Informed Difficulty-Aware Self-Consistency


1️⃣ 一句话总结

这篇论文提出了一种名为ACTSC的新方法,它通过分析大语言模型内部的神经元激活信号来智能判断问题的难度,从而动态调整推理路径的生成数量,在保持答案准确性的同时,显著降低了自一致性解码策略的计算开销,且无需预先采样。

源自 arXiv: 2602.09438