缓解视频大语言模型幻觉:放松锚定帧的主导地位 / Relaxing Anchor-Frame Dominance for Mitigating Hallucinations in Video Large Language Models
1️⃣ 一句话总结
这篇论文发现视频大语言模型在回答问题时,会过度依赖视频中的某一个关键帧(锚定帧),导致回答出现幻觉,并提出了一种无需额外训练、在解码阶段就能自动平衡各帧注意力权重的方法,有效减少了幻觉并保持了模型性能。
Recent Video Large Language Models (Video-LLMs) have demonstrated strong capability in video understanding, yet they still suffer from hallucinations. Existing mitigation methods typically rely on training, input modification, auxiliary guidance, or additional decoding procedures, while largely overlooking a more fundamental challenge. During generation, Video-LLMs tend to over-rely on a limited portion of temporal evidence, leading to temporally imbalanced evidence aggregation across the video. To address this issue, we investigate a decoder-side phenomenon in which the model exhibits a temporally imbalanced concentration pattern. We term the frame with the highest aggregated frame-level attention mass the anchor frame. We find that this bias is largely independent of the input video and instead appears to reflect a persistent, model-specific structural or positional bias, whose over-dominance is closely associated with hallucination-prone generation. Motivated by this insight, we propose Decoder-side Temporal Rebalancing (DTR), a training-free, layer-selective inference method that rebalances temporal evidence allocation in middle-to-late decoder layers without altering visual encoding or requiring auxiliary models. DTR adaptively calibrates decoder-side visual attention to alleviate temporally imbalanced concentration and encourage under-attended frames to contribute more effectively to response generation. In this way, DTR guides the decoder to ground its outputs in temporally broader and more balanced video evidence. Extensive experiments on hallucination and video understanding benchmarks show that DTR consistently improves hallucination robustness across diverse Video-LLM families, while preserving competitive video understanding performance and high inference efficiency.
缓解视频大语言模型幻觉:放松锚定帧的主导地位 / Relaxing Anchor-Frame Dominance for Mitigating Hallucinations in Video Large Language Models
这篇论文发现视频大语言模型在回答问题时,会过度依赖视频中的某一个关键帧(锚定帧),导致回答出现幻觉,并提出了一种无需额外训练、在解码阶段就能自动平衡各帧注意力权重的方法,有效减少了幻觉并保持了模型性能。
源自 arXiv: 2604.12582