通过解耦信息瓶颈缓解多语言LLM作为评判者时的翻译腔偏见 / Mitigating Translationese Bias in Multilingual LLM-as-a-Judge via Disentangled Information Bottleneck
1️⃣ 一句话总结
这篇论文提出了一种名为DIBJudge的微调框架,通过信息压缩和解耦技术,有效减少了大语言模型在多语言评估中普遍存在的、倾向于机器翻译文本而非人工参考译文的系统性偏见。
Large language models (LLMs) have become a standard for multilingual evaluation, yet they exhibit a severe systematic translationese bias. In this paper, translationese bias is characterized as LLMs systematically favoring machine-translated text over human-authored references, particularly in low-resource languages. We attribute this bias to spurious correlations with (i) latent manifold alignment with English and (ii) cross-lingual predictability. To mitigate this bias, we propose DIBJudge, a robust fine-tuning framework that learns a minimally sufficient, judgment-critical representation via variational information compression, while explicitly isolating spurious factors into the dedicated bias branch. Furthermore, we incorporate a cross-covariance penalty that explicitly suppresses statistical dependence between robust and bias representations, thereby encouraging effective disentanglement. Extensive evaluations on multilingual reward modeling benchmarks and a dedicated translationese bias evaluation suite demonstrate that the proposed DIBJudge consistently outperforms strong baselines and substantially mitigates translationese bias.
通过解耦信息瓶颈缓解多语言LLM作为评判者时的翻译腔偏见 / Mitigating Translationese Bias in Multilingual LLM-as-a-Judge via Disentangled Information Bottleneck
这篇论文提出了一种名为DIBJudge的微调框架,通过信息压缩和解耦技术,有效减少了大语言模型在多语言评估中普遍存在的、倾向于机器翻译文本而非人工参考译文的系统性偏见。
源自 arXiv: 2603.10351