菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-03-11
📄 Abstract - Mitigating Translationese Bias in Multilingual LLM-as-a-Judge via Disentangled Information Bottleneck

Large language models (LLMs) have become a standard for multilingual evaluation, yet they exhibit a severe systematic translationese bias. In this paper, translationese bias is characterized as LLMs systematically favoring machine-translated text over human-authored references, particularly in low-resource languages. We attribute this bias to spurious correlations with (i) latent manifold alignment with English and (ii) cross-lingual predictability. To mitigate this bias, we propose DIBJudge, a robust fine-tuning framework that learns a minimally sufficient, judgment-critical representation via variational information compression, while explicitly isolating spurious factors into the dedicated bias branch. Furthermore, we incorporate a cross-covariance penalty that explicitly suppresses statistical dependence between robust and bias representations, thereby encouraging effective disentanglement. Extensive evaluations on multilingual reward modeling benchmarks and a dedicated translationese bias evaluation suite demonstrate that the proposed DIBJudge consistently outperforms strong baselines and substantially mitigates translationese bias.

顶级标签: llm model evaluation natural language processing
详细标签: translationese bias multilingual evaluation information bottleneck disentangled representation reward modeling 或 搜索:

通过解耦信息瓶颈缓解多语言LLM作为评判者时的翻译腔偏见 / Mitigating Translationese Bias in Multilingual LLM-as-a-Judge via Disentangled Information Bottleneck


1️⃣ 一句话总结

这篇论文提出了一种名为DIBJudge的微调框架,通过信息压缩和解耦技术,有效减少了大语言模型在多语言评估中普遍存在的、倾向于机器翻译文本而非人工参考译文的系统性偏见。

源自 arXiv: 2603.10351