菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-02-11
📄 Abstract - Towards Reliable Machine Translation: Scaling LLMs for Critical Error Detection and Safety

Machine Translation (MT) plays a pivotal role in cross-lingual information access, public policy communication, and equitable knowledge dissemination. However, critical meaning errors, such as factual distortions, intent reversals, or biased translations, can undermine the reliability, fairness, and safety of multilingual systems. In this work, we explore the capacity of instruction-tuned Large Language Models (LLMs) to detect such critical errors, evaluating models across a range of parameters using the publicly accessible data sets. Our findings show that model scaling and adaptation strategies (zero-shot, few-shot, fine-tuning) yield consistent improvements, outperforming encoder-only baselines like XLM-R and ModernBERT. We argue that improving critical error detection in MT contributes to safer, more trustworthy, and socially accountable information systems by reducing the risk of disinformation, miscommunication, and linguistic harm, especially in high-stakes or underrepresented contexts. This work positions error detection not merely as a technical challenge, but as a necessary safeguard in the pursuit of just and responsible multilingual AI. The code will be made available at GitHub.

顶级标签: llm natural language processing model evaluation
详细标签: machine translation error detection safety multilingual model scaling 或 搜索:

迈向可靠的机器翻译:利用大语言模型扩展关键错误检测与安全性 / Towards Reliable Machine Translation: Scaling LLMs for Critical Error Detection and Safety


1️⃣ 一句话总结

这篇论文研究了如何利用指令微调的大语言模型来检测机器翻译中可能导致事实扭曲或意图反转的关键错误,发现扩大模型规模和采用合适的适应策略能有效提升检测性能,从而有助于构建更安全、可信的多语言信息系统。

源自 arXiv: 2602.11444