📄
Abstract - ExPosST: Explicit Positioning with Adaptive Masking for LLM-Based Simultaneous Machine Translation
Large language models (LLMs) have recently demonstrated promising performance in simultaneous machine translation (SimulMT). However, applying decoder-only LLMs to SimulMT introduces a positional mismatch, which leads to a dilemma between decoding efficiency and positional consistency. Existing approaches often rely on specific positional encodings or carefully designed prompting schemes, and thus fail to simultaneously achieve inference efficiency, positional consistency, and broad model compatibility. In this work, we propose ExPosST, a general framework that resolves this dilemma through explicit position allocation. ExPosST reserves fixed positional slots for incoming source tokens, enabling efficient decoding with KV cache across different positional encoding methods. To further bridge the gap between fine-tuning and inference, we introduce a policy-consistent fine-tuning strategy that aligns training with inference-time decoding behavior. Experiments across multiple language pairs demonstrate that ExPosST effectively supports simultaneous translation under diverse policies.
ExPosST:基于大语言模型的同声传译显式定位与自适应掩码框架 /
ExPosST: Explicit Positioning with Adaptive Masking for LLM-Based Simultaneous Machine Translation
1️⃣ 一句话总结
这篇论文提出了一个名为ExPosST的新框架,通过为输入源语言词分配固定位置并采用策略一致的微调方法,解决了大语言模型在同声传译任务中解码效率与位置一致性难以兼顾的难题,从而在多种语言对上都实现了高效且准确的实时翻译。