Cross-Preference Learning for Sentence-Level and Context-Aware Machine Translation
AI 摘要
提出Cross-Preference Learning框架,通过显式建模sentence-level和context-aware翻译的互补优势,提升机器翻译质量。
主要贡献
- 提出Cross-Preference Learning (CPL)框架
- 引入intra- and cross-condition preferences优化目标
- 无需架构修改即可提升多种模型在上下文机器翻译任务上的性能
方法论
CPL利用基于偏好的训练框架,整合sentence-level和context-aware翻译的偏好,通过显式监督学习如何利用上下文信息。
原文摘要
Context-aware machine translation (MT) leverages document-level information, yet it does not consistently outperform sentence-level MT, as contextual signals are unevenly beneficial across sentences. Existing training objectives do not explicitly model this variability, limiting a model's ability to adaptively exploit context. In this paper, we propose Cross-Preference Learning (CPL), a preference-based training framework that explicitly captures the complementary benefits of sentence-level and context-aware MT. CPL achieves this by integrating both intra- and cross-condition preferences into the preference optimization objective. The introduction of intra- and cross-condition preferences provides explicit supervision on when and how contextual information improves translation quality. We validate the proposed approach on several public context-aware MT tasks using multiple models, including Qwen3-4B, Qwen3-8B, and Llama-3-8B. Experimental results demonstrate consistent improvements in translation quality and robustness across both input conditions, achieved without any architectural modifications.