Agent Tuning & Optimization 相关度: 7/10

Cross-Preference Learning for Sentence-Level and Context-Aware Machine Translation

Ying Li, Xinglin Lyu, Junhui Li, Jinlong Yang, Hengchao Shang, Min Zhang, Shimin Tao, Daimeng Wei
arXiv: 2603.25183v1 发布: 2026-03-26 更新: 2026-03-26

AI 摘要

提出Cross-Preference Learning框架,通过显式建模sentence-level和context-aware翻译的互补优势,提升机器翻译质量。

主要贡献

  • 提出Cross-Preference Learning (CPL)框架
  • 引入intra- and cross-condition preferences优化目标
  • 无需架构修改即可提升多种模型在上下文机器翻译任务上的性能

方法论

CPL利用基于偏好的训练框架,整合sentence-level和context-aware翻译的偏好,通过显式监督学习如何利用上下文信息。

原文摘要

Context-aware machine translation (MT) leverages document-level information, yet it does not consistently outperform sentence-level MT, as contextual signals are unevenly beneficial across sentences. Existing training objectives do not explicitly model this variability, limiting a model's ability to adaptively exploit context. In this paper, we propose Cross-Preference Learning (CPL), a preference-based training framework that explicitly captures the complementary benefits of sentence-level and context-aware MT. CPL achieves this by integrating both intra- and cross-condition preferences into the preference optimization objective. The introduction of intra- and cross-condition preferences provides explicit supervision on when and how contextual information improves translation quality. We validate the proposed approach on several public context-aware MT tasks using multiple models, including Qwen3-4B, Qwen3-8B, and Llama-3-8B. Experimental results demonstrate consistent improvements in translation quality and robustness across both input conditions, achieved without any architectural modifications.

标签

机器翻译 上下文感知 偏好学习 自然语言处理

arXiv 分类

cs.CL