LLM Reasoning 相关度: 9/10

Unlocking Reasoning Capability on Machine Translation in Large Language Models

Sara Rajaee, Sebastian Vincent, Alexandre Berard, Marzieh Fadaee, Kelly Marchisio, Tom Kocmi
arXiv: 2602.14763v1 发布: 2026-02-16 更新: 2026-02-16

AI 摘要

该论文研究了大型语言模型推理能力在机器翻译中的应用,并提出了针对机器翻译的结构化推理框架。

主要贡献

  • 发现通用推理在机器翻译中效果不佳
  • 提出了针对机器翻译的结构化推理框架
  • 构建了结构化推理轨迹的合成数据集

方法论

通过实验评估现有推理模型在机器翻译上的表现,并基于分析结果设计结构化推理框架,通过合成数据进行训练。

原文摘要

Reasoning-oriented large language models (RLMs) achieve strong gains on tasks such as mathematics and coding by generating explicit intermediate reasoning. However, their impact on machine translation (MT) remains underexplored. We systematically evaluate several open- and closed-weights RLMs on the WMT24++ benchmark and find that enabling explicit reasoning consistently degrades translation quality across languages and models. Analysis reveals that MT reasoning traces are highly linear, lacking revision, self-correction and exploration of alternative translations, which limits their usefulness. Furthermore, injecting higher-quality reasoning traces from stronger models does not reliably improve weaker models' performance. To address this mismatch, we propose a structured reasoning framework tailored to translation, based on multi-step drafting, adequacy refinement, fluency improvement, and selective iterative revision. We curate a synthetic dataset of dynamic structured reasoning traces and post-train a large reasoning model on this data. Experiments show significant improvements over standard translation fine-tuning and injected generic reasoning baselines. Our findings demonstrate that reasoning must be task-structured to benefit MT.

标签

machine translation large language models reasoning

arXiv 分类

cs.CL cs.AI