Gender Disambiguation in Machine Translation: Diagnostic Evaluation in Decoder-Only Architectures
AI 摘要
该论文研究Decoder-only模型在机器翻译中存在的性别偏见问题,并提出一种新的评估指标。
主要贡献
- 提出新的“Prior Bias”指标
- 评估Decoder-only模型在机器翻译中的性别偏见
- 研究指令微调对缓解性别偏见的影响
方法论
扩展现有偏见评估框架,引入Prior Bias指标,并在Decoder-only模型上进行实验分析。
原文摘要
While Large Language Models achieve state-of-the-art results across a wide range of NLP tasks, they remain prone to systematic biases. Among these, gender bias is particularly salient in MT, due to systematic differences across languages in whether and how gender is marked. As a result, translation often requires disambiguating implicit source signals into explicit gender-marked forms. In this context, standard benchmarks may capture broad disparities but fail to reflect the full complexity of gender bias in modern MT. In this paper, we extend recent frameworks on bias evaluation by: (i) introducing a novel measure coined "Prior Bias", capturing a model's default gender assumptions, and (ii) applying the framework to decoder-only MT models. Our results show that, despite their scale and state-of-the-art status, decoder-only models do not generally outperform encoder-decoder architectures on gender-specific metrics; however, post-training (e.g., instruction tuning) not only improves contextual awareness but also reduces the masculine Prior Bias.