Agent Tuning & Optimization 相关度: 9/10

Prompt Programming for Cultural Bias and Alignment of Large Language Models

Maksim Eren, Eric Michalak, Brian Cook, Johnny Seales
arXiv: 2603.16827v1 发布: 2026-03-17 更新: 2026-03-17

AI 摘要

该论文研究了大型语言模型中的文化偏见问题,并提出利用DSPy进行提示编程以优化文化对齐。

主要贡献

  • 验证并扩展了文化对齐框架
  • 引入了基于DSPy的提示编程方法
  • 实验证明提示优化优于人工提示工程

方法论

使用社会科学调查投影和距离指标评估文化偏差,并利用DSPy优化提示以减少偏差。

原文摘要

Culture shapes reasoning, values, prioritization, and strategic decision-making, yet large language models (LLMs) often exhibit cultural biases that misalign with target populations. As LLMs are increasingly used for strategic decision-making, policy support, and document engineering tasks such as summarization, categorization, and compliance-oriented auditing, improving cultural alignment is important for ensuring that downstream analyses and recommendations reflect target-population value profiles rather than default model priors. Previous work introduced a survey-grounded cultural alignment framework and showed that culture-specific prompting can reduce misalignment, but it primarily evaluated proprietary models and relied on manual prompt engineering. In this paper, we validate and extend that framework by reproducing its social sciences survey based projection and distance metrics on open-weight LLMs, testing whether the same cultural skew and benefits of culture conditioning persist outside closed LLM systems. Building on this foundation, we introduce use of prompt programming with DSPy for this problem-treating prompts as modular, optimizable programs-to systematically tune cultural conditioning by optimizing against cultural-distance objectives. In our experiments, we show that prompt optimization often improves upon cultural prompt engineering, suggesting prompt compilation with DSPy can provide a more stable and transferable route to culturally aligned LLM responses.

标签

文化偏见 大型语言模型 提示编程 DSPy 文化对齐

arXiv 分类

cs.AI cs.CL