Fairness over Equality: Correcting Social Incentives in Asymmetric Sequential Social Dilemmas
AI 摘要
针对非对称社会困境,论文提出了一种考虑奖励范围和局部反馈的公平性学习方法。
主要贡献
- 提出了针对非对称社会困境的公平性定义
- 引入了基于agent的权重机制来处理不对称性
- 实现了局部社交反馈,无需全局信息共享
方法论
通过改进奖励函数和引入权重机制,在多智能体强化学习框架下训练智能体。
原文摘要
Sequential Social Dilemmas (SSDs) provide a key framework for studying how cooperation emerges when individual incentives conflict with collective welfare. In Multi-Agent Reinforcement Learning, these problems are often addressed by incorporating intrinsic drives that encourage prosocial or fair behavior. However, most existing methods assume that agents face identical incentives in the dilemma and require continuous access to global information about other agents to assess fairness. In this work, we introduce asymmetric variants of well-known SSD environments and examine how natural differences between agents influence cooperation dynamics. Our findings reveal that existing fairness-based methods struggle to adapt under asymmetric conditions by enforcing raw equality that wrongfully incentivize defection. To address this, we propose three modifications: (i) redefining fairness by accounting for agents' reward ranges, (ii) introducing an agent-based weighting mechanism to better handle inherent asymmetries, and (iii) localizing social feedback to make the methods effective under partial observability without requiring global information sharing. Experimental results show that in asymmetric scenarios, our method fosters faster emergence of cooperative policies compared to existing approaches, without sacrificing scalability or practicality.