Rethinking Role-Playing Evaluation: Anonymous Benchmarking and a Systematic Study of Personality Effects
AI 摘要
提出匿名评估方法,研究人格对角色扮演agent性能的影响,并验证了自生成人格的有效性。
主要贡献
- 提出了一种匿名角色扮演评估方法,消除了对预训练知识的依赖。
- 系统性地研究了人格特征对角色扮演agent性能的影响。
- 验证了自生成人格在提升角色扮演性能方面的有效性。
方法论
设计匿名角色扮演实验,对比人类标注和模型自生成的人格特征对agent性能的影响,使用多个benchmark评估。
原文摘要
Large language models (LLMs) have demonstrated significant potential in developing Role-Playing Agents (RPAs). However, current research primarily evaluates RPAs using famous fictional characters, allowing models to rely on memory associated with character names. This dependency creates a bias that limits the generalization of RPAs to unseen personas. To address this issue, we propose an anonymous evaluation method. Experiments across multiple benchmarks reveal that anonymization significantly degrades role-playing performance, confirming that name exposure carries implicit information. Furthermore, we investigate personality augmentation to enhance role fidelity under anonymous setting. We systematically compare the efficacy of personality traits derived from human annotations versus those self-generated by the model. Our results demonstrate that incorporating personality information consistently improves RPA performance. Crucially, self-generated personalities achieve performance comparable to human-annotated ones. This work establishes a fairer evaluation protocol and validates a scalable, personality-enhanced framework for constructing robust RPAs.