PICon: A Multi-Turn Interrogation Framework for Evaluating Persona Agent Consistency
AI 摘要
PICon提出了一种多轮提问框架,用于评估人格化AI代理的一致性。
主要贡献
- 提出了PICon评估框架,用于评估人格化AI代理的一致性
- 揭示了现有系统在一致性方面与人类基线存在差距
- 提供了评估人格化AI代理的实用方法论
方法论
通过设计逻辑链式多轮提问,从内部一致性、外部一致性和重测一致性三个维度评估AI代理。
原文摘要
Large language model (LLM)-based persona agents are rapidly being adopted as scalable proxies for human participants across diverse domains. Yet there is no systematic method for verifying whether a persona agent's responses remain free of contradictions and factual inaccuracies throughout an interaction. A principle from interrogation methodology offers a lens: no matter how elaborate a fabricated identity, systematic interrogation will expose its contradictions. We apply this principle to propose PICon, an evaluation framework that probes persona agents through logically chained multi-turn questioning. PICon evaluates consistency along three core dimensions: internal consistency (freedom from self-contradiction), external consistency (alignment with real-world facts), and retest consistency (stability under repetition). Evaluating seven groups of persona agents alongside 63 real human participants, we find that even systems previously reported as highly consistent fail to meet the human baseline across all three dimensions, revealing contradictions and evasive responses under chained questioning. This work provides both a conceptual foundation and a practical methodology for evaluating persona agents before trusting them as substitutes for human participants. We provide the source code and an interactive demo at: https://kaist-edlab.github.io/picon/