LLM Reasoning 相关度: 9/10

GPT-4o Lacks Core Features of Theory of Mind

John Muchovej, Amanda Royka, Shane Lee, Julian Jara-Ettinger
arXiv: 2602.12150v1 发布: 2026-02-12 更新: 2026-02-12

AI 摘要

GPT-4o在理论推理(ToM)的核心能力上存在缺陷,缺乏一致且泛化的心理状态行为模型。

主要贡献

  • 提出了评估LLM的ToM的新框架
  • 揭示了LLM在简单ToM任务上取得成功,但在逻辑等价任务上失败
  • 证明了LLM的社交能力并非源于领域泛化或一致的ToM

方法论

采用认知科学基础的ToM定义,设计测试框架,探究LLM是否具有一致的、领域泛化的心理状态行为因果模型。

原文摘要

Do Large Language Models (LLMs) possess a Theory of Mind (ToM)? Research into this question has focused on evaluating LLMs against benchmarks and found success across a range of social tasks. However, these evaluations do not test for the actual representations posited by ToM: namely, a causal model of mental states and behavior. Here, we use a cognitively-grounded definition of ToM to develop and test a new evaluation framework. Specifically, our approach probes whether LLMs have a coherent, domain-general, and consistent model of how mental states cause behavior -- regardless of whether that model matches a human-like ToM. We find that even though LLMs succeed in approximating human judgments in a simple ToM paradigm, they fail at a logically equivalent task and exhibit low consistency between their action predictions and corresponding mental state inferences. As such, these findings suggest that the social proficiency exhibited by LLMs is not the result of an domain-general or consistent ToM.

标签

Theory of Mind LLM evaluation Reasoning Causal Models Social Cognition

arXiv 分类

cs.AI cs.CL cs.LG