Skip to content
← Back to explorer

GPT-4o Lacks Core Features of Theory of Mind

John Muchovej, Amanda Royka, Shane Lee, Julian Jara-Ettinger · Feb 12, 2026 · Citations: 0

Abstract

Do Large Language Models (LLMs) possess a Theory of Mind (ToM)? Research into this question has focused on evaluating LLMs against benchmarks and found success across a range of social tasks. However, these evaluations do not test for the actual representations posited by ToM: namely, a causal model of mental states and behavior. Here, we use a cognitively-grounded definition of ToM to develop and test a new evaluation framework. Specifically, our approach probes whether LLMs have a coherent, domain-general, and consistent model of how mental states cause behavior -- regardless of whether that model matches a human-like ToM. We find that even though LLMs succeed in approximating human judgments in a simple ToM paradigm, they fail at a logically equivalent task and exhibit low consistency between their action predictions and corresponding mental state inferences. As such, these findings suggest that the social proficiency exhibited by LLMs is not the result of a domain-general or consistent ToM.

Human Data Lens

  • Uses human feedback: No
  • Feedback types: None
  • Rater population: Unknown
  • Unit of annotation: Unknown
  • Expertise required: General

Evaluation Lens

  • Evaluation modes: Automatic Metrics
  • Agentic eval: None
  • Quality controls: Not reported
  • Confidence: 0.30
  • Flags: low_signal, possible_false_positive

Research Summary

Contribution Summary

  • Do Large Language Models (LLMs) possess a Theory of Mind (ToM)?
  • Research into this question has focused on evaluating LLMs against benchmarks and found success across a range of social tasks.
  • However, these evaluations do not test for the actual representations posited by ToM: namely, a causal model of mental states and behavior.

Why It Matters For Eval

  • Research into this question has focused on evaluating LLMs against benchmarks and found success across a range of social tasks.
  • However, these evaluations do not test for the actual representations posited by ToM: namely, a causal model of mental states and behavior.

Related Papers