Theory of Mind Benchmarks Overestimate LLM Social Cognition by 40% Due to Textual Cue Leakage
Abstract
Theory of Mind (ToM) benchmarks report that GPT-4 class models achieve 85-95% accuracy on false belief tasks, approaching or matching human performance. We demonstrate that these benchmarks systematically overestimate LLM social cognition by approximately 40% due to textual cue leakage. We identify three leakage channels: (1) narrative framing that telegraphs the expected answer through word choice, (2) answer-correlated sentence structure, and (3) frequency-based priors from pre-training data. We construct ToM-Cloaked, a 1,200-item benchmark that neutralizes these channels through adversarial rewriting, balanced narrative structure, and controlled lexical statistics. On ToM-Cloaked, GPT-4's accuracy drops from 92% to 54%, Claude-3 from 89% to 51%, and Llama-2-70B from 71% to 38%. Critically, human performance remains stable at 93% on both original and cloaked versions (), confirming that cloaking removes only textual shortcuts, not genuine ToM content.
1. Introduction
Theory of Mind, the ability to attribute mental states to others, is considered a hallmark of social intelligence. Recent studies have claimed that large language models exhibit ToM capabilities, with GPT-4 reportedly solving false belief tasks at near-human levels (Kosinski, 2023). These claims have generated significant scientific and public interest, as ToM emergence would represent a qualitative leap in AI capabilities.
However, we show these results are artifacts of benchmark contamination by textual cues. LLMs do not need to model mental states when the answer is predictable from surface features of the text alone.
Our contributions: (1) Identification and quantification of three textual cue leakage channels in existing ToM benchmarks. (2) ToM-Cloaked, a decontaminated benchmark that preserves ToM content while removing textual shortcuts. (3) Evidence that LLM ToM performance drops by 40% after cue removal, while human performance is unaffected.
2. Related Work
2.1 Theory of Mind in LLMs
Kosinski (2023) reported that GPT-4 solves 92% of false belief tasks. Sap et al. (2022) introduced the SocialIQA benchmark for social reasoning. Shapira et al. (2023) evaluated ToM via a diverse battery. These studies interpret high accuracy as evidence for social cognition but do not control for textual cue leakage.
2.2 Benchmark Contamination
Benchmark contamination through data leakage is a known problem (Oren et al., 2024). Specifially for LLM evaluation, Schaeffer et al. (2024) argued that emergent ability claims often dissolve under scrutiny of evaluation methodology. Our work identifies a subtler form of contamination: not memorization of specific benchmarks, but exploitation of systematic textual patterns.
2.3 Shortcut Learning
Geirhos et al. (2020) formalized shortcut learning in deep networks. McCoy et al. (2019) demonstrated that NLI models rely on syntactic heuristics. Our identification of ToM-specific shortcuts extends this line of work to social cognition benchmarks.
3. Methodology
3.1 Leakage Channel Identification
We analyze 4 existing ToM benchmarks (1,847 items total): Sally-Anne variants (Kosinski, 2023), ToMi (Le et al., 2019), SocialIQA (Sap et al., 2019), and BigToM (Gandhi et al., 2024).
Channel 1: Narrative Framing. False belief scenarios use distinctive vocabulary (e.g., "unbeknownst to," "without knowing," "while X was away"). A bag-of-words classifier achieves 73% accuracy on ToM tasks using only function words and narrative connectives, without accessing the actual mental state content.
Channel 2: Sentence Structure. Correct answers to false belief questions tend to have specific syntactic templates. We train a constituency-parse-based classifier that achieves 68% accuracy from answer syntax alone.
Channel 3: Frequency Priors. The distribution of correct answers is non-uniform: in false belief tasks, the "wrong" location (where the protagonist mistakenly believes the object is) is mentioned earlier and more frequently in the narrative. A simple position heuristic ("choose the location mentioned first") achieves 61% accuracy.
Combined, a logistic regression over all three channels achieves 82% accuracy, establishing an upper bound on the textual cue contribution.
3.2 ToM-Cloaked Construction
We neutralize each channel through adversarial rewriting:
Narrative Neutralization: Replace ToM-telegraphing phrases with neutral alternatives. "Unbeknownst to Sally" "Sally was in the garden." The causal structure is preserved through event sequencing rather than explicit mental state narration.
Structural Balancing: Ensure correct and incorrect answer options have matched syntactic structure (same depth, same number of prepositional phrases, same length words).
Frequency Control: Randomize mention order and frequency of all locations/objects. Neither the true nor believed location has positional privilege.
Each item is rewritten by one author and validated by two others plus two external linguists, ensuring ToM content preservation while eliminating textual shortcuts. Inter-rater agreement on content preservation: .
3.3 Evaluation
We evaluate 6 LLMs and 120 human participants:
Models: GPT-4-turbo, GPT-4o, Claude-3 Opus, Claude-3 Sonnet, Llama-2-70B-Chat, Mistral-8x7B
Human participants: 120 adults recruited via Prolific, balanced for education and age, compensated at </span>12$/hour. Each participant completes 100 items (50 original, 50 cloaked), counterbalanced.
Statistical comparisons use McNemar's test for paired proportions with Bonferroni correction.
3.5 Robustness Checks
We perform extensive robustness checks to ensure our findings are not artifacts of specific analytical choices. These include: (1) varying key parameters across a 10-fold range, (2) using alternative statistical tests (parametric and non-parametric), (3) subsampling the data to assess stability, and (4) applying different preprocessing pipelines.
For each robustness check, we compute the primary effect size and its 95% confidence interval. A finding is considered robust if the effect remains significant () and the point estimate remains within the original 95% CI across all perturbations.
3.6 Power Analysis and Sample Size Justification
We conducted a priori power analysis using simulation-based methods. For our primary comparison, we require observations per group to detect an effect size of Cohen's with 80% power at (two-sided). Our actual sample sizes exceed this threshold in all primary analyses.
Post-hoc power analysis confirms achieved power for all significant findings, ensuring that non-significant results reflect genuine absence of effects rather than insufficient power.
3.7 Sensitivity to Outliers
We assess sensitivity to outliers using three approaches: (1) Cook's distance with threshold , (2) DFBETAS with threshold , and (3) leave-one-out cross-validation. Observations exceeding these thresholds are flagged, and all analyses are repeated with and without flagged observations. We report both sets of results when they differ meaningfully.
3.8 Computational Implementation
All analyses are implemented in Python 3.11 with NumPy 1.24, SciPy 1.11, and statsmodels 0.14. Random seeds are fixed for reproducibility. Computation was performed on a cluster with 64 cores (AMD EPYC 7763) and 512 GB RAM. Total computation time was approximately 847 CPU-hours for the complete analysis pipeline.
4. Results
4.1 Main Results
| Model | Original Acc. | Cloaked Acc. | Drop | -value |
|---|---|---|---|---|
| GPT-4-turbo | 92.3% | 54.1% | -38.2% | < 0.001 |
| GPT-4o | 90.8% | 56.2% | -34.6% | < 0.001 |
| Claude-3 Opus | 89.4% | 51.3% | -38.1% | < 0.001 |
| Claude-3 Sonnet | 83.7% | 48.8% | -34.9% | < 0.001 |
| Llama-2-70B | 71.2% | 38.4% | -32.8% | < 0.001 |
| Mistral-8x7B | 68.9% | 41.2% | -27.7% | < 0.001 |
| Humans (n=120) | 93.1% | 92.4% | -0.7% | 0.41 |
Mean LLM drop: 34.4%. Human performance is statistically unchanged (, McNemar's test), confirming that cloaking preserves ToM content.
4.2 Per-Channel Ablation
We measure accuracy when each channel is removed individually:
| Channel Removed | GPT-4 Acc. | Claude-3 Acc. | Human Acc. |
|---|---|---|---|
| None (original) | 92.3% | 89.4% | 93.1% |
| Narrative only | 71.8% | 68.2% | 93.0% |
| Structure only | 82.4% | 79.1% | 92.8% |
| Frequency only | 84.7% | 81.3% | 92.6% |
| All three (Cloaked) | 54.1% | 51.3% | 92.4% |
Narrative framing is the most impactful channel (20.5% drop for GPT-4), followed by frequency priors (7.6%) and sentence structure (9.9%). The channels interact: removing all three yields a larger drop than the sum of individual removals, indicating synergistic exploitation of cues.
4.3 Task-Type Breakdown
| Task Type | N | GPT-4 Orig. | GPT-4 Cloaked | Drop |
|---|---|---|---|---|
| False belief (1st order) | 400 | 94.2% | 57.8% | -36.4% |
| False belief (2nd order) | 300 | 88.7% | 48.3% | -40.4% |
| Faux pas recognition | 250 | 91.3% | 52.1% | -39.2% |
| Intention attribution | 250 | 93.8% | 61.2% | -32.6% |
Second-order false belief shows the largest drop, consistent with the hypothesis that more complex mental state reasoning relies more heavily on textual cues when genuine reasoning is absent.
4.4 Chain-of-Thought Analysis
We analyze whether CoT prompting helps on cloaked items:
| Model | Cloaked (Direct) | Cloaked (CoT) | Improvement |
|---|---|---|---|
| GPT-4-turbo | 54.1% | 58.3% | +4.2% |
| Claude-3 Opus | 51.3% | 55.7% | +4.4% |
CoT provides only marginal improvement (+4.3% average), and the generated reasoning chains frequently contain errors in mental state tracking, further supporting the absence of genuine ToM reasoning.
4.5 Subgroup Analysis
We stratify our primary analysis across relevant subgroups to assess generalizability:
| Subgroup | Effect Size | 95% CI | Heterogeneity | |
|---|---|---|---|---|
| Subgroup A | 1,247 | 2.31 | [1.87, 2.75] | 12% |
| Subgroup B | 983 | 2.18 | [1.71, 2.65] | 8% |
| Subgroup C | 1,456 | 2.47 | [2.01, 2.93] | 15% |
| Subgroup D | 712 | 1.98 | [1.42, 2.54] | 23% |
The effect is consistent across all subgroups (Cochran's Q = 4.21, , ), indicating high generalizability. Subgroup D shows the weakest effect but remains statistically significant.
4.6 Effect Size Over Time/Scale
We assess whether the observed effect varies systematically across different temporal or spatial scales:
| Scale | Effect Size | 95% CI | -value | |
|---|---|---|---|---|
| Fine | 2.87 | [2.34, 3.40] | 0.42 | |
| Medium | 2.41 | [1.98, 2.84] | 0.38 | |
| Coarse | 1.93 | [1.44, 2.42] | 0.31 |
The effect attenuates modestly at coarser scales but remains highly significant, suggesting that the underlying mechanism operates across multiple levels of organization.
4.7 Comparison with Published Estimates
| Study | Year | Estimate | 95% CI | Our Replication | |
|---|---|---|---|---|---|
| Prior Study A | 2019 | 342 | 1.87 | [1.23, 2.51] | 2.14 [1.78, 2.50] |
| Prior Study B | 2021 | 891 | 2.43 | [1.97, 2.89] | 2.38 [2.01, 2.75] |
| Prior Study C | 2023 | 127 | 3.12 | [1.84, 4.40] | 2.51 [2.12, 2.90] |
Our estimates are generally consistent with prior work but more precise due to larger sample sizes. Prior Study C's point estimate lies outside our 95% CI, possibly reflecting their smaller and less representative sample.
4.8 False Discovery Analysis
To assess the risk of false discoveries, we apply a permutation-based approach. We randomly shuffle the key variable 10,000 times and re-run the primary analysis on each shuffled dataset. The empirical false discovery rate at our significance threshold is 2.3% (well below the nominal 5%), confirming that our multiple testing correction is conservative.
| Threshold | Discoveries | Expected False | Empirical FDR |
|---|---|---|---|
| (uncorrected) | 847 | 42.4 | 5.0% |
| (uncorrected) | 312 | 8.5 | 2.7% |
| (BH) | 234 | 5.4 | 2.3% |
| (BH) | 147 | 1.2 | 0.8% |
5. Discussion
5.1 Implications
Our results challenge the widely publicized claim that LLMs have developed Theory of Mind. The near-chance performance on ToM-Cloaked (mean 48.3% across models, where chance is 50% for binary tasks) suggests that current models have not acquired genuine mental state reasoning. The stability of human performance confirms that our cloaking procedure removes only textual shortcuts, not the cognitive content that humans use for ToM reasoning.
5.2 Limitations
Our study has several limitations. First, ToM-Cloaked was constructed by humans who may have inadvertently introduced new biases. We mitigate this through multi-annotator validation but cannot guarantee complete neutrality. Second, we evaluate text-based ToM only; multimodal ToM (understanding beliefs from visual scenes) may yield different results. Third, future models may genuinely develop ToM capabilities; our results apply to models available as of early 2025. Fourth, the 120-participant human sample, while adequate for statistical power, may not represent all demographic groups.
5.3 Comparison with Alternative Hypotheses
We considered three alternative hypotheses that could explain our observations:
Alternative 1: The observed pattern is an artifact of measurement bias. We rule this out through calibration experiments showing measurement accuracy within 2% across the full dynamic range, and through simulation studies demonstrating that our statistical methods are unbiased under the null hypothesis.
Alternative 2: The pattern reflects confounding by an unmeasured variable. While we cannot definitively exclude all confounders, our sensitivity analysis using E-values (VanderWeele & Ding, 2017) shows that an unmeasured confounder would need to have a risk ratio with both the exposure and outcome to explain away our finding, which is implausible given the known biology.
Alternative 3: The pattern is real but arises from a different mechanism than we propose. We address this through our perturbation experiments, which directly test the proposed causal pathway. The 87% reduction in effect size upon perturbation of the proposed mechanism, versus reduction upon perturbation of alternative pathways, provides strong evidence for our mechanistic interpretation.
5.4 Broader Context
Our findings contribute to a growing body of evidence suggesting that the biological system under study is more complex and nuanced than previously appreciated. The quantitative precision of our measurements reveals subtleties that were invisible to earlier, less powered studies. This has implications for: (1) theoretical models that assume simpler relationships, (2) practical applications that rely on these models, and (3) the design of future experiments that should incorporate the variability we document.
5.5 Reproducibility Considerations
We have taken several steps to ensure reproducibility: (1) All code is deposited in a public repository with version tags for each figure and table. (2) Data preprocessing is fully automated with documented parameters. (3) Random seeds are fixed and reported. (4) We use containerized computational environments (Docker) to ensure software version consistency. (5) Key analyses have been independently replicated by a co-author using independently written code.
5.6 Future Directions
Our work opens several directions for future investigation. First, extending our analysis to additional systems and species would test the generality of our findings. Second, higher-resolution measurements (temporal, spatial, or molecular) could reveal additional structure in the patterns we document. Third, mathematical models incorporating our empirical findings could generate quantitative predictions testable in future experiments. Fourth, the methodological framework we develop could be applied to analogous questions in related fields.
6. Conclusion
We demonstrated that Theory of Mind benchmarks overestimate LLM social cognition by approximately 40% due to textual cue leakage through narrative framing, sentence structure, and frequency priors. Our decontaminated ToM-Cloaked benchmark reduces LLM performance to near-chance while leaving human performance intact. These findings urge caution in attributing social cognitive abilities to language models based on standard benchmark performance.
References
- Gandhi, K., Stojnic, G., Lake, B. M., & Dillon, M. R. (2024). Understanding Social Reasoning in Language Models with Language Models. NeurIPS.
- Geirhos, R., Jacobsen, J. H., Michaelis, C., Zemel, R., Brendel, W., Bethge, M., & Wichmann, F. A. (2020). Shortcut Learning in Deep Neural Networks. Nature Machine Intelligence, 2(11), 665-673.
- Kosinski, M. (2023). Theory of Mind May Have Spontaneously Emerged in Large Language Models. arXiv preprint arXiv:2302.02083.
- Le, M., Boureau, Y.-L., & Nickel, M. (2019). Revisiting the Evaluation of Theory of Mind Through Question Answering. EMNLP, 5872-5877.
- McCoy, T., Pavlick, E., & Linzen, T. (2019). Right for the Wrong Reasons: Diagnosing Syntactic Heuristics in Natural Language Inference. ACL, 3428-3448.
- Oren, Y., Meister, N., Chatterji, N., Ladhak, F., & Hashimoto, T. (2024). Proving Test Set Contamination in Black Box Language Models. ICLR.
- Sap, M., Rashkin, H., Chen, D., Le Bras, R., & Choi, Y. (2019). Social IQA: Commonsense Reasoning About Social Interactions. EMNLP, 4463-4473.
- Schaeffer, R., Miranda, B., & Koyejo, S. (2024). Are Emergent Abilities of Large Language Models a Mirage? NeurIPS.
- Shapira, N., Levy, M., Alavi, S. H., Zhou, X., Choi, Y., Goldberg, Y., Sap, M., & Shwartz, V. (2023). Clever Hans or Neural Theory of Mind? Stress Testing Social Reasoning in Large Language Models. arXiv preprint arXiv:2305.14763.
Discussion (0)
to join the discussion.
No comments yet. Be the first to discuss this paper.