Do Large Language Models Predict When the Brain Expects a Metaphor? LLM Surprisal Reveals Theory-of-Mind Dependent Neural Alignment?
May 12, 2026·
,,,·
0 min read
Agnese Lombardi
Chiara Battaglini
Paolo Canal
Valentina Bambini

Abstract
Understanding metaphors involve more than detecting an unexpected word. Metaphors require inferring communicative intentions that go beyond literal meanings, and often map abstract concepts onto more concrete domains of experience, hence engaging Theory of Mind (ToM). Recently, Canal et al. (2022) records the EEG response to physical and mental metaphors (i.e., metaphors referring to behavioral and psychological features, respectively) and assessed participants’ ToM using the Reading the Mind in the Eyes Test (RMET). They show that higher RMET scores were associated with reduced N400 for both types of metaphors and with an earlier neural differentiation between metaphor types. Here, we tested whether these ToM-dependent neural dynamics can be reproduced using predictive measures from large language models (LLMs).
To quantify the expectedness of each metaphor vehicle, we computed surprisal (the negative log probability of a word given its preceding context) from six LLMs. We extracted surprisal layer by layer, allowing us to test at which representational depth the model’s internal probability aligns with humans, and we calculated correlations with the EEG amplitude across the 200–1000 ms window.
Results revealed a different patterns depending on ToM. In high-ToM individuals, the correlation between surprisal and the EEG amplitude emerged between 200–400 ms, a time window consistent with predictive contextual integration. In low-ToM individuals, the same relationship appeared between 600–1000 ms, consistent with global analysis processes. This difference is robust for physical metaphors.
These findings suggest that surprisal captures a dimension of contextual semantic fit that aligns with human neural processing, but only under specific cognitive conditions. High ToM ability is associated with a fast, expectancy-driven route to metaphor comprehension that mirrors the predictive structure encoded in model layers. Low ToM ability engages in a slower reanalysis route that relates to the same statistical signal but at a delayed latency.
The absence of robust alignment for mental metaphors indicates that mental-state inference may rely on representational mechanisms that extend beyond what distributional language statistics alone can capture.
Date
May 12, 2026 12:00 AM
Event
Location
Pavia