Do LLMs Exhibit Human-Like Reasoning? Evaluating Theory of Mind in LLMs for Open-Ended Responses

M Amirizaniani, E Martin, M Sivachenko… - arXiv preprint arXiv …, 2024 - arxiv.org
Theory of Mind (ToM) reasoning entails recognizing that other individuals possess their own
intentions, emotions, and thoughts, which is vital for guiding one's own thought processes …

Understanding social reasoning in language models with language models

K Gandhi, JP Fränken… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract As Large Language Models (LLMs) become increasingly integrated into our
everyday lives, understanding their ability to comprehend human mental states becomes …

PHAnToM: Personality Has An Effect on Theory-of-Mind Reasoning in Large Language Models

FA Tan, GC Yeo, F Wu, W Xu, V Jain, A Chadha… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent advances in large language models (LLMs) demonstrate that their capabilities are
comparable, or even superior, to humans in many tasks in natural language processing …

Through the Theory of Mind's Eye: Reading Minds with Multimodal Video Large Language Models

Z Chen, T Wang, Y Wang, M Kosinski, X Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Can large multimodal models have a human-like ability for emotional and social reasoning,
and if so, how does it work? Recent research has discovered emergent theory-of-mind …

Think Twice: Perspective-Taking Improves Large Language Models' Theory-of-Mind Capabilities

A Wilf, SS Lee, PP Liang, LP Morency - arXiv preprint arXiv:2311.10227, 2023 - arxiv.org
Human interactions are deeply rooted in the interplay of thoughts, beliefs, and desires made
possible by Theory of Mind (ToM): our cognitive ability to understand the mental states of …

Tomchallenges: A principle-guided dataset and diverse evaluation tasks for exploring theory of mind

X Ma, L Gao, Q Xu - arXiv preprint arXiv:2305.15068, 2023 - arxiv.org
Theory of Mind (ToM), the capacity to comprehend the mental states of distinct individuals, is
essential for numerous practical applications. With the development of large language …

How FaR Are Large Language Models From Agents with Theory-of-Mind?

P Zhou, A Madaan, SP Potharaju, A Gupta… - arXiv preprint arXiv …, 2023 - arxiv.org
" Thinking is for Doing." Humans can infer other people's mental states from observations--
an ability called Theory-of-Mind (ToM)--and subsequently act pragmatically on those …

[HTML][HTML] Comparing Humans and Large Language Models on an Experimental Protocol Inventory for Theory of Mind Evaluation (EPITOME)

CR Jones, S Trott, B Bergen - Transactions of the Association for …, 2024 - direct.mit.edu
We address a growing debate about the extent to which large language models (LLMs)
produce behavior consistent with Theory of Mind (ToM) in humans. We present EPITOME: a …

FANToM: A benchmark for stress-testing machine theory of mind in interactions

H Kim, M Sclar, X Zhou, RL Bras, G Kim, Y Choi… - arXiv preprint arXiv …, 2023 - arxiv.org
Theory of mind (ToM) evaluations currently focus on testing models using passive narratives
that inherently lack interactivity. We introduce FANToM, a new benchmark designed to stress …

Boosting theory-of-mind performance in large language models via prompting

SR Moghaddam, CJ Honey - arXiv preprint arXiv:2304.11490, 2023 - arxiv.org
Large language models (LLMs) excel in many tasks in 2023, but they still face challenges in
complex reasoning. Theory-of-mind (ToM) tasks, which require understanding agents' …