Theory of Mind in AI Systems

Theory of Mind in AI Systems

How LLMs Understand Human Mental States

This systematic review examines how large language models (LLMs) perform on Theory of Mind (ToM) tasks—the ability to attribute mental states to oneself and others.

  • Current research reveals diverse methodologies for evaluating ToM capabilities in AI systems
  • LLMs show promising but inconsistent abilities to understand beliefs, intentions, and perspectives
  • Standardized evaluation frameworks are emerging but remain fragmented
  • Performance on ToM tasks varies significantly across different model architectures and training approaches

Medical Relevance: Understanding ToM in AI systems has significant implications for medical applications, including better modeling of cognitive disorders like autism and schizophrenia, and developing more empathetic healthcare AI assistants.

A Systematic Review on the Evaluation of Large Language Models in Theory of Mind Tasks

37 | 85