Brain-LLM Connections Through Explainable AI

Brain-LLM Connections Through Explainable AI

Using LLM explanations to predict brain activity during language processing

This research demonstrates how explainable AI techniques applied to large language models can predict actual brain activity patterns during language processing.

  • Uses attribution methods to quantify word influence in LLM predictions
  • Applies these explanations to successfully predict fMRI data from human participants
  • Establishes stronger links between computational models and biological language processing
  • Suggests new ways to understand both LLMs and human cognition

This work advances medical neuroscience by providing novel methods to analyze brain imaging data, potentially improving our understanding of language disorders and creating more brain-like AI systems.

Explanations of Large Language Models Explain Language Representations in the Brain

43 | 78