Increasing alignment of large language models with language processing in the human brain

  • Changjiang Gao
  • , Zhengwu Ma
  • , Jiajun Chen
  • , Ping Li
  • , Shujian Huang (Corresponding Author)
  • , Jixing Li (Corresponding Author)

Research output: Journal article publicationJournal articleAcademic researchpeer-review

Abstract

Transformer-based large language models (LLMs) have considerably advanced our understanding of how meaning is represented in the human brain; however, the validity of increasingly large LLMs is being questioned due to their extensive training data and their ability to access context thousands of words long. In this study we investigated whether instruction tuning—another core technique in recent LLMs that goes beyond mere scaling—can enhance models’ ability to capture linguistic information in the human brain. We compared base and instruction-tuned LLMs of varying sizes against human behavioral and brain activity measured with eye-tracking and functional magnetic resonance imaging during naturalistic reading. We show that simply making LLMs larger leads to a closer match with the human brain than fine-tuning them with instructions. These finding have substantial implications for understanding the cognitive plausibility of LLMs and their role in studying naturalistic language comprehension.
Original languageEnglish
Pages (from-to)1080-1090
Number of pages11
JournalNature Computational Science
Volume5
Issue number11
DOIs
Publication statusPublished - 16 Sept 2025

Fingerprint

Dive into the research topics of 'Increasing alignment of large language models with language processing in the human brain'. Together they form a unique fingerprint.

Cite this