Medical College of Wisconsin
CTSICores SearchResearch InformaticsREDCap

Deep Artificial Neural Networks Reveal a Distributed Cortical Network Encoding Propositional Sentence-Level Meaning. J Neurosci 2021 May 05;41(18):4100-4119

Date

03/24/2021

Pubmed ID

33753548

Pubmed Central ID

PMC8176751

DOI

10.1523/JNEUROSCI.1152-20.2021

Scopus ID

2-s2.0-85105889500 (requires institutional sign-in at Scopus site)   21 Citations

Abstract

Understanding how and where in the brain sentence-level meaning is constructed from words presents a major scientific challenge. Recent advances have begun to explain brain activation elicited by sentences using vector models of word meaning derived from patterns of word co-occurrence in text corpora. These studies have helped map out semantic representation across a distributed brain network spanning temporal, parietal, and frontal cortex. However, it remains unclear whether activation patterns within regions reflect unified representations of sentence-level meaning, as opposed to superpositions of context-independent component words. This is because models have typically represented sentences as "bags-of-words" that neglect sentence-level structure. To address this issue, we interrogated fMRI activation elicited as 240 sentences were read by 14 participants (9 female, 5 male), using sentences encoded by a recurrent deep artificial neural-network trained on a sentence inference task (InferSent). Recurrent connections and nonlinear filters enable InferSent to transform sequences of word vectors into unified "propositional" sentence representations suitable for evaluating intersentence entailment relations. Using voxelwise encoding modeling, we demonstrate that InferSent predicts elements of fMRI activation that cannot be predicted by bag-of-words models and sentence models using grammatical rules to assemble word vectors. This effect occurs throughout a distributed network, which suggests that propositional sentence-level meaning is represented within and across multiple cortical regions rather than at any single site. In follow-up analyses, we place results in the context of other deep network approaches (ELMo and BERT) and estimate the degree of unpredicted neural signal using an "experiential" semantic model and cross-participant encoding.SIGNIFICANCE STATEMENT A modern-day scientific challenge is to understand how the human brain transforms word sequences into representations of sentence meaning. A recent approach, emerging from advances in functional neuroimaging, big data, and machine learning, is to computationally model meaning, and use models to predict brain activity. Such models have helped map a cortical semantic information-processing network. However, how unified sentence-level information, as opposed to word-level units, is represented throughout this network remains unclear. This is because models have typically represented sentences as unordered "bags-of-words." Using a deep artificial neural network that recurrently and nonlinearly combines word representations into unified propositional sentence representations, we provide evidence that sentence-level information is encoded throughout a cortical network, rather than in a single region.

Author List

Anderson AJ, Kiela D, Binder JR, Fernandino L, Humphries CJ, Conant LL, Raizada RDS, Grimm S, Lalor EC

Authors

Andrew J. Anderson PhD Assistant Professor in the Neurology department at Medical College of Wisconsin
Jeffrey R. Binder MD Professor in the Neurology department at Medical College of Wisconsin
Leonardo Fernandino PhD Assistant Professor in the Neurology department at Medical College of Wisconsin




MESH terms used to index this publication - Major topics in bold

Adult
Cerebral Cortex
Comprehension
Computer Simulation
Female
Humans
Language
Magnetic Resonance Imaging
Male
Middle Aged
Reading
Semantics
Young Adult