Dimensionality and Ramping: Signatures of Sentence Integration in the Dynamics of Brains and Deep Language Models

T. Desbordes, Yair Lakretz, V. Chanoine, M. Oquab, J. Badier, A. Trébuchon, R. Carron, C. Bénar, S. Dehaene, J. King
{"title":"Dimensionality and Ramping: Signatures of Sentence Integration in the Dynamics of Brains and Deep Language Models","authors":"T. Desbordes, Yair Lakretz, V. Chanoine, M. Oquab, J. Badier, A. Trébuchon, R. Carron, C. Bénar, S. Dehaene, J. King","doi":"10.1101/2023.02.28.530443","DOIUrl":null,"url":null,"abstract":"A sentence is more than the sum of its words: its meaning depends on how they combine with one another. The brain mechanisms underlying such semantic composition remain poorly understood. To shed light on the neural vector code underlying semantic composition, we introduce two hypotheses: (1) the intrinsic dimensionality of the space of neural representations should increase as a sentence unfolds, paralleling the growing complexity of its semantic representation; and (2) this progressive integration should be reflected in ramping and sentence-final signals. To test these predictions, we designed a dataset of closely matched normal and jabberwocky sentences (composed of meaningless pseudo words) and displayed them to deep language models and to 11 human participants (5 men and 6 women) monitored with simultaneous MEG and intracranial EEG. In both deep language models and electrophysiological data, we found that representational dimensionality was higher for meaningful sentences than jabberwocky. Furthermore, multivariate decoding of normal versus jabberwocky confirmed three dynamic patterns: (1) a phasic pattern following each word, peaking in temporal and parietal areas; (2) a ramping pattern, characteristic of bilateral inferior and middle frontal gyri; and (3) a sentence-final pattern in left superior frontal gyrus and right orbitofrontal cortex. These results provide a first glimpse into the neural geometry of semantic integration and constrain the search for a neural code of linguistic composition. SIGNIFICANCE STATEMENT Starting from general linguistic concepts, we make two sets of predictions in neural signals evoked by reading multiword sentences. First, the intrinsic dimensionality of the representation should grow with additional meaningful words. Second, the neural dynamics should exhibit signatures of encoding, maintaining, and resolving semantic composition. We successfully validated these hypotheses in deep neural language models, artificial neural networks trained on text and performing very well on many natural language processing tasks. Then, using a unique combination of MEG and intracranial electrodes, we recorded high-resolution brain data from human participants while they read a controlled set of sentences. Time-resolved dimensionality analysis showed increasing dimensionality with meaning, and multivariate decoding allowed us to isolate the three dynamical patterns we had hypothesized.","PeriodicalId":22786,"journal":{"name":"The Journal of Neuroscience","volume":"11 1","pages":"5350 - 5364"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Journal of Neuroscience","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2023.02.28.530443","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

A sentence is more than the sum of its words: its meaning depends on how they combine with one another. The brain mechanisms underlying such semantic composition remain poorly understood. To shed light on the neural vector code underlying semantic composition, we introduce two hypotheses: (1) the intrinsic dimensionality of the space of neural representations should increase as a sentence unfolds, paralleling the growing complexity of its semantic representation; and (2) this progressive integration should be reflected in ramping and sentence-final signals. To test these predictions, we designed a dataset of closely matched normal and jabberwocky sentences (composed of meaningless pseudo words) and displayed them to deep language models and to 11 human participants (5 men and 6 women) monitored with simultaneous MEG and intracranial EEG. In both deep language models and electrophysiological data, we found that representational dimensionality was higher for meaningful sentences than jabberwocky. Furthermore, multivariate decoding of normal versus jabberwocky confirmed three dynamic patterns: (1) a phasic pattern following each word, peaking in temporal and parietal areas; (2) a ramping pattern, characteristic of bilateral inferior and middle frontal gyri; and (3) a sentence-final pattern in left superior frontal gyrus and right orbitofrontal cortex. These results provide a first glimpse into the neural geometry of semantic integration and constrain the search for a neural code of linguistic composition. SIGNIFICANCE STATEMENT Starting from general linguistic concepts, we make two sets of predictions in neural signals evoked by reading multiword sentences. First, the intrinsic dimensionality of the representation should grow with additional meaningful words. Second, the neural dynamics should exhibit signatures of encoding, maintaining, and resolving semantic composition. We successfully validated these hypotheses in deep neural language models, artificial neural networks trained on text and performing very well on many natural language processing tasks. Then, using a unique combination of MEG and intracranial electrodes, we recorded high-resolution brain data from human participants while they read a controlled set of sentences. Time-resolved dimensionality analysis showed increasing dimensionality with meaning, and multivariate decoding allowed us to isolate the three dynamical patterns we had hypothesized.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
维度与斜坡:脑动力学与深度语言模型中句子整合的特征
一个句子不仅仅是单词的总和,它的意义取决于单词如何相互组合。这种语义构成背后的大脑机制仍然知之甚少。为了阐明语义构成背后的神经向量代码,我们引入了两个假设:(1)随着句子的展开,神经表征空间的内在维度应该增加,与语义表征的复杂性并行;(2)这种渐进式的整合应该体现在斜坡和句末信号上。为了验证这些预测,我们设计了一个紧密匹配的正常和胡言乱语句子(由无意义的伪词组成)的数据集,并将它们展示给深度语言模型和11名人类参与者(5名男性和6名女性),同时监测MEG和颅内脑电图。在深度语言模型和电生理数据中,我们发现有意义句子的表征维数高于废话句子。此外,正常和废话的多变量解码证实了三种动态模式:(1)每个单词之后都有一个相位模式,在颞叶和顶叶区域达到峰值;(2)双侧额下回和额中回呈斜坡状;(3)左侧额上回和右侧眼窝额叶皮层的句末模式。这些结果提供了对语义整合的神经几何的第一次一瞥,并限制了对语言组成的神经代码的搜索。我们从一般的语言学概念出发,对阅读多词句子所引起的神经信号进行了两组预测。首先,表征的内在维度应该随着额外的有意义的词语而增长。其次,神经动力学应该表现出编码、维护和解析语义组成的特征。我们成功地在深度神经语言模型中验证了这些假设,在文本上训练的人工神经网络在许多自然语言处理任务上表现得很好。然后,使用MEG和颅内电极的独特组合,我们记录了人类参与者在阅读一组受控句子时的高分辨率大脑数据。时间分辨维度分析显示维度随意义的增加而增加,多元解码使我们能够分离出我们假设的三种动态模式。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Erratum: Schlüter et al., “Rabphilin Knock-Out Mice Reveal That Rabphilin Is Not Required for Rab3 Function in Regulating Neurotransmitter Release” Category-selective representation of relationships in visual cortex Phosphorylation of RPT6 controls its ability to bind DNA and regulate gene expression in the hippocampus of male rats during memory formation Neural network connectivity following opioid dependence is altered by a common genetic variant in the mu-opioid receptor,OPRM1A118G An Ascending Excitatory Circuit from the Dorsal Raphe for Sensory Modulation of Pain
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1