{"title":"Neural bases of proactive and predictive processing of meaningful sub-word units in speech comprehension.","authors":"Suhail Matar, Alec Marantz","doi":"10.1523/JNEUROSCI.0781-24.2024","DOIUrl":null,"url":null,"abstract":"<p><p>To comprehend speech, human brains identify meaningful units in the speech stream. But whereas the English '<i>She believed him.</i>' has 3 word-units, the Arabic equivalent '<i>ṣaddaqathu.</i>' is a single word-unit with 3 meaningful sub-word units, called morphemes: a verb stem ('<i>ṣaddaqa</i>'), a subject suffix ('-<i>t</i>-'), and a direct object pronoun ('-<i>hu</i>'). It remains unclear whether and how the brain processes morphemes, above and beyond other language units, during speech comprehension. Here, we propose and test hierarchically-nested encoding models of speech comprehension: a naïve model with word-, syllable-, and sound-level information; a bottom-up model with additional morpheme boundary information; and predictive models that process morphemes before these boundaries. We recorded magnetoencephalography (MEG) data as 27 participants (16 female) listened to Arabic sentences like '<i>ṣaddaqathu.</i>'. A temporal response function (TRF) analysis revealed that in temporal and left inferior frontal regions predictive models outperform the bottom-up model, which outperforms the naïve model. Moreover, verb stems were either length-ambiguous (e.g., '<i>ṣaddaqa</i>' could initially be mistaken for the shorter stem '<i>ṣadda</i>'='<i>blocked</i>') or length-unambiguous (e.g., '<i>qayyama</i>'='<i>evaluated</i>' cannot be mistaken for a shorter stem), but shared a uniqueness point, beyond which stem identity is fully disambiguated. Evoked analyses revealed differences between conditions before the uniqueness point, suggesting that, rather than await disambiguation, the brain employs proactive predictive strategies, processing accumulated input as soon as any possible stem is identifiable, even if not uniquely. These findings highlight the role of morphemes in speech, and the importance of including morpheme-level information in neural and computational models of speech comprehension.<b>Significance statement</b> Many leading models of speech comprehension include information about words, syllables and sounds. But languages vary considerably in the amount of meaning packed into word units. This work proposes speech comprehension models with information about meaningful sub-word units, called morphemes (e.g., '<i>bake-</i>' and '<i>-ing</i>' in '<i>baking</i>'), and shows that they explain significantly more neural activity than models without morpheme information. We also show how the brain predictively processes morphemic information. These findings highlight the role of morphemes in speech comprehension and emphasize the contributions of morpheme-level information-theoretic metrics, like surprisal and entropy. Our findings can be used to update current neural, cognitive, and computational models of speech comprehension, and constitute a step towards refining those models for naturalistic, connected speech.</p>","PeriodicalId":50114,"journal":{"name":"Journal of Neuroscience","volume":" ","pages":""},"PeriodicalIF":4.4000,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Neuroscience","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1523/JNEUROSCI.0781-24.2024","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"NEUROSCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
To comprehend speech, human brains identify meaningful units in the speech stream. But whereas the English 'She believed him.' has 3 word-units, the Arabic equivalent 'ṣaddaqathu.' is a single word-unit with 3 meaningful sub-word units, called morphemes: a verb stem ('ṣaddaqa'), a subject suffix ('-t-'), and a direct object pronoun ('-hu'). It remains unclear whether and how the brain processes morphemes, above and beyond other language units, during speech comprehension. Here, we propose and test hierarchically-nested encoding models of speech comprehension: a naïve model with word-, syllable-, and sound-level information; a bottom-up model with additional morpheme boundary information; and predictive models that process morphemes before these boundaries. We recorded magnetoencephalography (MEG) data as 27 participants (16 female) listened to Arabic sentences like 'ṣaddaqathu.'. A temporal response function (TRF) analysis revealed that in temporal and left inferior frontal regions predictive models outperform the bottom-up model, which outperforms the naïve model. Moreover, verb stems were either length-ambiguous (e.g., 'ṣaddaqa' could initially be mistaken for the shorter stem 'ṣadda'='blocked') or length-unambiguous (e.g., 'qayyama'='evaluated' cannot be mistaken for a shorter stem), but shared a uniqueness point, beyond which stem identity is fully disambiguated. Evoked analyses revealed differences between conditions before the uniqueness point, suggesting that, rather than await disambiguation, the brain employs proactive predictive strategies, processing accumulated input as soon as any possible stem is identifiable, even if not uniquely. These findings highlight the role of morphemes in speech, and the importance of including morpheme-level information in neural and computational models of speech comprehension.Significance statement Many leading models of speech comprehension include information about words, syllables and sounds. But languages vary considerably in the amount of meaning packed into word units. This work proposes speech comprehension models with information about meaningful sub-word units, called morphemes (e.g., 'bake-' and '-ing' in 'baking'), and shows that they explain significantly more neural activity than models without morpheme information. We also show how the brain predictively processes morphemic information. These findings highlight the role of morphemes in speech comprehension and emphasize the contributions of morpheme-level information-theoretic metrics, like surprisal and entropy. Our findings can be used to update current neural, cognitive, and computational models of speech comprehension, and constitute a step towards refining those models for naturalistic, connected speech.
期刊介绍:
JNeurosci (ISSN 0270-6474) is an official journal of the Society for Neuroscience. It is published weekly by the Society, fifty weeks a year, one volume a year. JNeurosci publishes papers on a broad range of topics of general interest to those working on the nervous system. Authors now have an Open Choice option for their published articles