The existence of a neural representation for whole words (i.e., a lexicon) is a common feature of many models of speech processing. Prior studies have provided evidence for a visual lexicon containing representations of whole written words in an area of the ventral visual stream known as the visual word form area. Similar experimental support for an auditory lexicon containing representations of spoken words has yet to be shown. Using functional magnetic resonance imaging rapid adaptation techniques, we provide evidence for an auditory lexicon in the auditory word form area in the human left anterior superior temporal gyrus that contains representations highly selective for individual spoken words. Furthermore, we show that familiarization with novel auditory words sharpens the selectivity of their representations in the auditory word form area. These findings reveal strong parallels in how the brain represents written and spoken words, showing convergent processing strategies across modalities in the visual and auditory ventral streams.
Studies have shown that the integrity of white matter tracts connecting different regions in the left cerebral hemisphere is important for aphasia recovery after stroke. However, the impact of the underlying structural connection between the cortex and the cerebellum in post-stroke aphasia is poorly understood. We studied the microstructural integrity of the cerebellum and the corticocerebellar connections and their role in picture naming. Fifty-six patients with left cerebral infarcts (sparing the cerebellum) underwent diffusion tensor imaging (DTI) and Boston Naming Test. We compared the fractional anisotropy (FA) and mean diffusivity (MD) values of the right and the left cerebellum (lobular gray and white matter structures) and cerebellocortical connections. Recursive feature elimination and Spearman correlation analyses were performed to evaluate the relationship between naming performance and the corticocerebellar connections. We found that the right, relative to left, cerebellar structures and their connections with the left cerebrum showed lower FA and higher MD values, both reflecting lower microstructural integrity. This trend was not observed in the healthy controls. Higher MD values of the right major cerebellar outflow tract were associated with poorer picture naming performance. Our study provides the first DTI data demonstrating the critical importance of ascending and descending corticocerebellar connections for naming outcomes after stroke.
Older adults typically exhibit decline in language production. However, how the brain supports or fails to support these processes is unclear. Moreover, there are competing hypotheses about the nature of age-related neural changes and whether age-related increases in neural activity reflect compensation or a decline in neural efficiency. In the current study, we investigated the neural bases of language production focusing on resting state functional connectivity. We hypothesized that language production performance, functional connectivity, and their relationship would differ as a function of age. Consistent with prior work, older age was associated with worse language production performance. Functional connectivity analyses showed that network segregation within the left hemisphere language network was maintained across adulthood. However, increased age was associated with lower whole brain network segregation. Moreover, network segregation was related to language production ability. In both network analyses, there were significant interactions with age-higher network segregation was associated with better language production abilities for younger and middle-aged adults, but not for older adults. Interestingly, there was a stronger relationship between language production and the whole brain network segregation than between production and the language network. These results highlight the utility of network segregation measures as an index of brain function, with higher network segregation associated with better language production ability. Moreover, these results are consistent with stability in the left hemisphere language network across adulthood and suggest that dedifferentiation among brain networks, outside of the language network, is a hallmark of aging and may contribute to age-related language production difficulties.
Considerable work suggests the dominant syllable rhythm of the acoustic envelope is remarkably similar across languages (∼4-5 Hz) and that oscillatory brain activity tracks these quasiperiodic rhythms to facilitate speech processing. However, whether this fundamental periodicity represents a common organizing principle in both auditory and motor systems involved in speech has not been explicitly tested. To evaluate relations between entrainment in the perceptual and production domains, we measured individuals' (i) neuroacoustic tracking of the EEG to speech trains and their (ii) simultaneous and non-simultaneous productions synchronized to syllable rates between 2.5 and 8.5 Hz. Productions made without concurrent auditory presentation isolated motor speech functions more purely. We show that neural synchronization flexibly adapts to the heard stimuli in a rate-dependent manner, but that phase locking is boosted near ∼4.5 Hz, the purported dominant rate of speech. Cued speech productions (recruit sensorimotor interaction) were optimal between 2.5 and 4.5 Hz, suggesting a low-frequency constraint on motor output and/or sensorimotor integration. In contrast, "pure" motor productions (without concurrent sound cues) were most precisely generated at rates of 4.5 and 5.5 Hz, paralleling the neuroacoustic data. Correlations further revealed strong links between receptive (EEG) and production synchronization abilities; individuals with stronger auditory-perceptual entrainment better matched speech rhythms motorically. Together, our findings support an intimate link between exogenous and endogenous rhythmic processing that is optimized at 4-5 Hz in both auditory and motor systems. Parallels across modalities could result from dynamics of the speech motor system coupled with experience-dependent tuning of the perceptual system via the sensorimotor interface.
One of the core features of brain maturation is functional specialization. Previous research has found that 7- to 8-year-old children start to specialize in both the temporal and frontal lobes. However, as children continue to develop their phonological and semantic skills rapidly until approximately 10 years old, it remained unclear whether any changes in specialization later in childhood would be detected. Thus, the goal of the current study was to examine phonological and semantic specialization in 9- to 10-year-old children during auditory word processing. Sixty-one children were included in the analysis. They were asked to perform a sound judgment task and a meaning judgment task, each with both hard and easy conditions to examine parametric effects. Consistent with previous results from 7- to 8-year-old children, direct task comparisons revealed language specialization in both the temporal and frontal lobes in 9- to 10-year-old children. Specifically, the left dorsal inferior frontal gyrus showed greater activation for the sound than the meaning task whereas the left middle temporal gyrus showed greater activation for the meaning than the sound task. Interestingly, in contrast to the previously reported finding that 7- to 8-year-old children primarily engage a general control region during the harder condition for both tasks, we showed that 9- to 10-year-old children recruited language-specific regions to process the more difficult task conditions. Specifically, the left superior temporal gyrus showed greater activation for the phonological parametric manipulation whereas the left ventral inferior frontal gyrus showed greater activation for the semantic parametric manipulation.
Speech processing often occurs amid competing inputs from other modalities, for example, listening to the radio while driving. We examined the extent to which dividing attention between auditory and visual modalities (bimodal divided attention) impacts neural processing of natural continuous speech from acoustic to linguistic levels of representation. We recorded electroencephalographic (EEG) responses when human participants performed a challenging primary visual task, imposing low or high cognitive load while listening to audiobook stories as a secondary task. The two dual-task conditions were contrasted with an auditory single-task condition in which participants attended to stories while ignoring visual stimuli. Behaviorally, the high load dual-task condition was associated with lower speech comprehension accuracy relative to the other two conditions. We fitted multivariate temporal response function encoding models to predict EEG responses from acoustic and linguistic speech features at different representation levels, including auditory spectrograms and information-theoretic models of sublexical-, word-form-, and sentence-level representations. Neural tracking of most acoustic and linguistic features remained unchanged with increasing dual-task load, despite unambiguous behavioral and neural evidence of the high load dual-task condition being more demanding. Compared to the auditory single-task condition, dual-task conditions selectively reduced neural tracking of only some acoustic and linguistic features, mainly at latencies >200 ms, while earlier latencies were surprisingly unaffected. These findings indicate that behavioral effects of bimodal divided attention on continuous speech processing occur not because of impaired early sensory representations but likely at later cognitive processing stages. Crossmodal attention-related mechanisms may not be uniform across different speech processing levels.
Diversity and variation in language experiences, such as bilingualism, contribute to heterogeneity in children's neural organization for language and brain development. To uncover sources of such heterogeneity in children's neural language networks, the present study examined the effects of bilingual proficiency on children's neural organization for language function. To do so, we took an innovative person-specific analytical approach to investigate young Chinese-English and Spanish-English bilingual learners of structurally distinct languages. Bilingual and English monolingual children (N = 152, M(SD)age = 7.71(1.32)) completed an English word recognition task during functional near-infrared spectroscopy neuroimaging, along with language and literacy tasks in each of their languages. Two key findings emerged. First, bilinguals' heritage language proficiency (Chinese or Spanish) made a unique contribution to children's language network density. Second, the findings reveal common and unique patterns in children's patterns of task-related functional connectivity. Common across all participants were short-distance neural connections within left hemisphere regions associated with semantic processes (within middle temporal and frontal regions). Unique to more proficient language users were additional long-distance connections between frontal, temporal, and bilateral regions within the broader language network. The study informs neurodevelopmental theories of language by revealing the effects of heterogeneity in language proficiency and experiences on the structure and quality of emerging language neural networks in linguistically diverse learners.
[This corrects the article DOI: 10.1162/nol_a_00081.].