首页 > 最新文献

Trends in Hearing最新文献

英文 中文
Impact of Hearing Aids on Language Outcomes in Preschool Children With Mild Bilateral Hearing Loss. 助听器对轻度双侧听力损失学龄前儿童语言成果的影响。
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241256721
Yu-Chen Hung, Pei-Hsuan Ho, Pei-Hua Chen, Yi-Shin Tsai, Yi-Jui Li, Hung-Ching Lin

This study aimed to investigate the role of hearing aid (HA) usage in language outcomes among preschool children aged 3-5 years with mild bilateral hearing loss (MBHL). The data were retrieved from a total of 52 children with MBHL and 30 children with normal hearing (NH). The association between demographical, audiological factors and language outcomes was examined. Analyses of variance were conducted to compare the language abilities of HA users, non-HA users, and their NH peers. Furthermore, regression analyses were performed to identify significant predictors of language outcomes. Aided better ear pure-tone average (BEPTA) was significantly correlated with language comprehension scores. Among children with MBHL, those who used HA outperformed the ones who did not use HA across all linguistic domains. The language skills of children with MBHL were comparable to those of their peers with NH. The degree of improvement in audibility in terms of aided BEPTA was a significant predictor of language comprehension. It is noteworthy that 50% of the parents expressed reluctance regarding HA use for their children with MBHL. The findings highlight the positive impact of HA usage on language development in this population. Professionals may therefore consider HAs as a viable treatment option for children with MBHL, especially when there is a potential risk of language delay due to hearing loss. It was observed that 25% of the children with MBHL had late-onset hearing loss. Consequently, the implementation of preschool screening or a listening performance checklist is recommended to facilitate early detection.

本研究旨在调查助听器(HA)的使用对 3-5 岁轻度双侧听力损失(MBHL)学龄前儿童语言能力的影响。数据来自 52 名轻度双侧听力损失(MBHL)儿童和 30 名听力正常(NH)儿童。研究考察了人口统计学、听力学因素与语言能力之间的关系。通过方差分析,比较了听力障碍儿童、非听力障碍儿童和听力正常儿童的语言能力。此外,还进行了回归分析,以确定语言能力的重要预测因素。辅助较好耳纯音平均值(BEPTA)与语言理解能力得分有明显的相关性。在患有 MBHL 的儿童中,使用 HA 的儿童在所有语言领域的表现均优于未使用 HA 的儿童。患有 MBHL 的儿童的语言能力与患有 NH 的儿童相当。从辅助 BEPTA 的角度来看,可听度的改善程度是语言理解能力的重要预测因素。值得注意的是,50% 的家长表示不愿意为患有 MBHL 的儿童使用助听器。研究结果凸显了使用助听器对该群体语言发展的积极影响。因此,专业人士可以考虑将助听器作为治疗 MBHL 儿童的一种可行方法,尤其是在听力损失可能导致语言发育迟缓的情况下。据观察,25%的 MBHL 儿童有迟发性听力损失。因此,建议实施学前筛查或听力表现检查表,以便及早发现。
{"title":"Impact of Hearing Aids on Language Outcomes in Preschool Children With Mild Bilateral Hearing Loss.","authors":"Yu-Chen Hung, Pei-Hsuan Ho, Pei-Hua Chen, Yi-Shin Tsai, Yi-Jui Li, Hung-Ching Lin","doi":"10.1177/23312165241256721","DOIUrl":"10.1177/23312165241256721","url":null,"abstract":"<p><p>This study aimed to investigate the role of hearing aid (HA) usage in language outcomes among preschool children aged 3-5 years with mild bilateral hearing loss (MBHL). The data were retrieved from a total of 52 children with MBHL and 30 children with normal hearing (NH). The association between demographical, audiological factors and language outcomes was examined. Analyses of variance were conducted to compare the language abilities of HA users, non-HA users, and their NH peers. Furthermore, regression analyses were performed to identify significant predictors of language outcomes. Aided better ear pure-tone average (BEPTA) was significantly correlated with language comprehension scores. Among children with MBHL, those who used HA outperformed the ones who did not use HA across all linguistic domains. The language skills of children with MBHL were comparable to those of their peers with NH. The degree of improvement in audibility in terms of aided BEPTA was a significant predictor of language comprehension. It is noteworthy that 50% of the parents expressed reluctance regarding HA use for their children with MBHL. The findings highlight the positive impact of HA usage on language development in this population. Professionals may therefore consider HAs as a viable treatment option for children with MBHL, especially when there is a potential risk of language delay due to hearing loss. It was observed that 25% of the children with MBHL had late-onset hearing loss. Consequently, the implementation of preschool screening or a listening performance checklist is recommended to facilitate early detection.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241256721"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11113073/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141076740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Easy as 1-2-3: Development and Evaluation of a Simple yet Valid Audiogram-Classification System. 简单如 1-2-3:开发和评估简单有效的听力图分类系统。
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241260041
Larry E Humes, David A Zapala

Almost since the inception of the modern-day electroacoustic audiometer a century ago the results of pure-tone audiometry have been characterized by an audiogram. For almost as many years, clinicians and researchers have sought ways to distill the volume and complexity of information on the audiogram. Commonly used approaches have made use of pure-tone averages (PTAs) for various frequency ranges with the PTA for 500, 1000, 2000 and 4000 Hz (PTA4) being the most widely used for the categorization of hearing loss severity. Here, a three-digit triad is proposed as a single-number summary of not only the severity, but also the configuration and bilateral symmetry of the hearing loss. Each digit in the triad ranges from 0 to 9, increasing as the level of the pure-tone hearing threshold level (HTL) increases from a range of optimal hearing (< 10 dB Hearing Level; HL) to complete hearing loss (≥ 90 dB HL). Each digit also represents a different frequency region of the audiogram proceeding from left to right as: (Low, L) PTA for 500, 1000, and 2000 Hz; (Center, C) PTA for 3000, 4000 and 6000 Hz; and (High, H) HTL at 8000 Hz. This LCH Triad audiogram-classification system is evaluated using a large United States (U.S.) national dataset (N = 8,795) from adults 20 to 80 + years of age and two large clinical datasets totaling 8,254 adults covering a similar age range. Its ability to capture variations in hearing function was found to be superior to that of the widely used PTA4.

几乎自一个世纪前现代电声测听仪问世以来,纯音测听的结果就以听力图为特征。几乎同样多年来,临床医生和研究人员一直在寻找方法来提炼听力图上的大量复杂信息。常用的方法是使用不同频率范围的纯音平均值(PTA),其中 500、1000、2000 和 4000 Hz 的纯音平均值(PTA4)最广泛用于听力损失严重程度的分类。在此,我们提出了一个三位数的三元组,它不仅是听力损失严重程度的单数总结,也是听力损失结构和双侧对称性的单数总结。三位数中的每个数字范围从 0 到 9,随着纯音听力阈值水平(HTL)的增加而增加,从最佳听力(< 10 dB 听力水平;HL)到完全听力损失(≥ 90 dB HL)。每个数字还代表听力图的不同频率区域,从左到右依次为(低,L)500、1000 和 2000 Hz 的 PTA;(中,C)3000、4000 和 6000 Hz 的 PTA;(高,H)8000 Hz 的 HTL。该 LCH 三联听力图分类系统通过一个大型的美国全国数据集(N = 8795)进行评估,该数据集来自 20 至 80 多岁的成年人,以及两个大型的临床数据集,共计 8254 名成年人,涵盖了相似的年龄范围。结果发现,它捕捉听力功能变化的能力优于广泛使用的 PTA4。
{"title":"Easy as 1-2-3: Development and Evaluation of a Simple yet Valid Audiogram-Classification System.","authors":"Larry E Humes, David A Zapala","doi":"10.1177/23312165241260041","DOIUrl":"10.1177/23312165241260041","url":null,"abstract":"<p><p>Almost since the inception of the modern-day electroacoustic audiometer a century ago the results of pure-tone audiometry have been characterized by an audiogram. For almost as many years, clinicians and researchers have sought ways to distill the volume and complexity of information on the audiogram. Commonly used approaches have made use of pure-tone averages (PTAs) for various frequency ranges with the PTA for 500, 1000, 2000 and 4000 Hz (PTA4) being the most widely used for the categorization of hearing loss severity. Here, a three-digit triad is proposed as a single-number summary of not only the severity, but also the configuration and bilateral symmetry of the hearing loss. Each digit in the triad ranges from 0 to 9, increasing as the level of the pure-tone hearing threshold level (HTL) increases from a range of optimal hearing (< 10 dB Hearing Level; HL) to complete hearing loss (≥ 90 dB HL). Each digit also represents a different frequency region of the audiogram proceeding from left to right as: (Low, L) PTA for 500, 1000, and 2000 Hz; (Center, C) PTA for 3000, 4000 and 6000 Hz; and (High, H) HTL at 8000 Hz. This LCH Triad audiogram-classification system is evaluated using a large United States (U.S.) national dataset (N = 8,795) from adults 20 to 80 + years of age and two large clinical datasets totaling 8,254 adults covering a similar age range. Its ability to capture variations in hearing function was found to be superior to that of the widely used PTA4.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241260041"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11179497/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141318660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Right-Ear Advantage in Static and Dynamic Cocktail-Party Situations. 右耳在静态和动态鸡尾酒会中的优势
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165231215916
Moritz Wächtler, Pascale Sandmann, Hartmut Meister

When presenting two competing speech stimuli, one to each ear, a right-ear advantage (REA) can often be observed, reflected in better speech recognition compared to the left ear. Considering the left-hemispheric dominance for language, the REA has been explained by superior contralateral pathways (structural models) and language-induced shifts of attention to the right (attentional models). There is some evidence that the REA becomes more pronounced, as cognitive load increases. Hence, it is interesting to investigate the REA in static (constant target talker) and dynamic (target changing pseudo-randomly) cocktail-party situations, as the latter is associated with a higher cognitive load than the former. Furthermore, previous research suggests an increasing REA, when listening becomes more perceptually challenging. The present study examined the REA by using virtual acoustics to simulate static and dynamic cocktail-party situations, with three spatially separated talkers uttering concurrent matrix sentences. Sentences were presented at low sound pressure levels or processed with a noise vocoder to increase perceptual load. Sixteen young normal-hearing adults participated in the study. The REA was assessed by means of word recognition scores and a detailed error analysis. Word recognition revealed a greater REA for the dynamic than for the static situations, compatible with the view that an increase in cognitive load results in a heightened REA. Also, the REA depended on the type of perceptual load, as indicated by a higher REA associated with vocoded compared to low-level stimuli. The results of the error analysis support both structural and attentional models of the REA.

当呈现两个相互竞争的语音刺激时,两只耳朵各接受一个刺激,通常可以观察到右耳优势(REA),这反映在与左耳相比,右耳的语音识别能力更强。考虑到左半球在语言方面的优势,REA 可通过对侧的优势通路(结构模型)和语言引起的注意力向右侧转移(注意模型)来解释。有证据表明,随着认知负荷的增加,REA 会变得更加明显。因此,研究静态(目标谈话者不变)和动态(目标伪随机变化)鸡尾酒会情况下的 REA 是很有意义的,因为后者比前者与更高的认知负荷相关。此外,以往的研究表明,当听力变得更具知觉挑战性时,REA 会增加。本研究通过使用虚拟声学模拟静态和动态鸡尾酒会情境,让三个空间上分开的说话者同时说出矩阵句子,来检验 REA。句子以低声压级呈现,或用噪声声码器处理,以增加知觉负荷。16 名听力正常的年轻成年人参与了研究。通过单词识别得分和详细的错误分析对 REA 进行了评估。单词识别结果显示,动态情况下的 REA 高于静态情况下的 REA,这与认知负荷增加会导致 REA 增加的观点相吻合。此外,REA 还取决于感知负荷的类型,如与低级刺激相比,与声码刺激相关的 REA 更高。误差分析的结果支持 REA 的结构模型和注意模型。
{"title":"The Right-Ear Advantage in Static and Dynamic Cocktail-Party Situations.","authors":"Moritz Wächtler, Pascale Sandmann, Hartmut Meister","doi":"10.1177/23312165231215916","DOIUrl":"10.1177/23312165231215916","url":null,"abstract":"<p><p>When presenting two competing speech stimuli, one to each ear, a right-ear advantage (REA) can often be observed, reflected in better speech recognition compared to the left ear. Considering the left-hemispheric dominance for language, the REA has been explained by superior contralateral pathways (structural models) and language-induced shifts of attention to the right (attentional models). There is some evidence that the REA becomes more pronounced, as cognitive load increases. Hence, it is interesting to investigate the REA in static (constant target talker) and dynamic (target changing pseudo-randomly) cocktail-party situations, as the latter is associated with a higher cognitive load than the former. Furthermore, previous research suggests an increasing REA, when listening becomes more perceptually challenging. The present study examined the REA by using virtual acoustics to simulate static and dynamic cocktail-party situations, with three spatially separated talkers uttering concurrent matrix sentences. Sentences were presented at low sound pressure levels or processed with a noise vocoder to increase perceptual load. Sixteen young normal-hearing adults participated in the study. The REA was assessed by means of word recognition scores and a detailed error analysis. Word recognition revealed a greater REA for the dynamic than for the static situations, compatible with the view that an increase in cognitive load results in a heightened REA. Also, the REA depended on the type of perceptual load, as indicated by a higher REA associated with vocoded compared to low-level stimuli. The results of the error analysis support both structural and attentional models of the REA.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165231215916"},"PeriodicalIF":2.6,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10826403/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139570355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Head and Eye Movements Reveal Compensatory Strategies for Acute Binaural Deficits During Sound Localization. 头部和眼球运动揭示了声音定位过程中急性双耳缺陷的补偿策略
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165231217910
Robel Z Alemu, Blake C Papsin, Robert V Harrison, Al Blakeman, Karen A Gordon

The present study aimed to define use of head and eye movements during sound localization in children and adults to: (1) assess effects of stationary versus moving sound and (2) define effects of binaural cues degraded through acute monaural ear plugging. Thirty-three youth (MAge = 12.9 years) and seventeen adults (MAge = 24.6 years) with typical hearing were recruited and asked to localize white noise anywhere within a horizontal arc from -60° (left) to +60° (right) azimuth in two conditions (typical binaural and right ear plugged). In each trial, sound was presented at an initial stationary position (L1) and then while moving at ∼4°/s until reaching a second position (L2). Sound moved in five conditions (±40°, ±20°, or 0°). Participants adjusted a laser pointer to indicate L1 and L2 positions. Unrestricted head and eye movements were collected with gyroscopic sensors on the head and eye-tracking glasses, respectively. Results confirmed that accurate sound localization of both stationary and moving sound is disrupted by acute monaural ear plugging. Eye movements preceded head movements for sound localization in normal binaural listening and head movements were larger than eye movements during monaural plugging. Head movements favored the unplugged left ear when stationary sounds were presented in the right hemifield and during sound motion in both hemifields regardless of the movement direction. Disrupted binaural cues have greater effects on localization of moving than stationary sound. Head movements reveal preferential use of the better-hearing ear and relatively stable eye positions likely reflect normal vestibular-ocular reflexes.

本研究旨在确定儿童和成人在声音定位过程中头部和眼球运动的使用情况,以便:(1)评估静止声音和运动声音的影响;(2)确定通过急性单耳堵塞降低的双耳线索的影响。研究人员招募了具有典型听力的 33 名青少年(平均年龄 = 12.9 岁)和 17 名成人(平均年龄 = 24.6 岁),要求他们在两种条件下(典型双耳和右耳堵塞)定位方位角从 -60°(左)到 +60°(右)水平弧线内任何地方的白噪声。在每次试验中,声音在初始静止位置(L1)出现,然后以 ∼4°/s 的速度移动,直到到达第二个位置(L2)。声音的移动有五种情况(±40°、±20° 或 0°)。参与者通过调整激光指示器来指示 L1 和 L2 位置。头部的陀螺仪传感器和眼球跟踪眼镜分别收集头部和眼球的无限制运动。结果证实,急性单声道耳塞会干扰对静止和移动声音的准确声音定位。在正常双耳聆听时,眼球运动先于头部运动进行声音定位,而在单耳堵塞时,头部运动大于眼球运动。当静止的声音出现在右半球时,头部运动偏向于未插耳的左耳;当声音运动出现在两个半球时,无论运动方向如何,头部运动都偏向于未插耳的左耳。中断的双耳线索对运动声音的定位影响大于静止声音。头部运动显示了听力较好的耳朵的优先使用,而相对稳定的眼球位置可能反映了正常的前庭-眼反射。
{"title":"Head and Eye Movements Reveal Compensatory Strategies for Acute Binaural Deficits During Sound Localization.","authors":"Robel Z Alemu, Blake C Papsin, Robert V Harrison, Al Blakeman, Karen A Gordon","doi":"10.1177/23312165231217910","DOIUrl":"10.1177/23312165231217910","url":null,"abstract":"<p><p>The present study aimed to define use of head and eye movements during sound localization in children and adults to: (1) assess effects of stationary versus moving sound and (2) define effects of binaural cues degraded through acute monaural ear plugging. Thirty-three youth (<i>M</i><sub>Age </sub>= 12.9 years) and seventeen adults (<i>M</i><sub>Age </sub>= 24.6 years) with typical hearing were recruited and asked to localize white noise anywhere within a horizontal arc from -60° (left) to +60° (right) azimuth in two conditions (typical binaural and right ear plugged). In each trial, sound was presented at an initial stationary position (L1) and then while moving at ∼4°/s until reaching a second position (L2). Sound moved in five conditions (±40°, ±20°, or 0°). Participants adjusted a laser pointer to indicate L1 and L2 positions. Unrestricted head and eye movements were collected with gyroscopic sensors on the head and eye-tracking glasses, respectively. Results confirmed that accurate sound localization of both stationary and moving sound is disrupted by acute monaural ear plugging. Eye movements preceded head movements for sound localization in normal binaural listening and head movements were larger than eye movements during monaural plugging. Head movements favored the unplugged left ear when stationary sounds were presented in the right hemifield and during sound motion in both hemifields regardless of the movement direction. Disrupted binaural cues have greater effects on localization of moving than stationary sound. Head movements reveal preferential use of the better-hearing ear and relatively stable eye positions likely reflect normal vestibular-ocular reflexes.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165231217910"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10832417/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139651917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Combining Cardiovascular and Pupil Features Using k-Nearest Neighbor Classifiers to Assess Task Demand, Social Context, and Sentence Accuracy During Listening. 利用 k 近邻分类器结合心血管和瞳孔特征,评估听力过程中的任务需求、社会背景和句子准确性。
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241232551
Bethany Plain, Hidde Pielage, Sophia E Kramer, Michael Richter, Gabrielle H Saunders, Niek J Versfeld, Adriana A Zekveld, Tanveer A Bhuiyan

In daily life, both acoustic factors and social context can affect listening effort investment. In laboratory settings, information about listening effort has been deduced from pupil and cardiovascular responses independently. The extent to which these measures can jointly predict listening-related factors is unknown. Here we combined pupil and cardiovascular features to predict acoustic and contextual aspects of speech perception. Data were collected from 29 adults (mean  =  64.6 years, SD  =  9.2) with hearing loss. Participants performed a speech perception task at two individualized signal-to-noise ratios (corresponding to 50% and 80% of sentences correct) and in two social contexts (the presence and absence of two observers). Seven features were extracted per trial: baseline pupil size, peak pupil dilation, mean pupil dilation, interbeat interval, blood volume pulse amplitude, pre-ejection period and pulse arrival time. These features were used to train k-nearest neighbor classifiers to predict task demand, social context and sentence accuracy. The k-fold cross validation on the group-level data revealed above-chance classification accuracies: task demand, 64.4%; social context, 78.3%; and sentence accuracy, 55.1%. However, classification accuracies diminished when the classifiers were trained and tested on data from different participants. Individually trained classifiers (one per participant) performed better than group-level classifiers: 71.7% (SD  =  10.2) for task demand, 88.0% (SD  =  7.5) for social context, and 60.0% (SD  =  13.1) for sentence accuracy. We demonstrated that classifiers trained on group-level physiological data to predict aspects of speech perception generalized poorly to novel participants. Individually calibrated classifiers hold more promise for future applications.

在日常生活中,声学因素和社会环境都会影响听力投入。在实验室环境中,有关倾听努力的信息是单独从瞳孔和心血管反应中推导出来的。这些测量方法能在多大程度上共同预测与聆听相关的因素,目前还不得而知。在这里,我们结合了瞳孔和心血管特征来预测语音感知的声学和语境方面。我们收集了 29 名听力损失的成年人(平均年龄为 64.6 岁,标准差为 9.2)的数据。受试者在两种个性化信噪比(分别对应 50% 和 80% 的句子正确率)和两种社会情境(有两名观察者在场和没有观察者在场)下完成了一项语音感知任务。每次试验提取七个特征:基线瞳孔大小、瞳孔放大峰值、平均瞳孔放大、搏动间隔、血容量脉搏振幅、射血前周期和脉搏到达时间。这些特征用于训练 k-近邻分类器,以预测任务需求、社会背景和句子准确性。对组级数据进行 k 倍交叉验证后发现,分类准确率高于平均值:任务需求为 64.4%;社会背景为 78.3%;句子准确率为 55.1%。然而,当分类器在来自不同参与者的数据上进行训练和测试时,分类准确率有所下降。单独训练的分类器(每个参与者一个)的表现优于群体级分类器:任务需求为 71.7%(标准差 = 10.2),社会背景为 88.0%(标准差 = 7.5),句子准确率为 60.0%(标准差 = 13.1)。我们证明,根据群体级生理数据训练的分类器在预测言语感知方面对新参与者的通用性很差。单独校准的分类器在未来的应用中更有前景。
{"title":"Combining Cardiovascular and Pupil Features Using k-Nearest Neighbor Classifiers to Assess Task Demand, Social Context, and Sentence Accuracy During Listening.","authors":"Bethany Plain, Hidde Pielage, Sophia E Kramer, Michael Richter, Gabrielle H Saunders, Niek J Versfeld, Adriana A Zekveld, Tanveer A Bhuiyan","doi":"10.1177/23312165241232551","DOIUrl":"10.1177/23312165241232551","url":null,"abstract":"<p><p>In daily life, both acoustic factors and social context can affect listening effort investment. In laboratory settings, information about listening effort has been deduced from pupil and cardiovascular responses independently. The extent to which these measures can jointly predict listening-related factors is unknown. Here we combined pupil and cardiovascular features to predict acoustic and contextual aspects of speech perception. Data were collected from 29 adults (mean  =  64.6 years, SD  =  9.2) with hearing loss. Participants performed a speech perception task at two individualized signal-to-noise ratios (corresponding to 50% and 80% of sentences correct) and in two social contexts (the presence and absence of two observers). Seven features were extracted per trial: baseline pupil size, peak pupil dilation, mean pupil dilation, interbeat interval, blood volume pulse amplitude, pre-ejection period and pulse arrival time. These features were used to train k-nearest neighbor classifiers to predict task demand, social context and sentence accuracy. The k-fold cross validation on the group-level data revealed above-chance classification accuracies: task demand, 64.4%; social context, 78.3%; and sentence accuracy, 55.1%. However, classification accuracies diminished when the classifiers were trained and tested on data from different participants. Individually trained classifiers (one per participant) performed better than group-level classifiers: 71.7% (SD  =  10.2) for task demand, 88.0% (SD  =  7.5) for social context, and 60.0% (SD  =  13.1) for sentence accuracy. We demonstrated that classifiers trained on group-level physiological data to predict aspects of speech perception generalized poorly to novel participants. Individually calibrated classifiers hold more promise for future applications.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241232551"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10981225/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140319548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Cochlear Implants and Music. 社论:人工耳蜗与音乐
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241231685
Deborah A Vickers, Brian C J Moore
{"title":"Editorial: Cochlear Implants and Music.","authors":"Deborah A Vickers, Brian C J Moore","doi":"10.1177/23312165241231685","DOIUrl":"10.1177/23312165241231685","url":null,"abstract":"","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241231685"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10874149/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139742320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Measurement of Speech Recognition, Reaction Time, and Speech Rate and Their Relation to Self-Reported Listening Effort for Normal-Hearing and Hearing-Impaired Listeners Using various Maskers. 使用各种掩码自动测量正常听力和听力受损听者的语音识别能力、反应时间和语速及其与自述听力努力的关系。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241276435
Inga Holube, Stefan Taesler, Saskia Ibelings, Martin Hansen, Jasper Ooster

In speech audiometry, the speech-recognition threshold (SRT) is usually established by adjusting the signal-to-noise ratio (SNR) until 50% of the words or sentences are repeated correctly. However, these conditions are rarely encountered in everyday situations. Therefore, for a group of 15 young participants with normal hearing and a group of 12 older participants with hearing impairment, speech-recognition scores were determined at SRT and at four higher SNRs using several stationary and fluctuating maskers. Participants' verbal responses were recorded, and participants were asked to self-report their listening effort on a categorical scale (self-reported listening effort, SR-LE). The responses were analyzed using an Automatic Speech Recognizer (ASR) and compared to the results of a human examiner. An intraclass correlation coefficient of r = .993 for the agreement between their corresponding speech-recognition scores was observed. As expected, speech-recognition scores increased with increasing SNR and decreased with increasing SR-LE. However, differences between speech-recognition scores for fluctuating and stationary maskers were observed as a function of SNR, but not as a function of SR-LE. The verbal response time (VRT) and the response speech rate (RSR) of the listeners' responses were measured using an ASR. The participants with hearing impairment showed significantly lower RSRs and higher VRTs compared to the participants with normal hearing. These differences may be attributed to differences in age, hearing, or both. With increasing SR-LE, VRT increased and RSR decreased. The results show the possibility of deriving a behavioral measure, VRT, measured directly from participants' verbal responses during speech audiometry, as a proxy for SR-LE.

在言语测听中,通常通过调整信噪比(SNR)来确定言语识别阈值(SRT),直到 50%的单词或句子被正确重复为止。然而,这种情况在日常生活中很少遇到。因此,我们对一组 15 名听力正常的年轻参与者和一组 12 名听力受损的老年参与者进行了测试,在 SRT 和四种较高信噪比条件下,使用几种固定和波动掩蔽器测定了他们的语音识别得分。参与者的口头回答都被记录下来,并要求他们用分类量表(自评听力强度,SR-LE)自我报告听力强度。使用自动语音识别器(ASR)对这些回答进行分析,并将其与人工检查员的结果进行比较。结果表明,相应的语音识别得分之间的类内相关系数为 r = 0.993。正如预期的那样,语音识别得分随着信噪比的增加而增加,随着 SR-LE 的增加而减少。然而,在信噪比(SNR)的函数作用下,可以观察到波动型掩蔽者和静止型掩蔽者的语音识别得分之间存在差异,但在 SR-LE 的函数作用下则没有这种差异。使用 ASR 测量了听者的言语反应时间(VRT)和反应语速(RSR)。与听力正常者相比,听力受损者的 RSR 明显较低,而 VRT 则较高。这些差异可能是由于年龄、听力或两者的差异造成的。随着 SR-LE 的增加,VRT 增加,RSR 减少。研究结果表明,可以从参与者在言语测听过程中的言语反应直接得出行为测量值 VRT,作为 SR-LE 的替代值。
{"title":"Automated Measurement of Speech Recognition, Reaction Time, and Speech Rate and Their Relation to Self-Reported Listening Effort for Normal-Hearing and Hearing-Impaired Listeners Using various Maskers.","authors":"Inga Holube, Stefan Taesler, Saskia Ibelings, Martin Hansen, Jasper Ooster","doi":"10.1177/23312165241276435","DOIUrl":"10.1177/23312165241276435","url":null,"abstract":"<p><p>In speech audiometry, the speech-recognition threshold (SRT) is usually established by adjusting the signal-to-noise ratio (SNR) until 50% of the words or sentences are repeated correctly. However, these conditions are rarely encountered in everyday situations. Therefore, for a group of 15 young participants with normal hearing and a group of 12 older participants with hearing impairment, speech-recognition scores were determined at SRT and at four higher SNRs using several stationary and fluctuating maskers. Participants' verbal responses were recorded, and participants were asked to self-report their listening effort on a categorical scale (self-reported listening effort, SR-LE). The responses were analyzed using an Automatic Speech Recognizer (ASR) and compared to the results of a human examiner. An intraclass correlation coefficient of <i>r </i>= .993 for the agreement between their corresponding speech-recognition scores was observed. As expected, speech-recognition scores increased with increasing SNR and decreased with increasing SR-LE. However, differences between speech-recognition scores for fluctuating and stationary maskers were observed as a function of SNR, but not as a function of SR-LE. The verbal response time (VRT) and the response speech rate (RSR) of the listeners' responses were measured using an ASR. The participants with hearing impairment showed significantly lower RSRs and higher VRTs compared to the participants with normal hearing. These differences may be attributed to differences in age, hearing, or both. With increasing SR-LE, VRT increased and RSR decreased. The results show the possibility of deriving a behavioral measure, VRT, measured directly from participants' verbal responses during speech audiometry, as a proxy for SR-LE.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241276435"},"PeriodicalIF":2.6,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11421406/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142299020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Speech Audiometry: Can It Work Using Open-Source Pre-Trained Kaldi-NL Automatic Speech Recognition? 自动语音测听:使用开源预训练的 Kaldi-NL 自动语音识别技术是否可行?
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241229057
Gloria Araiza-Illan, Luke Meyer, Khiet P Truong, Deniz Başkent

A practical speech audiometry tool is the digits-in-noise (DIN) test for hearing screening of populations of varying ages and hearing status. The test is usually conducted by a human supervisor (e.g., clinician), who scores the responses spoken by the listener, or online, where software scores the responses entered by the listener. The test has 24-digit triplets presented in an adaptive staircase procedure, resulting in a speech reception threshold (SRT). We propose an alternative automated DIN test setup that can evaluate spoken responses whilst conducted without a human supervisor, using the open-source automatic speech recognition toolkit, Kaldi-NL. Thirty self-reported normal-hearing Dutch adults (19-64 years) completed one DIN + Kaldi-NL test. Their spoken responses were recorded and used for evaluating the transcript of decoded responses by Kaldi-NL. Study 1 evaluated the Kaldi-NL performance through its word error rate (WER), percentage of summed decoding errors regarding only digits found in the transcript compared to the total number of digits present in the spoken responses. Average WER across participants was 5.0% (range 0-48%, SD = 8.8%), with average decoding errors in three triplets per participant. Study 2 analyzed the effect that triplets with decoding errors from Kaldi-NL had on the DIN test output (SRT), using bootstrapping simulations. Previous research indicated 0.70 dB as the typical within-subject SRT variability for normal-hearing adults. Study 2 showed that up to four triplets with decoding errors produce SRT variations within this range, suggesting that our proposed setup could be feasible for clinical applications.

噪声中数字(DIN)测试是一种实用的言语测听工具,用于对不同年龄和听力状况的人群进行听力筛查。该测试通常由人工监督员(如临床医生)或在线软件进行,人工监督员会对听者的回答进行评分,在线软件则会对听者输入的回答进行评分。该测试采用自适应阶梯程序呈现 24 位三连音,从而得出语音接收阈值 (SRT)。我们提出了另一种自动 DIN 测试设置,可以在没有人工监督的情况下,使用开源自动语音识别工具包 Kaldi-NL 评估口语回答。30 名自称听力正常的荷兰成年人(19-64 岁)完成了一次 DIN + Kaldi-NL 测试。他们的口语回答被录制下来,用于评估 Kaldi-NL 解码后的回答记录。研究 1 通过 Kaldi-NL 的单词错误率(WER)来评估 Kaldi-NL 的性能,WER 是指与口语回答中出现的数字总数相比,只涉及笔录中出现的数字的解码错误总和所占的百分比。参与者的平均 WER 为 5.0%(范围为 0-48%,SD = 8.8%),每位参与者平均在三个三连音中出现解码错误。研究 2 采用引导模拟法分析了 Kaldi-NL 解码错误的三连音对 DIN 测试输出(SRT)的影响。先前的研究表明,正常听力成人的典型受试者内 SRT 变异为 0.70 dB。研究 2 表明,多达四个三连音解码错误产生的 SRT 变异在此范围内,这表明我们建议的设置在临床应用中是可行的。
{"title":"Automated Speech Audiometry: Can It Work Using Open-Source Pre-Trained Kaldi-NL Automatic Speech Recognition?","authors":"Gloria Araiza-Illan, Luke Meyer, Khiet P Truong, Deniz Başkent","doi":"10.1177/23312165241229057","DOIUrl":"10.1177/23312165241229057","url":null,"abstract":"<p><p>A practical speech audiometry tool is the digits-in-noise (DIN) test for hearing screening of populations of varying ages and hearing status. The test is usually conducted by a human supervisor (e.g., clinician), who scores the responses spoken by the listener, or online, where software scores the responses entered by the listener. The test has 24-digit triplets presented in an adaptive staircase procedure, resulting in a speech reception threshold (SRT). We propose an alternative automated DIN test setup that can evaluate spoken responses whilst conducted without a human supervisor, using the open-source automatic speech recognition toolkit, Kaldi-NL. Thirty self-reported normal-hearing Dutch adults (19-64 years) completed one DIN + Kaldi-NL test. Their spoken responses were recorded and used for evaluating the transcript of decoded responses by Kaldi-NL. Study 1 evaluated the Kaldi-NL performance through its word error rate (WER), percentage of summed decoding errors regarding only digits found in the transcript compared to the total number of digits present in the spoken responses. Average WER across participants was 5.0% (range 0-48%, SD = 8.8%), with average decoding errors in three triplets per participant. Study 2 analyzed the effect that triplets with decoding errors from Kaldi-NL had on the DIN test output (SRT), using bootstrapping simulations. Previous research indicated 0.70 dB as the typical within-subject SRT variability for normal-hearing adults. Study 2 showed that up to four triplets with decoding errors produce SRT variations within this range, suggesting that our proposed setup could be feasible for clinical applications.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241229057"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10943752/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140132882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Review of Binaural Processing With Asymmetrical Hearing Outcomes in Patients With Bilateral Cochlear Implants. 双耳处理与双侧人工耳蜗患者不对称听力结果的回顾。
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241229880
Sean R Anderson, Emily Burg, Lukas Suveg, Ruth Y Litovsky

Bilateral cochlear implants (BiCIs) result in several benefits, including improvements in speech understanding in noise and sound source localization. However, the benefit bilateral implants provide among recipients varies considerably across individuals. Here we consider one of the reasons for this variability: difference in hearing function between the two ears, that is, interaural asymmetry. Thus far, investigations of interaural asymmetry have been highly specialized within various research areas. The goal of this review is to integrate these studies in one place, motivating future research in the area of interaural asymmetry. We first consider bottom-up processing, where binaural cues are represented using excitation-inhibition of signals from the left ear and right ear, varying with the location of the sound in space, and represented by the lateral superior olive in the auditory brainstem. We then consider top-down processing via predictive coding, which assumes that perception stems from expectations based on context and prior sensory experience, represented by cascading series of cortical circuits. An internal, perceptual model is maintained and updated in light of incoming sensory input. Together, we hope that this amalgamation of physiological, behavioral, and modeling studies will help bridge gaps in the field of binaural hearing and promote a clearer understanding of the implications of interaural asymmetry for future research on optimal patient interventions.

双侧人工耳蜗(BiCIs)可带来多种益处,包括改善噪音中的语音理解和声源定位。然而,双侧植入体给受助者带来的益处却因人而异。在此,我们将考虑造成这种差异的原因之一:双耳听力功能的差异,即耳间不对称。迄今为止,对耳间不对称的研究在不同的研究领域都非常专业。本综述的目的是将这些研究整合在一起,激励未来在耳间不对称领域的研究。我们首先考虑的是自下而上的处理过程,其中双耳线索是通过左耳和右耳信号的激发-抑制来表示的,随声音在空间中的位置而变化,并由听觉脑干的外侧上橄榄来表示。然后,我们考虑通过预测编码进行自上而下的处理,即假定知觉源于基于上下文和先前感官经验的预期,由一系列层叠的皮层电路表示。内部感知模型会根据输入的感官信息进行维护和更新。我们希望,将生理学、行为学和建模研究结合在一起,将有助于弥补双耳听力领域的不足,并促进人们更清楚地了解耳间不对称对未来最佳患者干预研究的影响。
{"title":"Review of Binaural Processing With Asymmetrical Hearing Outcomes in Patients With Bilateral Cochlear Implants.","authors":"Sean R Anderson, Emily Burg, Lukas Suveg, Ruth Y Litovsky","doi":"10.1177/23312165241229880","DOIUrl":"10.1177/23312165241229880","url":null,"abstract":"<p><p>Bilateral cochlear implants (BiCIs) result in several benefits, including improvements in speech understanding in noise and sound source localization. However, the benefit bilateral implants provide among recipients varies considerably across individuals. Here we consider one of the reasons for this variability: difference in hearing function between the two ears, that is, interaural asymmetry. Thus far, investigations of interaural asymmetry have been highly specialized within various research areas. The goal of this review is to integrate these studies in one place, motivating future research in the area of interaural asymmetry. We first consider bottom-up processing, where binaural cues are represented using excitation-inhibition of signals from the left ear and right ear, varying with the location of the sound in space, and represented by the lateral superior olive in the auditory brainstem. We then consider top-down processing via predictive coding, which assumes that perception stems from expectations based on context and prior sensory experience, represented by cascading series of cortical circuits. An internal, perceptual model is maintained and updated in light of incoming sensory input. Together, we hope that this amalgamation of physiological, behavioral, and modeling studies will help bridge gaps in the field of binaural hearing and promote a clearer understanding of the implications of interaural asymmetry for future research on optimal patient interventions.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241229880"},"PeriodicalIF":2.7,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10976506/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140307503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating Pitch Information From Simulated Cochlear Implant Signals With Deep Neural Networks. 利用深度神经网络从模拟人工耳蜗信号中估计音高信息
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-01-01 DOI: 10.1177/23312165241298606
Takanori Ashihara, Shigeto Furukawa, Makio Kashino

Cochlear implant (CI) users, even with substantial speech comprehension, generally have poor sensitivity to pitch information (or fundamental frequency, F0). This insensitivity is often attributed to limited spectral and temporal resolution in the CI signals. However, the pitch sensitivity markedly varies among individuals, and some users exhibit fairly good sensitivity. This indicates that the CI signal contains sufficient information about F0, and users' sensitivity is predominantly limited by other physiological conditions such as neuroplasticity or neural health. We estimated the upper limit of F0 information that a CI signal can convey by decoding F0 from simulated CI signals (multi-channel pulsatile signals) with a deep neural network model (referred to as the CI model). We varied the number of electrode channels and the pulse rate, which should respectively affect spectral and temporal resolutions of stimulus representations. The F0-estimation performance generally improved with increasing number of channels and pulse rate. For the sounds presented under quiet conditions, the model performance was at best comparable to that of a control waveform model, which received raw-waveform inputs. Under conditions in which background noise was imposed, the performance of the CI model generally degraded by a greater degree than that of the waveform model. The pulse rate had a particularly large effect on predicted performance. These observations indicate that the CI signal contains some information for predicting F0, which is particularly sufficient for targets under quiet conditions. The temporal resolution (represented as pulse rate) plays a critical role in pitch representation under noisy conditions.

人工耳蜗 (CI) 用户即使有很强的语音理解能力,一般对音高信息(或基频,F0)的敏感度也很低。这种不敏感通常归因于 CI 信号的频谱和时间分辨率有限。然而,不同个体的音调灵敏度存在明显差异,有些用户的灵敏度相当高。这表明 CI 信号包含足够的 F0 信息,而用户的灵敏度主要受到神经可塑性或神经健康等其他生理条件的限制。我们通过使用深度神经网络模型(简称 CI 模型)对模拟 CI 信号(多通道脉动信号)进行 F0 解码,从而估算出 CI 信号所能传达的 F0 信息上限。我们改变了电极通道的数量和脉冲频率,这将分别影响刺激表征的频谱和时间分辨率。随着通道数和脉冲频率的增加,F0 估算性能普遍提高。对于在安静条件下呈现的声音,模型性能最多只能与接收原始波形输入的对照波形模型相媲美。在有背景噪音的条件下,CI 模型的性能通常比波形模型的性能下降得更多。脉搏率对预测性能的影响尤其大。这些观察结果表明,CI 信号包含一些预测 F0 的信息,尤其是对安静条件下的目标而言,这些信息是足够的。时间分辨率(以脉搏率表示)在噪声条件下的音高表示中起着至关重要的作用。
{"title":"Estimating Pitch Information From Simulated Cochlear Implant Signals With Deep Neural Networks.","authors":"Takanori Ashihara, Shigeto Furukawa, Makio Kashino","doi":"10.1177/23312165241298606","DOIUrl":"https://doi.org/10.1177/23312165241298606","url":null,"abstract":"<p><p>Cochlear implant (CI) users, even with substantial speech comprehension, generally have poor sensitivity to pitch information (or fundamental frequency, F0). This insensitivity is often attributed to limited spectral and temporal resolution in the CI signals. However, the pitch sensitivity markedly varies among individuals, and some users exhibit fairly good sensitivity. This indicates that the CI signal contains sufficient information about F0, and users' sensitivity is predominantly limited by other physiological conditions such as neuroplasticity or neural health. We estimated the upper limit of F0 information that a CI signal can convey by decoding F0 from simulated CI signals (multi-channel pulsatile signals) with a deep neural network model (referred to as the CI model). We varied the number of electrode channels and the pulse rate, which should respectively affect spectral and temporal resolutions of stimulus representations. The F0-estimation performance generally improved with increasing number of channels and pulse rate. For the sounds presented under quiet conditions, the model performance was at best comparable to that of a control waveform model, which received raw-waveform inputs. Under conditions in which background noise was imposed, the performance of the CI model generally degraded by a greater degree than that of the waveform model. The pulse rate had a particularly large effect on predicted performance. These observations indicate that the CI signal contains some information for predicting F0, which is particularly sufficient for targets under quiet conditions. The temporal resolution (represented as pulse rate) plays a critical role in pitch representation under noisy conditions.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"28 ","pages":"23312165241298606"},"PeriodicalIF":2.6,"publicationDate":"2024-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142683025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Trends in Hearing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1