首页 > 最新文献

Trends in Hearing最新文献

英文 中文
Association of Tinnitus With Speech Recognition and Executive Functions in Older Adults. 耳鸣与老年人语音识别和执行功能的关系。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-11-13 DOI: 10.1177/23312165251389585
Nick Sommerhalder, Zbyněk Bureš, Oliver Profant, Tobias Kleinjung, Patrick Neff, Martin Meyer

Adults with chronic subjective tinnitus often struggle with speech recognition in challenging listening environments. While most research demonstrates deficits in speech recognition among individuals with tinnitus, studies focusing on older adults remain scarce. Besides speech recognition deficits, tinnitus has been linked to diminished cognitive performance, particularly in executive functions, yet its associations with specific cognitive domains in ageing populations are not fully understood. Our previous study of younger adults found that individuals with tinnitus exhibit deficits in speech recognition and interference control. Building on this, we hypothesized that these deficits are also present for older adults. We conducted a cross-sectional study of older adults (aged 60-79), 32 with tinnitus and 31 controls matched for age, gender, education, and approximately matched for hearing loss. Participants underwent audiometric, speech recognition, and cognitive tasks. The tinnitus participants performed more poorly in speech-in-noise and gated speech tasks, whereas no group differences were observed in the other suprathreshold auditory tasks. With regard to cognition, individuals with tinnitus showed reduced interference control, emotional interference, cognitive flexibility, and verbal working memory, correlating with tinnitus distress and loudness. It is concluded that tinnitus-related deficits persist and even worsen with age. Our results suggest that altered central mechanisms contribute to speech recognition difficulties in older adults with tinnitus.

患有慢性主观性耳鸣的成年人经常在具有挑战性的听力环境中挣扎于语音识别。虽然大多数研究表明耳鸣患者在语音识别方面存在缺陷,但针对老年人的研究仍然很少。除了语音识别缺陷外,耳鸣还与认知能力下降,特别是执行功能下降有关,但其与老年人特定认知领域的关系尚不完全清楚。我们之前对年轻人的研究发现,耳鸣患者在语音识别和干扰控制方面表现出缺陷。在此基础上,我们假设老年人也存在这些缺陷。我们对老年人(60-79岁)进行了一项横断面研究,其中32名耳鸣患者和31名年龄、性别、教育程度和听力损失大致匹配的对照组。参与者接受了听力测量、语音识别和认知任务。耳鸣参与者在噪音语音和门控语音任务中表现更差,而在其他超阈值听觉任务中没有观察到组间差异。在认知方面,耳鸣个体表现出干扰控制、情绪干扰、认知灵活性和言语工作记忆的减少,与耳鸣痛苦和响度相关。结论是,耳鸣相关的缺陷持续存在,甚至随着年龄的增长而恶化。我们的研究结果表明,中枢机制的改变导致了老年耳鸣患者的语音识别困难。
{"title":"Association of Tinnitus With Speech Recognition and Executive Functions in Older Adults.","authors":"Nick Sommerhalder, Zbyněk Bureš, Oliver Profant, Tobias Kleinjung, Patrick Neff, Martin Meyer","doi":"10.1177/23312165251389585","DOIUrl":"10.1177/23312165251389585","url":null,"abstract":"<p><p>Adults with chronic subjective tinnitus often struggle with speech recognition in challenging listening environments. While most research demonstrates deficits in speech recognition among individuals with tinnitus, studies focusing on older adults remain scarce. Besides speech recognition deficits, tinnitus has been linked to diminished cognitive performance, particularly in executive functions, yet its associations with specific cognitive domains in ageing populations are not fully understood. Our previous study of younger adults found that individuals with tinnitus exhibit deficits in speech recognition and interference control. Building on this, we hypothesized that these deficits are also present for older adults. We conducted a cross-sectional study of older adults (aged 60-79), 32 with tinnitus and 31 controls matched for age, gender, education, and approximately matched for hearing loss. Participants underwent audiometric, speech recognition, and cognitive tasks. The tinnitus participants performed more poorly in speech-in-noise and gated speech tasks, whereas no group differences were observed in the other suprathreshold auditory tasks. With regard to cognition, individuals with tinnitus showed reduced interference control, emotional interference, cognitive flexibility, and verbal working memory, correlating with tinnitus distress and loudness. It is concluded that tinnitus-related deficits persist and even worsen with age. Our results suggest that altered central mechanisms contribute to speech recognition difficulties in older adults with tinnitus.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251389585"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12615926/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145514780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
At-Home Auditory Assessment Using Portable Automated Rapid Testing (PART) to Understand Self-Reported Hearing Difficulties. 使用便携式自动快速测试的家庭听力评估(部分)来了解自我报告的听力困难。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-11-24 DOI: 10.1177/23312165251397373
E Sebastian Lelo de Larrea-Mancera, Tess K Koerner, William J Bologna, Sara Momtaz, Katherine N Menon, Audrey Carrillo, Eric C Hoover, G Christopher Stecker, Frederick J Gallun, Aaron R Seitz

Previous research has demonstrated that remote testing of suprathreshold auditory function using distributed technologies can produce results that closely match those obtained in laboratory settings with specialized, calibrated equipment. This work has facilitated the validation of various behavioral measures in remote settings that provide valuable insights into auditory function. In the current study, we sought to address whether a broad battery of auditory assessments could explain variance in self-report of hearing handicap. To address this, we used a portable psychophysics assessment tool along with an online recruitment tool (Prolific) to collect auditory task data from participants with (n= 84) and without (n= 108) self-reported hearing difficulty. Results indicate several measures of auditory processing differentiate participants with and without self-reported hearing difficulty. In addition, we report the factor structure of the test battery to clarify the underlying constructs and the extent to which they individually or jointly inform hearing function. Relationships between measures of auditory processing were found to be largely consistent with a hypothesized construct model that guided task selection. Overall, this study advances our understanding of the relationship between auditory and cognitive processing in those with and without subjective hearing difficulty. More broadly, these results indicate promise that these measures can be used in larger scale research studies in remote settings and have potential to contribute to telehealth approaches to better address people's hearing needs.

先前的研究表明,使用分布式技术对超阈听觉功能进行远程测试,可以产生与实验室设置中使用专业校准设备获得的结果密切匹配的结果。这项工作促进了远程环境中各种行为测量的验证,为听觉功能提供了有价值的见解。在当前的研究中,我们试图解决广泛的听觉评估是否可以解释听力障碍自我报告的差异。为了解决这个问题,我们使用了便携式心理物理评估工具和在线招募工具(多产),从有(n = 84)和没有(n = 108)自我报告听力困难的参与者中收集听觉任务数据。结果表明,听觉处理的几个措施区分参与者有和没有自我报告的听力困难。此外,我们报告了测试电池的因素结构,以澄清潜在的结构以及它们单独或共同通知听力功能的程度。听觉处理测量之间的关系被发现在很大程度上与指导任务选择的假设结构模型一致。总的来说,本研究促进了我们对主观听力障碍和非主观听力障碍的听觉和认知加工之间关系的理解。更广泛地说,这些结果表明,这些措施有望用于远程环境中的更大规模研究,并有可能促进远程保健方法,以更好地满足人们的听力需求。
{"title":"At-Home Auditory Assessment Using Portable Automated Rapid Testing (PART) to Understand Self-Reported Hearing Difficulties.","authors":"E Sebastian Lelo de Larrea-Mancera, Tess K Koerner, William J Bologna, Sara Momtaz, Katherine N Menon, Audrey Carrillo, Eric C Hoover, G Christopher Stecker, Frederick J Gallun, Aaron R Seitz","doi":"10.1177/23312165251397373","DOIUrl":"10.1177/23312165251397373","url":null,"abstract":"<p><p>Previous research has demonstrated that remote testing of suprathreshold auditory function using distributed technologies can produce results that closely match those obtained in laboratory settings with specialized, calibrated equipment. This work has facilitated the validation of various behavioral measures in remote settings that provide valuable insights into auditory function. In the current study, we sought to address whether a broad battery of auditory assessments could explain variance in self-report of hearing handicap. To address this, we used a portable psychophysics assessment tool along with an online recruitment tool (Prolific) to collect auditory task data from participants with (<i>n</i> <i>=</i> 84) and without (<i>n</i> <i>=</i> 108) self-reported hearing difficulty. Results indicate several measures of auditory processing differentiate participants with and without self-reported hearing difficulty. In addition, we report the factor structure of the test battery to clarify the underlying constructs and the extent to which they individually or jointly inform hearing function. Relationships between measures of auditory processing were found to be largely consistent with a hypothesized construct model that guided task selection. Overall, this study advances our understanding of the relationship between auditory and cognitive processing in those with and without subjective hearing difficulty. More broadly, these results indicate promise that these measures can be used in larger scale research studies in remote settings and have potential to contribute to telehealth approaches to better address people's hearing needs.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251397373"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12644446/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145597487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Objective Evaluation of a Deep Learning-Based Noise Reduction Algorithm for Hearing Aids Under Diverse Fitting and Listening Conditions. 不同拟合和听力条件下基于深度学习的助听器降噪算法的客观评价。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-11-25 DOI: 10.1177/23312165251396644
Vahid Ashkanichenarlogh, Paula Folkeard, Susan Scollie, Volker Kühnel, Vijay Parsa

This study evaluated a deep-neural-network denoising system using model-based design, comparing it with adaptive filtering and beamforming across various noise types, SNRs, and hearing-aid fittings. A KEMAR manikin fitted with five audiograms was recorded in reverberant and non-reverberant rooms, yielding 1,152 recordings. Speech intelligibility was estimated using the HASPI from 1,152 KEMAR manikin recordings. Effects of processing strategy and acoustic factors were tested with model-based within-device design that account for repeated recordings per device/program and fitting. Linear mixed model results showed that the DNN with beamforming outperformed conventional processing, with strongest gains at 0 and +5 dB SNR, moderate benefits at -5 dB in low reverberation, and none in medium reverberation. Across SNRs and noise types, the DNN combined with beamforming yielded the highest predicted intelligibility, with benefits attenuated under moderate reverberation. Azimuth effects varied; because estimates were derived from a better-ear metric on manikin recordings. Additionally, this paper reports comparisons using metrics of sound quality, for an intrusive metric (HASQI) and the pMOS non-intrusive metric. Results indicated that model type interacted with processing and acoustic factors. HASQI and pMOS scores increased with SNR and were moderately correlated (r² ≈ 0.479), supporting the use of non-intrusive metrics for large-scale assessment. However, pMOS showed greater variability across hearing aid programs and environments, suggesting non-intrusive models capture processing effects differently than intrusive metrics. These findings highlight the promise and limits of non-intrusive evaluation while emphasizing the benefit of combining deep learning with beamforming to improve intelligibility and quality.

本研究使用基于模型的设计评估了一种深度神经网络去噪系统,并将其与各种噪声类型、信噪比和助听器配件的自适应滤波和波束形成进行了比较。在混响和非混响房间中记录了一个装有5个听音仪的KEMAR人体模型,产生了1152个录音。使用来自1152个KEMAR人体录音的HASPI来估计语音可理解性。处理策略和声学因素的影响通过基于模型的设备内设计进行测试,该设计考虑了每个设备/程序和拟合的重复录音。线性混合模型结果表明,波束形成的深度神经网络优于常规处理,在0和+5 dB信噪比下增益最大,在低混响条件下-5 dB增益中等,在中混响条件下没有增益。在信噪比和噪声类型中,深度神经网络与波束形成相结合产生了最高的预测可理解性,在中度混响下其优势减弱。方位效应各不相同;因为估计是根据一个更好的耳朵度量模型录音得出的。此外,本文还报道了使用音质指标的比较,用于侵入性指标(HASQI)和pMOS非侵入性指标。结果表明,模型类型与加工和声学因素相互作用。HASQI和pMOS评分随信噪比增加而增加,并呈中等相关性(r²≈0.479),支持使用非侵入性指标进行大规模评估。然而,pMOS在助听器项目和环境中表现出更大的可变性,这表明非侵入性模型捕获的处理效果与侵入性指标不同。这些发现突出了非侵入性评估的前景和局限性,同时强调了将深度学习与波束成形相结合以提高可理解性和质量的好处。
{"title":"Objective Evaluation of a Deep Learning-Based Noise Reduction Algorithm for Hearing Aids Under Diverse Fitting and Listening Conditions.","authors":"Vahid Ashkanichenarlogh, Paula Folkeard, Susan Scollie, Volker Kühnel, Vijay Parsa","doi":"10.1177/23312165251396644","DOIUrl":"10.1177/23312165251396644","url":null,"abstract":"<p><p>This study evaluated a deep-neural-network denoising system using model-based design, comparing it with adaptive filtering and beamforming across various noise types, SNRs, and hearing-aid fittings. A KEMAR manikin fitted with five audiograms was recorded in reverberant and non-reverberant rooms, yielding 1,152 recordings. Speech intelligibility was estimated using the HASPI from 1,152 KEMAR manikin recordings. Effects of processing strategy and acoustic factors were tested with model-based within-device design that account for repeated recordings per device/program and fitting. Linear mixed model results showed that the DNN with beamforming outperformed conventional processing, with strongest gains at 0 and +5 dB SNR, moderate benefits at -5 dB in low reverberation, and none in medium reverberation. Across SNRs and noise types, the DNN combined with beamforming yielded the highest predicted intelligibility, with benefits attenuated under moderate reverberation. Azimuth effects varied; because estimates were derived from a better-ear metric on manikin recordings. Additionally, this paper reports comparisons using metrics of sound quality, for an intrusive metric (HASQI) and the pMOS non-intrusive metric. Results indicated that model type interacted with processing and acoustic factors. HASQI and pMOS scores increased with SNR and were moderately correlated (r² ≈ 0.479), supporting the use of non-intrusive metrics for large-scale assessment. However, pMOS showed greater variability across hearing aid programs and environments, suggesting non-intrusive models capture processing effects differently than intrusive metrics. These findings highlight the promise and limits of non-intrusive evaluation while emphasizing the benefit of combining deep learning with beamforming to improve intelligibility and quality.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251396644"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12647563/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145606795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Objectively Measuring Audiovisual Effects in Noise Using Virtual Human Speakers. 利用虚拟人说话者客观测量噪声中的视听效果。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-04-13 DOI: 10.1177/23312165251333528
John Kyle Cooper, Jonas Vanthornhout, Astrid van Wieringen, Tom Francart

Speech intelligibility in challenging listening environments relies on the integration of audiovisual cues. Measuring the effectiveness of audiovisual integration in these challenging listening environments can be difficult due to the complexity of such environments. The Audiovisual True-to-Life Assessment of Auditory Rehabilitation (AVATAR) is a paradigm that was developed to provide an ecological environment to capture both the audio and visual aspects of speech intelligibility measures. Previous research has shown the benefit from audiovisual cues can be measured using behavioral (e.g., word recognition) and electrophysiological (e.g., neural tracking) measures. The current research examines, when using the AVATAR paradigm, if electrophysiological measures of speech intelligibility yield similar outcomes as behavioral measures. We hypothesized visual cues would enhance both the behavioral and electrophysiological scores as the signal-to-noise ratio (SNR) of the speech signal decreased. Twenty young (18-25 years old) participants (1 male and 19 female) with normal hearing participated in our study. For our behavioral experiment, we administered lists of sentences using an adaptive procedure to estimate a speech reception threshold (SRT). For our electrophysiological experiment, we administered 35 lists of sentences randomized across five SNR levels (silence, 0, -3, -6, and -9 dB) and two visual conditions (audio-only and audiovisual). We used a neural tracking decoder to measure the reconstruction accuracies for each participant. We observed most participants had higher reconstruction accuracies for the audiovisual condition compared to the audio-only condition in conditions with moderate to high levels of noise. We found the electrophysiological measure may correlate with the behavioral measure that shows audiovisual benefit.

在具有挑战性的听力环境中,语音清晰度依赖于视听线索的整合。由于这些环境的复杂性,在这些具有挑战性的听力环境中测量视听整合的有效性可能很困难。听觉康复的视听逼真评估(AVATAR)是一个范例,旨在提供一个生态环境来捕捉语音可理解性措施的听觉和视觉方面。先前的研究表明,视听线索的好处可以通过行为(例如,单词识别)和电生理(例如,神经跟踪)测量来衡量。当前的研究在使用AVATAR范式时检验了语音可理解性的电生理测量是否与行为测量产生相似的结果。我们假设当语音信号的信噪比(SNR)降低时,视觉线索会提高行为和电生理评分。20名听力正常的年轻人(18-25岁)(男1名,女19名)参加了我们的研究。在我们的行为实验中,我们使用自适应程序管理句子列表来估计语音接收阈值(SRT)。在我们的电生理实验中,我们管理了35个句子列表,随机分为5种信噪比水平(沉默,0,-3,-6和-9 dB)和两种视觉条件(纯音频和视听)。我们使用神经跟踪解码器来测量每个参与者的重建精度。我们观察到,在中等到高水平的噪声条件下,大多数参与者在视听条件下的重建精度高于纯音频条件。我们发现电生理测量可能与显示视听效益的行为测量相关。
{"title":"Objectively Measuring Audiovisual Effects in Noise Using Virtual Human Speakers.","authors":"John Kyle Cooper, Jonas Vanthornhout, Astrid van Wieringen, Tom Francart","doi":"10.1177/23312165251333528","DOIUrl":"https://doi.org/10.1177/23312165251333528","url":null,"abstract":"<p><p>Speech intelligibility in challenging listening environments relies on the integration of audiovisual cues. Measuring the effectiveness of audiovisual integration in these challenging listening environments can be difficult due to the complexity of such environments. The Audiovisual True-to-Life Assessment of Auditory Rehabilitation (AVATAR) is a paradigm that was developed to provide an ecological environment to capture both the audio and visual aspects of speech intelligibility measures. Previous research has shown the benefit from audiovisual cues can be measured using behavioral (e.g., word recognition) and electrophysiological (e.g., neural tracking) measures. The current research examines, when using the AVATAR paradigm, if electrophysiological measures of speech intelligibility yield similar outcomes as behavioral measures. We hypothesized visual cues would enhance both the behavioral and electrophysiological scores as the signal-to-noise ratio (SNR) of the speech signal decreased. Twenty young (18-25 years old) participants (1 male and 19 female) with normal hearing participated in our study. For our behavioral experiment, we administered lists of sentences using an adaptive procedure to estimate a speech reception threshold (SRT). For our electrophysiological experiment, we administered 35 lists of sentences randomized across five SNR levels (silence, 0, -3, -6, and -9 dB) and two visual conditions (audio-only and audiovisual). We used a neural tracking decoder to measure the reconstruction accuracies for each participant. We observed most participants had higher reconstruction accuracies for the audiovisual condition compared to the audio-only condition in conditions with moderate to high levels of noise. We found the electrophysiological measure may correlate with the behavioral measure that shows audiovisual benefit.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251333528"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12033406/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144043708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measuring Speech Discrimination Ability in Sleeping Infants Using fNIRS-A Proof of Principle. 用fnirs测量睡眠婴儿言语辨别能力的原理证明。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 DOI: 10.1177/23312165241311721
Onn Wah Lee, Demi Gao, Tommy Peng, Julia Wunderlich, Darren Mao, Gautam Balasubramanian, Colette M McKay

This study used functional near-infrared spectroscopy (fNIRS) to measure aspects of the speech discrimination ability of sleeping infants. We examined the morphology of the fNIRS response to three different speech contrasts, namely "Tea/Ba," "Bee/Ba," and "Ga/Ba." Sixteen infants aged between 3 and 13 months old were included in this study and their fNIRS data were recorded during natural sleep. The stimuli were presented using a nonsilence baseline paradigm, where repeated standard stimuli were presented between the novel stimuli blocks without any silence periods. The morphology of fNIRS responses varied between speech contrasts. The data were fit with a model in which the responses were the sum of two independent and concurrent response mechanisms that were derived from previously published fNIRS detection responses. These independent components were an oxyhemoglobin (HbO)-positive early-latency response and an HbO-negative late latency response, hypothesized to be related to an auditory canonical response and a brain arousal response, respectively. The goodness of fit of the model with the data was high with median goodness of fit of 81%. The data showed that both response components had later latency when the left ear was the test ear (p < .05) compared to the right ear and that the negative component, due to brain arousal, was smallest for the most subtle contrast, "Ga/Ba" (p = .003).

本研究使用功能近红外光谱(fNIRS)来测量睡眠婴儿的语言识别能力。我们研究了三种不同语音对比的fNIRS反应形态,即“Tea/Ba”、“Bee/Ba”和“Ga/Ba”。16名年龄在3到13个月之间的婴儿参与了这项研究,并记录了他们在自然睡眠期间的近红外光谱数据。刺激采用非沉默基线范式,在新刺激块之间重复呈现标准刺激,没有任何沉默期。不同语音对照的fNIRS反应形态不同。数据拟合了一个模型,其中响应是两个独立且并发的响应机制的总和,这些响应机制来源于先前发表的fNIRS检测响应。这些独立的成分是一个氧合血红蛋白(HbO)阳性的早期潜伏期反应和一个氧合血红蛋白阴性的晚期潜伏期反应,假设分别与听觉规范反应和大脑唤醒反应有关。模型与数据的拟合优度较高,中位拟合优度为81%。数据显示,当左耳为测试耳时,两种反应成分的潜伏期均较晚(p p = 0.003)。
{"title":"Measuring Speech Discrimination Ability in Sleeping Infants Using fNIRS-A Proof of Principle.","authors":"Onn Wah Lee, Demi Gao, Tommy Peng, Julia Wunderlich, Darren Mao, Gautam Balasubramanian, Colette M McKay","doi":"10.1177/23312165241311721","DOIUrl":"10.1177/23312165241311721","url":null,"abstract":"<p><p>This study used functional near-infrared spectroscopy (fNIRS) to measure aspects of the speech discrimination ability of sleeping infants. We examined the morphology of the fNIRS response to three different speech contrasts, namely \"Tea/Ba,\" \"Bee/Ba,\" and \"Ga/Ba.\" Sixteen infants aged between 3 and 13 months old were included in this study and their fNIRS data were recorded during natural sleep. The stimuli were presented using a nonsilence baseline paradigm, where repeated standard stimuli were presented between the novel stimuli blocks without any silence periods. The morphology of fNIRS responses varied between speech contrasts. The data were fit with a model in which the responses were the sum of two independent and concurrent response mechanisms that were derived from previously published fNIRS detection responses. These independent components were an oxyhemoglobin (HbO)-positive early-latency response and an HbO-negative late latency response, hypothesized to be related to an auditory canonical response and a brain arousal response, respectively. The goodness of fit of the model with the data was high with median goodness of fit of 81%. The data showed that both response components had later latency when the left ear was the test ear (<i>p</i> < .05) compared to the right ear and that the negative component, due to brain arousal, was smallest for the most subtle contrast, \"Ga/Ba\" (<i>p</i> = .003).</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165241311721"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11758514/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143030151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of Speaker-Conditioned Target Speaker Extraction Algorithms for Hearing-Impaired Listeners. 针对听障听众的说话人条件目标说话人提取算法评价。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-08-11 DOI: 10.1177/23312165251365802
Ragini Sinha, Ann-Christin Scherer, Simon Doclo, Christian Rollwage, Jan Rennies

Speaker-conditioned target speaker extraction algorithms aim at extracting the target speaker from a mixture of multiple speakers by using additional information about the target speaker. Previous studies have evaluated the performance of these algorithms using either instrumental measures or subjective assessments with normal-hearing listeners or with hearing-impaired listeners. Notably, a previous study employing a quasicausal algorithm reported significant intelligibility improvements for both normal-hearing and hearing-impaired listeners, while another study demonstrated that a fully causal algorithm could enhance speech intelligibility and reduce listening effort for normal-hearing listeners. Building on these findings, this study focuses on an in-depth subjective assessment of two fully causal deep neural network-based speaker-conditioned target speaker extraction algorithms with hearing-impaired listeners, both without hearing loss compensation (unaided) and with linear hearing loss compensation (aided). Three different subjective performance measurement methods were used to cover a broad range of listening conditions, namely paired comparison, speech recognition thresholds, and categorically scaled perceived listening effort. The subjective evaluation results with 15 hearing-impaired listeners showed that one algorithm significantly reduced listening effort and improved intelligibility compared to unprocessed stimuli and the other algorithm. The data also suggest that hearing-impaired listeners experience a greater benefit in terms of listening effort (for both male and female interfering speakers) and speech recognition thresholds, especially in the presence of female interfering speakers than normal-hearing listeners, and that hearing loss compensation (linear amplification) is not required to obtain an algorithm benefit.

基于说话人条件的目标说话人提取算法旨在利用目标说话人的附加信息从混合的多个说话人中提取目标说话人。以前的研究使用仪器测量或主观评估对听力正常或听力受损的听众评估这些算法的性能。值得注意的是,先前一项采用准因果算法的研究报告了正常听力和听力受损听众的可理解性显著提高,而另一项研究表明,完全因果算法可以提高正常听力听众的语音可理解性,减少听力努力。在这些发现的基础上,本研究重点对两种完全因果的基于深度神经网络的说话人条件目标说话人提取算法进行了深入的主观评估,这两种算法都是针对听力受损的听众,没有听力损失补偿(无辅助)和线性听力损失补偿(辅助)。我们使用了三种不同的主观表现测量方法来涵盖广泛的听力条件,即配对比较、语音识别阈值和分类缩放的感知听力努力。对15名听障听众的主观评价结果表明,与未处理的刺激和另一种算法相比,一种算法显著减少了听力努力,提高了可理解性。数据还表明,听力受损的听众在听力努力(男性和女性干扰扬声器)和语音识别阈值方面比正常听力的听众获得更大的好处,特别是在女性干扰扬声器存在的情况下,并且不需要听力损失补偿(线性放大)来获得算法优势。
{"title":"Evaluation of Speaker-Conditioned Target Speaker Extraction Algorithms for Hearing-Impaired Listeners.","authors":"Ragini Sinha, Ann-Christin Scherer, Simon Doclo, Christian Rollwage, Jan Rennies","doi":"10.1177/23312165251365802","DOIUrl":"10.1177/23312165251365802","url":null,"abstract":"<p><p>Speaker-conditioned target speaker extraction algorithms aim at extracting the target speaker from a mixture of multiple speakers by using additional information about the target speaker. Previous studies have evaluated the performance of these algorithms using either instrumental measures or subjective assessments with normal-hearing listeners or with hearing-impaired listeners. Notably, a previous study employing a quasicausal algorithm reported significant intelligibility improvements for both normal-hearing and hearing-impaired listeners, while another study demonstrated that a fully causal algorithm could enhance speech intelligibility and reduce listening effort for normal-hearing listeners. Building on these findings, this study focuses on an in-depth subjective assessment of two fully causal deep neural network-based speaker-conditioned target speaker extraction algorithms with hearing-impaired listeners, both without hearing loss compensation (unaided) and with linear hearing loss compensation (aided). Three different subjective performance measurement methods were used to cover a broad range of listening conditions, namely paired comparison, speech recognition thresholds, and categorically scaled perceived listening effort. The subjective evaluation results with 15 hearing-impaired listeners showed that one algorithm significantly reduced listening effort and improved intelligibility compared to unprocessed stimuli and the other algorithm. The data also suggest that hearing-impaired listeners experience a greater benefit in terms of listening effort (for both male and female interfering speakers) and speech recognition thresholds, especially in the presence of female interfering speakers than normal-hearing listeners, and that hearing loss compensation (linear amplification) is not required to obtain an algorithm benefit.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251365802"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12340209/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144817996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Time Course of the Pupillary Response to Auditory Emotions in Pseudospeech, Music, and Vocalizations. 假性言语、音乐和发声中瞳孔对听觉情绪反应的时间过程。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-08-10 DOI: 10.1177/23312165251365824
Julie Kirwan, Deniz Başkent, Anita Wagner

Emotions can be communicated through visual and dynamic characteristics such as smiles and gestures, but also through auditory channels such as laughter, music, and human speech. Pupil dilation has become a notable marker for visual emotion processing; however the pupil's sensitivity to emotional sounds, specifically speech, remains largely underexplored. This study investigated the processing of emotional pseudospeech, which are speech-like sentences devoid of semantic content. We measured participants' pupil dilations while they listened to pseudospeech, music, and human vocalizations, and subsequently performed an emotion recognition task. Our results showed that emotional pseudospeech can trigger increases of pupil dilation compared to neutral pseudospeech, supporting the use of pupillometry as a tool for indexing prosodic emotion processing in the absence of semantics. However, pupil responses to pseudospeech were smaller and slower than the responses evoked by human vocalizations. The pupillary response was not sensitive enough to distinguish between emotion categories in pseudospeech, but pupil dilations to music and vocalizations reflected some emotion-specific pupillary curves. The valence of the stimulus had a stronger overall influence on pupil size than arousal. These results highlight the potential for pupillometry in studying auditory emotion processing and provide a foundation for contextualizing pseudospeech alongside other affective auditory stimuli.

情感可以通过视觉和动态特征(如微笑和手势)进行交流,也可以通过听觉渠道(如笑声、音乐和人类语言)进行交流。瞳孔扩张已成为视觉情绪加工的显著标志;然而,学生对情感声音的敏感性,特别是言语,在很大程度上仍未得到充分研究。情感伪言语是一种缺乏语义内容的类言语句子,研究了情感伪言语的加工过程。我们测量了参与者在听假话、音乐和人类发声时的瞳孔扩张,并随后执行了一项情绪识别任务。我们的研究结果表明,与中性假语相比,情绪性假语可以触发瞳孔扩张的增加,这支持了在没有语义的情况下使用瞳孔测量法作为索引韵律情绪处理的工具。然而,瞳孔对假言语的反应比人类发声引起的反应更小、更慢。假性言语的瞳孔反应不够敏感,无法区分情绪类别,但对音乐和发声的瞳孔扩张反映了一些特定情绪的瞳孔曲线。刺激的效价对瞳孔大小的总体影响强于唤醒。这些结果突出了瞳孔测量在研究听觉情绪加工方面的潜力,并为假性言语与其他情感听觉刺激的语境化提供了基础。
{"title":"The Time Course of the Pupillary Response to Auditory Emotions in Pseudospeech, Music, and Vocalizations.","authors":"Julie Kirwan, Deniz Başkent, Anita Wagner","doi":"10.1177/23312165251365824","DOIUrl":"10.1177/23312165251365824","url":null,"abstract":"<p><p>Emotions can be communicated through visual and dynamic characteristics such as smiles and gestures, but also through auditory channels such as laughter, music, and human speech. Pupil dilation has become a notable marker for visual emotion processing; however the pupil's sensitivity to emotional sounds, specifically speech, remains largely underexplored. This study investigated the processing of emotional pseudospeech, which are speech-like sentences devoid of semantic content. We measured participants' pupil dilations while they listened to pseudospeech, music, and human vocalizations, and subsequently performed an emotion recognition task. Our results showed that emotional pseudospeech can trigger increases of pupil dilation compared to neutral pseudospeech, supporting the use of pupillometry as a tool for indexing prosodic emotion processing in the absence of semantics. However, pupil responses to pseudospeech were smaller and slower than the responses evoked by human vocalizations. The pupillary response was not sensitive enough to distinguish between emotion categories in pseudospeech, but pupil dilations to music and vocalizations reflected some emotion-specific pupillary curves. The valence of the stimulus had a stronger overall influence on pupil size than arousal. These results highlight the potential for pupillometry in studying auditory emotion processing and provide a foundation for contextualizing pseudospeech alongside other affective auditory stimuli.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251365824"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12340197/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144817997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Validation of a Self-Fitting Over-the-Counter Hearing Aid Intervention Compared with a Clinician-Fitted Hearing Aid Intervention: A Within-Subjects Crossover Design Using the Same Device. 自配非处方助听器干预与临床配装助听器干预的验证:使用相同设备的受试者交叉设计。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-03-25 DOI: 10.1177/23312165251328055
Lucas S Baltzell, Kosta Kokkinakis, Amy Li, Anusha Yellamsetty, Katherine Teece, Peggy B Nelson

In October of 2022, the US Food and Drug Administration finalized regulations establishing the category of self-fitting over-the-counter (OTC) hearing aids, intended to reduce barriers to hearing aid adoption for individuals with self-perceived mild to moderate hearing loss. Since then a number of self-fitting OTC hearing aids have entered the market, and a small number of published studies have demonstrated the effectiveness of a self-fitted OTC intervention against a traditional clinician-fitted intervention. Given the variety of self-fitting approaches available, and the small number of studies demonstrating effectiveness, the goal of the present study was to evaluate the effectiveness of a commercially available self-fitting OTC hearing aid intervention against a clinician-fitted intervention. Consistent with previous studies, we found that the self-fitted intervention was not inferior to the clinician-fitted intervention for self-reported benefit and objective speech-in-noise outcomes. We found statistically significant improvements in self-fitted outcomes compared to clinician-fitted outcomes, though deviations from best audiological practices in our clinician-fitted intervention may have influenced our results. In addition to presenting our results, we discuss the state of evaluating the noninferiority of self-fitted interventions and offer some new perspectives.

2022年10月,美国食品和药物管理局(fda)最终确定了建立自配非处方药(OTC)助听器类别的法规,旨在减少自认为有轻度至中度听力损失的个人采用助听器的障碍。从那时起,一些自适应的OTC助听器进入了市场,少数已发表的研究表明,自适应的OTC干预措施比传统的临床医生干预措施更有效。考虑到各种各样的自配方法,以及证明有效性的少量研究,本研究的目的是评估市售自配OTC助听器干预与临床医生安装干预的有效性。与先前的研究一致,我们发现自我拟合干预在自我报告的获益和客观噪音语音结果方面并不亚于临床拟合干预。我们发现自拟结果与临床拟合结果相比有统计学上的显著改善,尽管在临床拟合干预中偏离最佳听力学实践可能会影响我们的结果。除了介绍我们的研究结果,我们还讨论了评估自拟合干预措施的非劣效性的现状,并提出了一些新的观点。
{"title":"Validation of a Self-Fitting Over-the-Counter Hearing Aid Intervention Compared with a Clinician-Fitted Hearing Aid Intervention: A Within-Subjects Crossover Design Using the Same Device.","authors":"Lucas S Baltzell, Kosta Kokkinakis, Amy Li, Anusha Yellamsetty, Katherine Teece, Peggy B Nelson","doi":"10.1177/23312165251328055","DOIUrl":"10.1177/23312165251328055","url":null,"abstract":"<p><p>In October of 2022, the US Food and Drug Administration finalized regulations establishing the category of self-fitting over-the-counter (OTC) hearing aids, intended to reduce barriers to hearing aid adoption for individuals with self-perceived mild to moderate hearing loss. Since then a number of self-fitting OTC hearing aids have entered the market, and a small number of published studies have demonstrated the effectiveness of a self-fitted OTC intervention against a traditional clinician-fitted intervention. Given the variety of self-fitting approaches available, and the small number of studies demonstrating effectiveness, the goal of the present study was to evaluate the effectiveness of a commercially available self-fitting OTC hearing aid intervention against a clinician-fitted intervention. Consistent with previous studies, we found that the self-fitted intervention was not inferior to the clinician-fitted intervention for self-reported benefit and objective speech-in-noise outcomes. We found statistically significant improvements in self-fitted outcomes compared to clinician-fitted outcomes, though deviations from best audiological practices in our clinician-fitted intervention may have influenced our results. In addition to presenting our results, we discuss the state of evaluating the noninferiority of self-fitted interventions and offer some new perspectives.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251328055"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11938855/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143701449","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Masked-speech Recognition Using Human and Synthetic Cloned Speech. 基于人类和合成克隆语音的屏蔽语音识别。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-12-08 DOI: 10.1177/23312165251403080
Lauren Calandruccio, Mohsen Hariri, Emily Buss, Vipin Chaudhary

Voice cloning is used to generate synthetic speech that mimics vocal characteristics of human talkers. This experiment used voice cloning to compare human and synthetic speech for intelligibility, human-likeness, and perceptual similarity, all tested in young adults with normal hearing. Masked-sentence recognition was evaluated using speech produced by five human talkers and their synthetically generated voice clones presented in speech-shaped noise at -6 dB signal-to-noise ratio. There were two types of sentences: semantically meaningful and nonsense. Human and automatic speech recognition scoring was used to evaluate performance. Participants were asked to rate human-likeness and determine whether pairs of sentences were produced by the same versus different people. As expected, sentence-recognition scores were worse for nonsense sentences compared to meaningful sentences, but they were similar for speech produced by human talkers and voice clones. Human-likeness scores were also similar for speech produced by human talkers and their voice clones. Participants were very good at identifying differences between voices but were less accurate at distinguishing between human/clone pairs, often leaning towards thinking they were produced by the same person. Reliability scoring by automatic speech recognition agreed with human reliability scoring for 98% of keywords and was minimally dependent on the context of the target sentences. Results provide preliminary support for the use of voice clones when evaluating the recognition of human and synthetic speech. More generally, voice synthesis and automatic speech recognition are promising tools for evaluating speech recognition in human listeners.

语音克隆用于生成模仿人类说话者声音特征的合成语音。该实验使用语音克隆来比较人类和合成语音的可理解性、人类相似性和感知相似性,所有这些都在听力正常的年轻成年人中进行了测试。利用5位人类说话者的语音及其合成的语音克隆,在信噪比为- 6db的情况下,以语音形状的噪声呈现,对屏蔽句识别进行了评估。有两种句子:语义上有意义的和无意义的。使用人工和自动语音识别评分来评估性能。参与者被要求评价与人类的相似度,并确定成对的句子是由同一个人还是不同的人说的。正如预期的那样,与有意义的句子相比,无意义句子的句子识别得分更低,但对于人类说话者和语音克隆产生的语音来说,它们是相似的。人类说话者和他们的声音克隆体的语音相似度得分也相似。参与者非常擅长识别声音之间的差异,但在区分人类/克隆对方面不太准确,通常倾向于认为它们是同一个人发出的。在98%的关键词上,自动语音识别的可靠性评分与人类的可靠性评分一致,并且对目标句子的上下文依赖最小。结果为语音克隆技术在评估人类语音和合成语音识别中的应用提供了初步支持。更一般地说,语音合成和自动语音识别是评估人类听众语音识别的有前途的工具。
{"title":"Masked-speech Recognition Using Human and Synthetic Cloned Speech.","authors":"Lauren Calandruccio, Mohsen Hariri, Emily Buss, Vipin Chaudhary","doi":"10.1177/23312165251403080","DOIUrl":"10.1177/23312165251403080","url":null,"abstract":"<p><p>Voice cloning is used to generate synthetic speech that mimics vocal characteristics of human talkers. This experiment used voice cloning to compare human and synthetic speech for intelligibility, human-likeness, and perceptual similarity, all tested in young adults with normal hearing. Masked-sentence recognition was evaluated using speech produced by five human talkers and their synthetically generated voice clones presented in speech-shaped noise at -6 dB signal-to-noise ratio. There were two types of sentences: semantically meaningful and nonsense. Human and automatic speech recognition scoring was used to evaluate performance. Participants were asked to rate human-likeness and determine whether pairs of sentences were produced by the same versus different people. As expected, sentence-recognition scores were worse for nonsense sentences compared to meaningful sentences, but they were similar for speech produced by human talkers and voice clones. Human-likeness scores were also similar for speech produced by human talkers and their voice clones. Participants were very good at identifying differences between voices but were less accurate at distinguishing between human/clone pairs, often leaning towards thinking they were produced by the same person. Reliability scoring by automatic speech recognition agreed with human reliability scoring for 98% of keywords and was minimally dependent on the context of the target sentences. Results provide preliminary support for the use of voice clones when evaluating the recognition of human and synthetic speech. More generally, voice synthesis and automatic speech recognition are promising tools for evaluating speech recognition in human listeners.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251403080"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12686364/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145702670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptation to Noise in Spectrotemporal Modulation Detection and Word Recognition 谱时调制检测和单词识别中的噪声适应性
IF 2.7 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2024-09-14 DOI: 10.1177/23312165241266322
David López-Ramos, Miriam I. Marrufo-Pérez, Almudena Eustaquio-Martín, Luis E. López-Bascuas, Enrique A. Lopez-Poveda
Noise adaptation is the improvement in auditory function as the signal of interest is delayed in the noise. Here, we investigated if noise adaptation occurs in spectral, temporal, and spectrotemporal modulation detection as well as in speech recognition. Eighteen normal-hearing adults participated in the experiments. In the modulation detection tasks, the signal was a 200ms spectrally and/or temporally modulated ripple noise. The spectral modulation rate was two cycles per octave, the temporal modulation rate was 10 Hz, and the spectrotemporal modulations combined these two modulations, which resulted in a downward-moving ripple. A control experiment was performed to determine if the results generalized to upward-moving ripples. In the speech recognition task, the signal consisted of disyllabic words unprocessed or vocoded to maintain only envelope cues. Modulation detection thresholds at 0 dB signal-to-noise ratio and speech reception thresholds were measured in quiet and in white noise (at 60 dB SPL) for noise-signal onset delays of 50 ms (early condition) and 800 ms (late condition). Adaptation was calculated as the threshold difference between the early and late conditions. Adaptation in word recognition was statistically significant for vocoded words (2.1 dB) but not for natural words (0.6 dB). Adaptation was found to be statistically significant in spectral (2.1 dB) and temporal (2.2 dB) modulation detection but not in spectrotemporal modulation detection (downward ripple: 0.0 dB, upward ripple: −0.4 dB). Findings suggest that noise adaptation in speech recognition is unrelated to improvements in the encoding of spectrotemporal modulation cues.
噪声适应是指当感兴趣的信号在噪声中延迟时,听觉功能得到改善。在此,我们研究了噪声适应是否发生在频谱、时间和频谱时空调制检测以及语音识别中。18 名听力正常的成年人参加了实验。在调制检测任务中,信号为 200ms 的频谱和/或时间调制波纹噪声。频谱调制率为每倍频程两个周期,时间调制率为 10 赫兹,频谱-时间调制将这两种调制结合在一起,形成一个向下移动的波纹。我们还进行了对照实验,以确定实验结果是否适用于向上移动的波纹。在语音识别任务中,信号由未经处理或仅保留包络线索的声码字组成。在安静和白噪声(60 dB SPL)条件下,噪声-信号开始延迟为 50 ms(早期条件)和 800 ms(晚期条件)时,测量信噪比为 0 dB 时的调制检测阈值和语音接收阈值。适应度以早期和晚期条件下的阈值差计算。词汇识别的适应性对词汇编码(2.1 dB)有显著的统计学意义,而对自然词汇(0.6 dB)则没有。在频谱(2.1 dB)和时间(2.2 dB)调制检测中,发现适应具有显著的统计意义,但在频谱-时间调制检测中却没有发现适应(向下波纹:0.0 dB,向上波纹:-0.4 dB)。研究结果表明,语音识别中的噪声适应与谱时调制线索编码的改进无关。
{"title":"Adaptation to Noise in Spectrotemporal Modulation Detection and Word Recognition","authors":"David López-Ramos, Miriam I. Marrufo-Pérez, Almudena Eustaquio-Martín, Luis E. López-Bascuas, Enrique A. Lopez-Poveda","doi":"10.1177/23312165241266322","DOIUrl":"https://doi.org/10.1177/23312165241266322","url":null,"abstract":"Noise adaptation is the improvement in auditory function as the signal of interest is delayed in the noise. Here, we investigated if noise adaptation occurs in spectral, temporal, and spectrotemporal modulation detection as well as in speech recognition. Eighteen normal-hearing adults participated in the experiments. In the modulation detection tasks, the signal was a 200ms spectrally and/or temporally modulated ripple noise. The spectral modulation rate was two cycles per octave, the temporal modulation rate was 10 Hz, and the spectrotemporal modulations combined these two modulations, which resulted in a downward-moving ripple. A control experiment was performed to determine if the results generalized to upward-moving ripples. In the speech recognition task, the signal consisted of disyllabic words unprocessed or vocoded to maintain only envelope cues. Modulation detection thresholds at 0 dB signal-to-noise ratio and speech reception thresholds were measured in quiet and in white noise (at 60 dB SPL) for noise-signal onset delays of 50 ms (early condition) and 800 ms (late condition). Adaptation was calculated as the threshold difference between the early and late conditions. Adaptation in word recognition was statistically significant for vocoded words (2.1 dB) but not for natural words (0.6 dB). Adaptation was found to be statistically significant in spectral (2.1 dB) and temporal (2.2 dB) modulation detection but not in spectrotemporal modulation detection (downward ripple: 0.0 dB, upward ripple: −0.4 dB). Findings suggest that noise adaptation in speech recognition is unrelated to improvements in the encoding of spectrotemporal modulation cues.","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"44 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142256779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Trends in Hearing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1