首页 > 最新文献

Trends in Hearing最新文献

英文 中文
Is Noise Exposure Associated With Impaired Extended High Frequency Hearing Despite a Normal Audiogram? A Systematic Review and Meta-Analysis. 尽管听力图正常,噪声暴露是否与延伸高频听力受损有关?系统回顾和荟萃分析。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-05-16 DOI: 10.1177/23312165251343757
Sajana Aryal, Monica Trevino, Hansapani Rodrigo, Srikanta Mishra

Understanding the initial signature of noise-induced auditory damage remains a significant priority. Animal models suggest the cochlear base is particularly vulnerable to noise, raising the possibility that early-stage noise exposure could be linked to basal cochlear dysfunction, even when thresholds at 0.25-8 kHz are normal. To investigate this in humans, we conducted a meta-analysis following a systematic review, examining the association between noise exposure and hearing in frequencies from 9 to 20 kHz as a marker for basal cochlear dysfunction. Systematic review and meta-analysis followed PRISMA guidelines and the PICOS framework. Studies on noise exposure and hearing in the 9 to 20 kHz region in adults with clinically normal audiograms were included by searching five electronic databases (e.g., PubMed). Cohorts from 30 studies, comprising approximately 2,500 participants, were systematically reviewed. Meta-analysis was conducted on 23 studies using a random-effects model for occupational and recreational noise exposure. Analysis showed a significant positive association between occupational noise and hearing thresholds, with medium effect sizes at 9 and 11.2 kHz and large effect sizes at 10, 12, 14, and 16 kHz. However, the association with recreational noise was less consistent, with significant effects only at 12, 12.5, and 16 kHz. Egger's test indicated some publication bias, specifically at 10 kHz. Findings suggest thresholds above 8 kHz may indicate early noise exposure effects, even when lower-frequency (≤8 kHz) thresholds remain normal. Longitudinal studies incorporating noise dosimetry are crucial to establish causality and further support the clinical utility of extended high-frequency testing.

了解噪音引起的听觉损伤的初始特征仍然是一个重要的优先事项。动物模型表明,耳蜗基底特别容易受到噪音的影响,这增加了早期接触噪音可能与基底耳蜗功能障碍有关的可能性,即使阈值在0.25-8 kHz是正常的。为了在人类中研究这一点,我们在系统回顾之后进行了一项荟萃分析,研究了噪音暴露与听力之间的关系,频率从9到20 kHz,作为基础耳蜗功能障碍的标志。系统评价和荟萃分析遵循PRISMA指南和PICOS框架。通过检索5个电子数据库(例如PubMed),纳入了临床听力学正常的成人在9至20 kHz区域的噪声暴露和听力的研究。系统地回顾了来自30项研究的队列,包括大约2,500名参与者。采用随机效应模型对23项研究进行了meta分析。分析显示,职业噪声与听力阈值之间存在显著的正相关,在9和11.2 kHz时具有中等效应,在10、12、14和16 kHz时具有较大效应。然而,与娱乐性噪音的关联不太一致,只有在12、12.5和16 kHz时才有显著影响。埃格的测试显示出一些发表偏倚,特别是在10khz时。研究结果表明,高于8 kHz的阈值可能表明早期噪声暴露效应,即使低频(≤8 kHz)阈值保持正常。结合噪声剂量学的纵向研究对于建立因果关系和进一步支持扩展高频测试的临床应用至关重要。
{"title":"Is Noise Exposure Associated With Impaired Extended High Frequency Hearing Despite a Normal Audiogram? A Systematic Review and Meta-Analysis.","authors":"Sajana Aryal, Monica Trevino, Hansapani Rodrigo, Srikanta Mishra","doi":"10.1177/23312165251343757","DOIUrl":"10.1177/23312165251343757","url":null,"abstract":"<p><p>Understanding the initial signature of noise-induced auditory damage remains a significant priority. Animal models suggest the cochlear base is particularly vulnerable to noise, raising the possibility that early-stage noise exposure could be linked to basal cochlear dysfunction, even when thresholds at 0.25-8 kHz are normal. To investigate this in humans, we conducted a meta-analysis following a systematic review, examining the association between noise exposure and hearing in frequencies from 9 to 20 kHz as a marker for basal cochlear dysfunction. Systematic review and meta-analysis followed PRISMA guidelines and the PICOS framework. Studies on noise exposure and hearing in the 9 to 20 kHz region in adults with clinically normal audiograms were included by searching five electronic databases (e.g., PubMed). Cohorts from 30 studies, comprising approximately 2,500 participants, were systematically reviewed. Meta-analysis was conducted on 23 studies using a random-effects model for occupational and recreational noise exposure. Analysis showed a significant positive association between occupational noise and hearing thresholds, with medium effect sizes at 9 and 11.2 kHz and large effect sizes at 10, 12, 14, and 16 kHz. However, the association with recreational noise was less consistent, with significant effects only at 12, 12.5, and 16 kHz. Egger's test indicated some publication bias, specifically at 10 kHz. Findings suggest thresholds above 8 kHz may indicate early noise exposure effects, even when lower-frequency (≤8 kHz) thresholds remain normal. Longitudinal studies incorporating noise dosimetry are crucial to establish causality and further support the clinical utility of extended high-frequency testing.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251343757"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12084714/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144081423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pupillary Responses During a Dual Task: Effect of Noise Attenuation on the Timing of Cognitive Resource Allocation. 双重任务中的瞳孔反应:噪声衰减对认知资源分配时间的影响。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-08-14 DOI: 10.1177/23312165251367630
Federica Bianchi, Sindri Jonsson, Torben Christiansen, Elaine Hoi Ning Ng

Although multitasking is a common everyday activity, it is often challenging. The aim of this study was to evaluate the effect of noise attenuation during an audio-visual dual task and investigate cognitive resource allocation over time via pupillometry. Twenty-six normal hearing participants performed a dual task consisting of a primary speech recognition task and a secondary visual reaction-time task, as well as a visual-only task. Four conditions were tested in the dual task: two speech levels (60- and 64-dB SPL) and two noise conditions (No Attenuation with noise at 70 dB SPL; Attenuation condition with noise attenuated by passive damping). Elevated pupillary responses for the No Attenuation condition relative to the Attenuation and visual-only conditions indicated that participants allocated additional resources on the primary task during the playback of the first part of the sentence, while reaction time to the secondary task increased significantly relative to the visual-only task. In the Attenuation condition, participants performed the secondary task with a similar reaction time relative to the visual-only task (no dual-task cost), while pupillary responses revealed allocation of resources on the primary task after completion of the secondary task. These findings reveal that the temporal dynamics of cognitive resource allocation between primary and secondary task were affected by the level of background noise in the primary task. This study demonstrates that noise attenuation, as offered for example by audio devices, frees up cognitive resources in noisy listening environments and may be beneficial to improve performance and decrease dual-task costs during multitasking.

虽然多任务处理是一种常见的日常活动,但它往往具有挑战性。本研究的目的是评估噪声衰减在视听双重任务中的效果,并通过瞳孔测量法研究认知资源随时间的分配。26名听力正常的参与者执行了一项双重任务,包括主要的语音识别任务和次要的视觉反应时间任务,以及一个仅限视觉的任务。在双重任务中测试了四种条件:两种语音水平(60和64 dB SPL)和两种噪声条件(70 dB SPL噪声无衰减;噪声经被动阻尼衰减后的衰减状态)。受试者在无衰减条件下的瞳孔反应明显高于衰减条件和仅视觉条件下的瞳孔反应,这表明受试者在回放句子第一部分时在主要任务上分配了额外的资源,而对次要任务的反应时间则明显高于仅视觉条件下的反应时间。在衰减条件下,受试者在完成次要任务时的反应时间与仅视觉任务相似(没有双任务成本),而瞳孔反应显示完成次要任务后资源在主要任务上的分配。研究结果表明,主次任务间认知资源分配的时间动态受到主次任务背景噪声水平的影响。这项研究表明,噪音衰减,如音频设备提供的,在嘈杂的听力环境中释放认知资源,可能有利于提高性能,减少多任务处理时的双重任务成本。
{"title":"Pupillary Responses During a Dual Task: Effect of Noise Attenuation on the Timing of Cognitive Resource Allocation.","authors":"Federica Bianchi, Sindri Jonsson, Torben Christiansen, Elaine Hoi Ning Ng","doi":"10.1177/23312165251367630","DOIUrl":"10.1177/23312165251367630","url":null,"abstract":"<p><p>Although multitasking is a common everyday activity, it is often challenging. The aim of this study was to evaluate the effect of noise attenuation during an audio-visual dual task and investigate cognitive resource allocation over time via pupillometry. Twenty-six normal hearing participants performed a dual task consisting of a primary speech recognition task and a secondary visual reaction-time task, as well as a visual-only task. Four conditions were tested in the dual task: two speech levels (60- and 64-dB SPL) and two noise conditions (<i>No Attenuation</i> with noise at 70 dB SPL<i>; Attenuation</i> condition with noise attenuated by passive damping). Elevated pupillary responses for the N<i>o Attenuation</i> condition relative to the A<i>ttenuation</i> and visual-only conditions indicated that participants allocated additional resources on the primary task during the playback of the first part of the sentence, while reaction time to the secondary task increased significantly relative to the visual-only task. In the A<i>ttenuation</i> condition, participants performed the secondary task with a similar reaction time relative to the visual-only task (no dual-task cost), while pupillary responses revealed allocation of resources on the primary task after completion of the secondary task. These findings reveal that the temporal dynamics of cognitive resource allocation between primary and secondary task were affected by the level of background noise in the primary task. This study demonstrates that noise attenuation, as offered for example by audio devices, frees up cognitive resources in noisy listening environments and may be beneficial to improve performance and decrease dual-task costs during multitasking.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251367630"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12357024/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144849442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comprehensive Measurements and Analyses of Ear Canal Geometry From Late Infancy Through Late Adulthood: Age-Related Variations and Implications for Basic Science and Audiological Measurements. 从婴儿晚期到成年晚期耳道几何形状的综合测量和分析:与年龄相关的变化和基础科学和听力学测量的含义。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-06-25 DOI: 10.1177/23312165251345572
Susan E Voss, Aaron K Remenschneider, Rebecca M Farrar, Soomin Myoung, Nicholas J Horton

This study provides a comprehensive analysis of ear canal geometry from 0.7 to 91 years, based on high-resolution computed tomography scans of 221 ears. Quantified features include cross-sectional areas along the canal's length, total canal length, curvature, and key anatomical landmarks such as the first and second bends and the cartilage-to-bone transition. Significant developmental changes occur during the first 10 years of life, with adult-like characteristics emerging between ages 10 and 15 years, likely coinciding with puberty. Substantial interindividual variability is observed across all ages, particularly in the canal area. The canal becomes fully cartilaginous at and lateral to the second bend by 0.7 years, with further growth occurring only in the bony segment thereafter. These anatomical findings have important implications for audiologic threshold assessments, wideband acoustic immitance measures, age-appropriate hearing aid fitting schedules, and surgical planning, particularly in pediatric populations where anatomical variation is greatest.

本研究基于221只耳朵的高分辨率计算机断层扫描,对0.7至91岁的耳道几何形状进行了全面分析。量化特征包括沿根管长度的横截面积、根管总长度、曲率和关键解剖标志,如第一和第二弯以及软骨到骨的过渡。重大的发育变化发生在生命的前10年,在10到15岁之间出现成年人的特征,可能与青春期同时发生。在所有年龄段都观察到大量的个体间差异,特别是在运河区域。到0.7岁时,在第二次弯曲处和外侧的椎管完全成为软骨,此后仅在骨段进一步生长。这些解剖学发现对听力学阈值评估、宽带声阻抗测量、适龄助听器安装计划和手术计划具有重要意义,特别是在解剖差异最大的儿科人群中。
{"title":"Comprehensive Measurements and Analyses of Ear Canal Geometry From Late Infancy Through Late Adulthood: Age-Related Variations and Implications for Basic Science and Audiological Measurements.","authors":"Susan E Voss, Aaron K Remenschneider, Rebecca M Farrar, Soomin Myoung, Nicholas J Horton","doi":"10.1177/23312165251345572","DOIUrl":"10.1177/23312165251345572","url":null,"abstract":"<p><p>This study provides a comprehensive analysis of ear canal geometry from 0.7 to 91 years, based on high-resolution computed tomography scans of 221 ears. Quantified features include cross-sectional areas along the canal's length, total canal length, curvature, and key anatomical landmarks such as the first and second bends and the cartilage-to-bone transition. Significant developmental changes occur during the first 10 years of life, with adult-like characteristics emerging between ages 10 and 15 years, likely coinciding with puberty. Substantial interindividual variability is observed across all ages, particularly in the canal area. The canal becomes fully cartilaginous at and lateral to the second bend by 0.7 years, with further growth occurring only in the bony segment thereafter. These anatomical findings have important implications for audiologic threshold assessments, wideband acoustic immitance measures, age-appropriate hearing aid fitting schedules, and surgical planning, particularly in pediatric populations where anatomical variation is greatest.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251345572"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12198549/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144486732","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Language-agnostic, Automated Assessment of Listeners' Speech Recall Using Large Language Models. 语言不可知论,使用大型语言模型对听者语音回忆的自动评估。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-05-30 DOI: 10.1177/23312165251347131
Björn Herrmann

Speech-comprehension difficulties are common among older people. Standard speech tests do not fully capture such difficulties because the tests poorly resemble the context-rich, story-like nature of ongoing conversation and are typically available only in a country's dominant/official language (e.g., English), leading to inaccurate scores for native speakers of other languages. Assessments for naturalistic, story speech in multiple languages require accurate, time-efficient scoring. The current research leverages modern large language models (LLMs) in native English speakers and native speakers of 10 other languages to automate the generation of high-quality, spoken stories and scoring of speech recall in different languages. Participants listened to and freely recalled short stories (in quiet/clear and in babble noise) in their native language. Large language model text-embeddings and LLM prompt engineering with semantic similarity analyses to score speech recall revealed sensitivity to known effects of temporal order, primacy/recency, and background noise, and high similarity of recall scores across languages. The work overcomes limitations associated with simple speech materials and testing of closed native-speaker groups because recall data of varying length and details can be mapped across languages with high accuracy. The full automation of speech generation and recall scoring provides an important step toward comprehension assessments of naturalistic speech with clinical applicability.

语言理解困难在老年人中很常见。标准的语言测试不能完全捕捉到这些困难,因为这些测试与正在进行的对话的上下文丰富、故事般的性质很不相似,而且通常只适用于一个国家的主导语言/官方语言(例如英语),导致以其他语言为母语的人得分不准确。对多种语言的自然主义、故事演讲的评估需要准确、省时的评分。目前的研究利用现代大型语言模型(llm),以英语为母语和其他10种语言为母语的人,自动生成高质量的口语故事,并对不同语言的语音回忆进行评分。参与者用母语听并自由回忆短篇故事(安静/清晰,咿呀学语)。大型语言模型文本嵌入和LLM提示工程利用语义相似度分析对语音回忆进行评分,揭示了对时间顺序、首因/近因和背景噪声的已知影响的敏感性,以及不同语言之间回忆分数的高相似性。这项工作克服了简单的语音材料和封闭的母语人群测试的局限性,因为不同长度和细节的回忆数据可以高精度地映射到不同的语言中。语音生成和回忆评分的完全自动化为具有临床适用性的自然语音理解评估提供了重要的一步。
{"title":"Language-agnostic, Automated Assessment of Listeners' Speech Recall Using Large Language Models.","authors":"Björn Herrmann","doi":"10.1177/23312165251347131","DOIUrl":"10.1177/23312165251347131","url":null,"abstract":"<p><p>Speech-comprehension difficulties are common among older people. Standard speech tests do not fully capture such difficulties because the tests poorly resemble the context-rich, story-like nature of ongoing conversation and are typically available only in a country's dominant/official language (e.g., English), leading to inaccurate scores for native speakers of other languages. Assessments for naturalistic, story speech in multiple languages require accurate, time-efficient scoring. The current research leverages modern large language models (LLMs) in native English speakers and native speakers of 10 other languages to automate the generation of high-quality, spoken stories and scoring of speech recall in different languages. Participants listened to and freely recalled short stories (in quiet/clear and in babble noise) in their native language. Large language model text-embeddings and LLM prompt engineering with semantic similarity analyses to score speech recall revealed sensitivity to known effects of temporal order, primacy/recency, and background noise, and high similarity of recall scores across languages. The work overcomes limitations associated with simple speech materials and testing of closed native-speaker groups because recall data of varying length and details can be mapped across languages with high accuracy. The full automation of speech generation and recall scoring provides an important step toward comprehension assessments of naturalistic speech with clinical applicability.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251347131"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12125525/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144192395","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Association of Tinnitus With Speech Recognition and Executive Functions in Older Adults. 耳鸣与老年人语音识别和执行功能的关系。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-11-13 DOI: 10.1177/23312165251389585
Nick Sommerhalder, Zbyněk Bureš, Oliver Profant, Tobias Kleinjung, Patrick Neff, Martin Meyer

Adults with chronic subjective tinnitus often struggle with speech recognition in challenging listening environments. While most research demonstrates deficits in speech recognition among individuals with tinnitus, studies focusing on older adults remain scarce. Besides speech recognition deficits, tinnitus has been linked to diminished cognitive performance, particularly in executive functions, yet its associations with specific cognitive domains in ageing populations are not fully understood. Our previous study of younger adults found that individuals with tinnitus exhibit deficits in speech recognition and interference control. Building on this, we hypothesized that these deficits are also present for older adults. We conducted a cross-sectional study of older adults (aged 60-79), 32 with tinnitus and 31 controls matched for age, gender, education, and approximately matched for hearing loss. Participants underwent audiometric, speech recognition, and cognitive tasks. The tinnitus participants performed more poorly in speech-in-noise and gated speech tasks, whereas no group differences were observed in the other suprathreshold auditory tasks. With regard to cognition, individuals with tinnitus showed reduced interference control, emotional interference, cognitive flexibility, and verbal working memory, correlating with tinnitus distress and loudness. It is concluded that tinnitus-related deficits persist and even worsen with age. Our results suggest that altered central mechanisms contribute to speech recognition difficulties in older adults with tinnitus.

患有慢性主观性耳鸣的成年人经常在具有挑战性的听力环境中挣扎于语音识别。虽然大多数研究表明耳鸣患者在语音识别方面存在缺陷,但针对老年人的研究仍然很少。除了语音识别缺陷外,耳鸣还与认知能力下降,特别是执行功能下降有关,但其与老年人特定认知领域的关系尚不完全清楚。我们之前对年轻人的研究发现,耳鸣患者在语音识别和干扰控制方面表现出缺陷。在此基础上,我们假设老年人也存在这些缺陷。我们对老年人(60-79岁)进行了一项横断面研究,其中32名耳鸣患者和31名年龄、性别、教育程度和听力损失大致匹配的对照组。参与者接受了听力测量、语音识别和认知任务。耳鸣参与者在噪音语音和门控语音任务中表现更差,而在其他超阈值听觉任务中没有观察到组间差异。在认知方面,耳鸣个体表现出干扰控制、情绪干扰、认知灵活性和言语工作记忆的减少,与耳鸣痛苦和响度相关。结论是,耳鸣相关的缺陷持续存在,甚至随着年龄的增长而恶化。我们的研究结果表明,中枢机制的改变导致了老年耳鸣患者的语音识别困难。
{"title":"Association of Tinnitus With Speech Recognition and Executive Functions in Older Adults.","authors":"Nick Sommerhalder, Zbyněk Bureš, Oliver Profant, Tobias Kleinjung, Patrick Neff, Martin Meyer","doi":"10.1177/23312165251389585","DOIUrl":"10.1177/23312165251389585","url":null,"abstract":"<p><p>Adults with chronic subjective tinnitus often struggle with speech recognition in challenging listening environments. While most research demonstrates deficits in speech recognition among individuals with tinnitus, studies focusing on older adults remain scarce. Besides speech recognition deficits, tinnitus has been linked to diminished cognitive performance, particularly in executive functions, yet its associations with specific cognitive domains in ageing populations are not fully understood. Our previous study of younger adults found that individuals with tinnitus exhibit deficits in speech recognition and interference control. Building on this, we hypothesized that these deficits are also present for older adults. We conducted a cross-sectional study of older adults (aged 60-79), 32 with tinnitus and 31 controls matched for age, gender, education, and approximately matched for hearing loss. Participants underwent audiometric, speech recognition, and cognitive tasks. The tinnitus participants performed more poorly in speech-in-noise and gated speech tasks, whereas no group differences were observed in the other suprathreshold auditory tasks. With regard to cognition, individuals with tinnitus showed reduced interference control, emotional interference, cognitive flexibility, and verbal working memory, correlating with tinnitus distress and loudness. It is concluded that tinnitus-related deficits persist and even worsen with age. Our results suggest that altered central mechanisms contribute to speech recognition difficulties in older adults with tinnitus.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251389585"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12615926/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145514780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
At-Home Auditory Assessment Using Portable Automated Rapid Testing (PART) to Understand Self-Reported Hearing Difficulties. 使用便携式自动快速测试的家庭听力评估(部分)来了解自我报告的听力困难。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-11-24 DOI: 10.1177/23312165251397373
E Sebastian Lelo de Larrea-Mancera, Tess K Koerner, William J Bologna, Sara Momtaz, Katherine N Menon, Audrey Carrillo, Eric C Hoover, G Christopher Stecker, Frederick J Gallun, Aaron R Seitz

Previous research has demonstrated that remote testing of suprathreshold auditory function using distributed technologies can produce results that closely match those obtained in laboratory settings with specialized, calibrated equipment. This work has facilitated the validation of various behavioral measures in remote settings that provide valuable insights into auditory function. In the current study, we sought to address whether a broad battery of auditory assessments could explain variance in self-report of hearing handicap. To address this, we used a portable psychophysics assessment tool along with an online recruitment tool (Prolific) to collect auditory task data from participants with (n= 84) and without (n= 108) self-reported hearing difficulty. Results indicate several measures of auditory processing differentiate participants with and without self-reported hearing difficulty. In addition, we report the factor structure of the test battery to clarify the underlying constructs and the extent to which they individually or jointly inform hearing function. Relationships between measures of auditory processing were found to be largely consistent with a hypothesized construct model that guided task selection. Overall, this study advances our understanding of the relationship between auditory and cognitive processing in those with and without subjective hearing difficulty. More broadly, these results indicate promise that these measures can be used in larger scale research studies in remote settings and have potential to contribute to telehealth approaches to better address people's hearing needs.

先前的研究表明,使用分布式技术对超阈听觉功能进行远程测试,可以产生与实验室设置中使用专业校准设备获得的结果密切匹配的结果。这项工作促进了远程环境中各种行为测量的验证,为听觉功能提供了有价值的见解。在当前的研究中,我们试图解决广泛的听觉评估是否可以解释听力障碍自我报告的差异。为了解决这个问题,我们使用了便携式心理物理评估工具和在线招募工具(多产),从有(n = 84)和没有(n = 108)自我报告听力困难的参与者中收集听觉任务数据。结果表明,听觉处理的几个措施区分参与者有和没有自我报告的听力困难。此外,我们报告了测试电池的因素结构,以澄清潜在的结构以及它们单独或共同通知听力功能的程度。听觉处理测量之间的关系被发现在很大程度上与指导任务选择的假设结构模型一致。总的来说,本研究促进了我们对主观听力障碍和非主观听力障碍的听觉和认知加工之间关系的理解。更广泛地说,这些结果表明,这些措施有望用于远程环境中的更大规模研究,并有可能促进远程保健方法,以更好地满足人们的听力需求。
{"title":"At-Home Auditory Assessment Using Portable Automated Rapid Testing (PART) to Understand Self-Reported Hearing Difficulties.","authors":"E Sebastian Lelo de Larrea-Mancera, Tess K Koerner, William J Bologna, Sara Momtaz, Katherine N Menon, Audrey Carrillo, Eric C Hoover, G Christopher Stecker, Frederick J Gallun, Aaron R Seitz","doi":"10.1177/23312165251397373","DOIUrl":"10.1177/23312165251397373","url":null,"abstract":"<p><p>Previous research has demonstrated that remote testing of suprathreshold auditory function using distributed technologies can produce results that closely match those obtained in laboratory settings with specialized, calibrated equipment. This work has facilitated the validation of various behavioral measures in remote settings that provide valuable insights into auditory function. In the current study, we sought to address whether a broad battery of auditory assessments could explain variance in self-report of hearing handicap. To address this, we used a portable psychophysics assessment tool along with an online recruitment tool (Prolific) to collect auditory task data from participants with (<i>n</i> <i>=</i> 84) and without (<i>n</i> <i>=</i> 108) self-reported hearing difficulty. Results indicate several measures of auditory processing differentiate participants with and without self-reported hearing difficulty. In addition, we report the factor structure of the test battery to clarify the underlying constructs and the extent to which they individually or jointly inform hearing function. Relationships between measures of auditory processing were found to be largely consistent with a hypothesized construct model that guided task selection. Overall, this study advances our understanding of the relationship between auditory and cognitive processing in those with and without subjective hearing difficulty. More broadly, these results indicate promise that these measures can be used in larger scale research studies in remote settings and have potential to contribute to telehealth approaches to better address people's hearing needs.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251397373"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12644446/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145597487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Objective Evaluation of a Deep Learning-Based Noise Reduction Algorithm for Hearing Aids Under Diverse Fitting and Listening Conditions. 不同拟合和听力条件下基于深度学习的助听器降噪算法的客观评价。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-11-25 DOI: 10.1177/23312165251396644
Vahid Ashkanichenarlogh, Paula Folkeard, Susan Scollie, Volker Kühnel, Vijay Parsa

This study evaluated a deep-neural-network denoising system using model-based design, comparing it with adaptive filtering and beamforming across various noise types, SNRs, and hearing-aid fittings. A KEMAR manikin fitted with five audiograms was recorded in reverberant and non-reverberant rooms, yielding 1,152 recordings. Speech intelligibility was estimated using the HASPI from 1,152 KEMAR manikin recordings. Effects of processing strategy and acoustic factors were tested with model-based within-device design that account for repeated recordings per device/program and fitting. Linear mixed model results showed that the DNN with beamforming outperformed conventional processing, with strongest gains at 0 and +5 dB SNR, moderate benefits at -5 dB in low reverberation, and none in medium reverberation. Across SNRs and noise types, the DNN combined with beamforming yielded the highest predicted intelligibility, with benefits attenuated under moderate reverberation. Azimuth effects varied; because estimates were derived from a better-ear metric on manikin recordings. Additionally, this paper reports comparisons using metrics of sound quality, for an intrusive metric (HASQI) and the pMOS non-intrusive metric. Results indicated that model type interacted with processing and acoustic factors. HASQI and pMOS scores increased with SNR and were moderately correlated (r² ≈ 0.479), supporting the use of non-intrusive metrics for large-scale assessment. However, pMOS showed greater variability across hearing aid programs and environments, suggesting non-intrusive models capture processing effects differently than intrusive metrics. These findings highlight the promise and limits of non-intrusive evaluation while emphasizing the benefit of combining deep learning with beamforming to improve intelligibility and quality.

本研究使用基于模型的设计评估了一种深度神经网络去噪系统,并将其与各种噪声类型、信噪比和助听器配件的自适应滤波和波束形成进行了比较。在混响和非混响房间中记录了一个装有5个听音仪的KEMAR人体模型,产生了1152个录音。使用来自1152个KEMAR人体录音的HASPI来估计语音可理解性。处理策略和声学因素的影响通过基于模型的设备内设计进行测试,该设计考虑了每个设备/程序和拟合的重复录音。线性混合模型结果表明,波束形成的深度神经网络优于常规处理,在0和+5 dB信噪比下增益最大,在低混响条件下-5 dB增益中等,在中混响条件下没有增益。在信噪比和噪声类型中,深度神经网络与波束形成相结合产生了最高的预测可理解性,在中度混响下其优势减弱。方位效应各不相同;因为估计是根据一个更好的耳朵度量模型录音得出的。此外,本文还报道了使用音质指标的比较,用于侵入性指标(HASQI)和pMOS非侵入性指标。结果表明,模型类型与加工和声学因素相互作用。HASQI和pMOS评分随信噪比增加而增加,并呈中等相关性(r²≈0.479),支持使用非侵入性指标进行大规模评估。然而,pMOS在助听器项目和环境中表现出更大的可变性,这表明非侵入性模型捕获的处理效果与侵入性指标不同。这些发现突出了非侵入性评估的前景和局限性,同时强调了将深度学习与波束成形相结合以提高可理解性和质量的好处。
{"title":"Objective Evaluation of a Deep Learning-Based Noise Reduction Algorithm for Hearing Aids Under Diverse Fitting and Listening Conditions.","authors":"Vahid Ashkanichenarlogh, Paula Folkeard, Susan Scollie, Volker Kühnel, Vijay Parsa","doi":"10.1177/23312165251396644","DOIUrl":"10.1177/23312165251396644","url":null,"abstract":"<p><p>This study evaluated a deep-neural-network denoising system using model-based design, comparing it with adaptive filtering and beamforming across various noise types, SNRs, and hearing-aid fittings. A KEMAR manikin fitted with five audiograms was recorded in reverberant and non-reverberant rooms, yielding 1,152 recordings. Speech intelligibility was estimated using the HASPI from 1,152 KEMAR manikin recordings. Effects of processing strategy and acoustic factors were tested with model-based within-device design that account for repeated recordings per device/program and fitting. Linear mixed model results showed that the DNN with beamforming outperformed conventional processing, with strongest gains at 0 and +5 dB SNR, moderate benefits at -5 dB in low reverberation, and none in medium reverberation. Across SNRs and noise types, the DNN combined with beamforming yielded the highest predicted intelligibility, with benefits attenuated under moderate reverberation. Azimuth effects varied; because estimates were derived from a better-ear metric on manikin recordings. Additionally, this paper reports comparisons using metrics of sound quality, for an intrusive metric (HASQI) and the pMOS non-intrusive metric. Results indicated that model type interacted with processing and acoustic factors. HASQI and pMOS scores increased with SNR and were moderately correlated (r² ≈ 0.479), supporting the use of non-intrusive metrics for large-scale assessment. However, pMOS showed greater variability across hearing aid programs and environments, suggesting non-intrusive models capture processing effects differently than intrusive metrics. These findings highlight the promise and limits of non-intrusive evaluation while emphasizing the benefit of combining deep learning with beamforming to improve intelligibility and quality.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251396644"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12647563/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145606795","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Objectively Measuring Audiovisual Effects in Noise Using Virtual Human Speakers. 利用虚拟人说话者客观测量噪声中的视听效果。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-04-13 DOI: 10.1177/23312165251333528
John Kyle Cooper, Jonas Vanthornhout, Astrid van Wieringen, Tom Francart

Speech intelligibility in challenging listening environments relies on the integration of audiovisual cues. Measuring the effectiveness of audiovisual integration in these challenging listening environments can be difficult due to the complexity of such environments. The Audiovisual True-to-Life Assessment of Auditory Rehabilitation (AVATAR) is a paradigm that was developed to provide an ecological environment to capture both the audio and visual aspects of speech intelligibility measures. Previous research has shown the benefit from audiovisual cues can be measured using behavioral (e.g., word recognition) and electrophysiological (e.g., neural tracking) measures. The current research examines, when using the AVATAR paradigm, if electrophysiological measures of speech intelligibility yield similar outcomes as behavioral measures. We hypothesized visual cues would enhance both the behavioral and electrophysiological scores as the signal-to-noise ratio (SNR) of the speech signal decreased. Twenty young (18-25 years old) participants (1 male and 19 female) with normal hearing participated in our study. For our behavioral experiment, we administered lists of sentences using an adaptive procedure to estimate a speech reception threshold (SRT). For our electrophysiological experiment, we administered 35 lists of sentences randomized across five SNR levels (silence, 0, -3, -6, and -9 dB) and two visual conditions (audio-only and audiovisual). We used a neural tracking decoder to measure the reconstruction accuracies for each participant. We observed most participants had higher reconstruction accuracies for the audiovisual condition compared to the audio-only condition in conditions with moderate to high levels of noise. We found the electrophysiological measure may correlate with the behavioral measure that shows audiovisual benefit.

在具有挑战性的听力环境中,语音清晰度依赖于视听线索的整合。由于这些环境的复杂性,在这些具有挑战性的听力环境中测量视听整合的有效性可能很困难。听觉康复的视听逼真评估(AVATAR)是一个范例,旨在提供一个生态环境来捕捉语音可理解性措施的听觉和视觉方面。先前的研究表明,视听线索的好处可以通过行为(例如,单词识别)和电生理(例如,神经跟踪)测量来衡量。当前的研究在使用AVATAR范式时检验了语音可理解性的电生理测量是否与行为测量产生相似的结果。我们假设当语音信号的信噪比(SNR)降低时,视觉线索会提高行为和电生理评分。20名听力正常的年轻人(18-25岁)(男1名,女19名)参加了我们的研究。在我们的行为实验中,我们使用自适应程序管理句子列表来估计语音接收阈值(SRT)。在我们的电生理实验中,我们管理了35个句子列表,随机分为5种信噪比水平(沉默,0,-3,-6和-9 dB)和两种视觉条件(纯音频和视听)。我们使用神经跟踪解码器来测量每个参与者的重建精度。我们观察到,在中等到高水平的噪声条件下,大多数参与者在视听条件下的重建精度高于纯音频条件。我们发现电生理测量可能与显示视听效益的行为测量相关。
{"title":"Objectively Measuring Audiovisual Effects in Noise Using Virtual Human Speakers.","authors":"John Kyle Cooper, Jonas Vanthornhout, Astrid van Wieringen, Tom Francart","doi":"10.1177/23312165251333528","DOIUrl":"https://doi.org/10.1177/23312165251333528","url":null,"abstract":"<p><p>Speech intelligibility in challenging listening environments relies on the integration of audiovisual cues. Measuring the effectiveness of audiovisual integration in these challenging listening environments can be difficult due to the complexity of such environments. The Audiovisual True-to-Life Assessment of Auditory Rehabilitation (AVATAR) is a paradigm that was developed to provide an ecological environment to capture both the audio and visual aspects of speech intelligibility measures. Previous research has shown the benefit from audiovisual cues can be measured using behavioral (e.g., word recognition) and electrophysiological (e.g., neural tracking) measures. The current research examines, when using the AVATAR paradigm, if electrophysiological measures of speech intelligibility yield similar outcomes as behavioral measures. We hypothesized visual cues would enhance both the behavioral and electrophysiological scores as the signal-to-noise ratio (SNR) of the speech signal decreased. Twenty young (18-25 years old) participants (1 male and 19 female) with normal hearing participated in our study. For our behavioral experiment, we administered lists of sentences using an adaptive procedure to estimate a speech reception threshold (SRT). For our electrophysiological experiment, we administered 35 lists of sentences randomized across five SNR levels (silence, 0, -3, -6, and -9 dB) and two visual conditions (audio-only and audiovisual). We used a neural tracking decoder to measure the reconstruction accuracies for each participant. We observed most participants had higher reconstruction accuracies for the audiovisual condition compared to the audio-only condition in conditions with moderate to high levels of noise. We found the electrophysiological measure may correlate with the behavioral measure that shows audiovisual benefit.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251333528"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12033406/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144043708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Measuring Speech Discrimination Ability in Sleeping Infants Using fNIRS-A Proof of Principle. 用fnirs测量睡眠婴儿言语辨别能力的原理证明。
IF 2.6 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 DOI: 10.1177/23312165241311721
Onn Wah Lee, Demi Gao, Tommy Peng, Julia Wunderlich, Darren Mao, Gautam Balasubramanian, Colette M McKay

This study used functional near-infrared spectroscopy (fNIRS) to measure aspects of the speech discrimination ability of sleeping infants. We examined the morphology of the fNIRS response to three different speech contrasts, namely "Tea/Ba," "Bee/Ba," and "Ga/Ba." Sixteen infants aged between 3 and 13 months old were included in this study and their fNIRS data were recorded during natural sleep. The stimuli were presented using a nonsilence baseline paradigm, where repeated standard stimuli were presented between the novel stimuli blocks without any silence periods. The morphology of fNIRS responses varied between speech contrasts. The data were fit with a model in which the responses were the sum of two independent and concurrent response mechanisms that were derived from previously published fNIRS detection responses. These independent components were an oxyhemoglobin (HbO)-positive early-latency response and an HbO-negative late latency response, hypothesized to be related to an auditory canonical response and a brain arousal response, respectively. The goodness of fit of the model with the data was high with median goodness of fit of 81%. The data showed that both response components had later latency when the left ear was the test ear (p < .05) compared to the right ear and that the negative component, due to brain arousal, was smallest for the most subtle contrast, "Ga/Ba" (p = .003).

本研究使用功能近红外光谱(fNIRS)来测量睡眠婴儿的语言识别能力。我们研究了三种不同语音对比的fNIRS反应形态,即“Tea/Ba”、“Bee/Ba”和“Ga/Ba”。16名年龄在3到13个月之间的婴儿参与了这项研究,并记录了他们在自然睡眠期间的近红外光谱数据。刺激采用非沉默基线范式,在新刺激块之间重复呈现标准刺激,没有任何沉默期。不同语音对照的fNIRS反应形态不同。数据拟合了一个模型,其中响应是两个独立且并发的响应机制的总和,这些响应机制来源于先前发表的fNIRS检测响应。这些独立的成分是一个氧合血红蛋白(HbO)阳性的早期潜伏期反应和一个氧合血红蛋白阴性的晚期潜伏期反应,假设分别与听觉规范反应和大脑唤醒反应有关。模型与数据的拟合优度较高,中位拟合优度为81%。数据显示,当左耳为测试耳时,两种反应成分的潜伏期均较晚(p p = 0.003)。
{"title":"Measuring Speech Discrimination Ability in Sleeping Infants Using fNIRS-A Proof of Principle.","authors":"Onn Wah Lee, Demi Gao, Tommy Peng, Julia Wunderlich, Darren Mao, Gautam Balasubramanian, Colette M McKay","doi":"10.1177/23312165241311721","DOIUrl":"10.1177/23312165241311721","url":null,"abstract":"<p><p>This study used functional near-infrared spectroscopy (fNIRS) to measure aspects of the speech discrimination ability of sleeping infants. We examined the morphology of the fNIRS response to three different speech contrasts, namely \"Tea/Ba,\" \"Bee/Ba,\" and \"Ga/Ba.\" Sixteen infants aged between 3 and 13 months old were included in this study and their fNIRS data were recorded during natural sleep. The stimuli were presented using a nonsilence baseline paradigm, where repeated standard stimuli were presented between the novel stimuli blocks without any silence periods. The morphology of fNIRS responses varied between speech contrasts. The data were fit with a model in which the responses were the sum of two independent and concurrent response mechanisms that were derived from previously published fNIRS detection responses. These independent components were an oxyhemoglobin (HbO)-positive early-latency response and an HbO-negative late latency response, hypothesized to be related to an auditory canonical response and a brain arousal response, respectively. The goodness of fit of the model with the data was high with median goodness of fit of 81%. The data showed that both response components had later latency when the left ear was the test ear (<i>p</i> < .05) compared to the right ear and that the negative component, due to brain arousal, was smallest for the most subtle contrast, \"Ga/Ba\" (<i>p</i> = .003).</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165241311721"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11758514/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143030151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of Speaker-Conditioned Target Speaker Extraction Algorithms for Hearing-Impaired Listeners. 针对听障听众的说话人条件目标说话人提取算法评价。
IF 3 2区 医学 Q1 AUDIOLOGY & SPEECH-LANGUAGE PATHOLOGY Pub Date : 2025-01-01 Epub Date: 2025-08-11 DOI: 10.1177/23312165251365802
Ragini Sinha, Ann-Christin Scherer, Simon Doclo, Christian Rollwage, Jan Rennies

Speaker-conditioned target speaker extraction algorithms aim at extracting the target speaker from a mixture of multiple speakers by using additional information about the target speaker. Previous studies have evaluated the performance of these algorithms using either instrumental measures or subjective assessments with normal-hearing listeners or with hearing-impaired listeners. Notably, a previous study employing a quasicausal algorithm reported significant intelligibility improvements for both normal-hearing and hearing-impaired listeners, while another study demonstrated that a fully causal algorithm could enhance speech intelligibility and reduce listening effort for normal-hearing listeners. Building on these findings, this study focuses on an in-depth subjective assessment of two fully causal deep neural network-based speaker-conditioned target speaker extraction algorithms with hearing-impaired listeners, both without hearing loss compensation (unaided) and with linear hearing loss compensation (aided). Three different subjective performance measurement methods were used to cover a broad range of listening conditions, namely paired comparison, speech recognition thresholds, and categorically scaled perceived listening effort. The subjective evaluation results with 15 hearing-impaired listeners showed that one algorithm significantly reduced listening effort and improved intelligibility compared to unprocessed stimuli and the other algorithm. The data also suggest that hearing-impaired listeners experience a greater benefit in terms of listening effort (for both male and female interfering speakers) and speech recognition thresholds, especially in the presence of female interfering speakers than normal-hearing listeners, and that hearing loss compensation (linear amplification) is not required to obtain an algorithm benefit.

基于说话人条件的目标说话人提取算法旨在利用目标说话人的附加信息从混合的多个说话人中提取目标说话人。以前的研究使用仪器测量或主观评估对听力正常或听力受损的听众评估这些算法的性能。值得注意的是,先前一项采用准因果算法的研究报告了正常听力和听力受损听众的可理解性显著提高,而另一项研究表明,完全因果算法可以提高正常听力听众的语音可理解性,减少听力努力。在这些发现的基础上,本研究重点对两种完全因果的基于深度神经网络的说话人条件目标说话人提取算法进行了深入的主观评估,这两种算法都是针对听力受损的听众,没有听力损失补偿(无辅助)和线性听力损失补偿(辅助)。我们使用了三种不同的主观表现测量方法来涵盖广泛的听力条件,即配对比较、语音识别阈值和分类缩放的感知听力努力。对15名听障听众的主观评价结果表明,与未处理的刺激和另一种算法相比,一种算法显著减少了听力努力,提高了可理解性。数据还表明,听力受损的听众在听力努力(男性和女性干扰扬声器)和语音识别阈值方面比正常听力的听众获得更大的好处,特别是在女性干扰扬声器存在的情况下,并且不需要听力损失补偿(线性放大)来获得算法优势。
{"title":"Evaluation of Speaker-Conditioned Target Speaker Extraction Algorithms for Hearing-Impaired Listeners.","authors":"Ragini Sinha, Ann-Christin Scherer, Simon Doclo, Christian Rollwage, Jan Rennies","doi":"10.1177/23312165251365802","DOIUrl":"10.1177/23312165251365802","url":null,"abstract":"<p><p>Speaker-conditioned target speaker extraction algorithms aim at extracting the target speaker from a mixture of multiple speakers by using additional information about the target speaker. Previous studies have evaluated the performance of these algorithms using either instrumental measures or subjective assessments with normal-hearing listeners or with hearing-impaired listeners. Notably, a previous study employing a quasicausal algorithm reported significant intelligibility improvements for both normal-hearing and hearing-impaired listeners, while another study demonstrated that a fully causal algorithm could enhance speech intelligibility and reduce listening effort for normal-hearing listeners. Building on these findings, this study focuses on an in-depth subjective assessment of two fully causal deep neural network-based speaker-conditioned target speaker extraction algorithms with hearing-impaired listeners, both without hearing loss compensation (unaided) and with linear hearing loss compensation (aided). Three different subjective performance measurement methods were used to cover a broad range of listening conditions, namely paired comparison, speech recognition thresholds, and categorically scaled perceived listening effort. The subjective evaluation results with 15 hearing-impaired listeners showed that one algorithm significantly reduced listening effort and improved intelligibility compared to unprocessed stimuli and the other algorithm. The data also suggest that hearing-impaired listeners experience a greater benefit in terms of listening effort (for both male and female interfering speakers) and speech recognition thresholds, especially in the presence of female interfering speakers than normal-hearing listeners, and that hearing loss compensation (linear amplification) is not required to obtain an algorithm benefit.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251365802"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12340209/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144817996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Trends in Hearing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1