Pub Date : 2025-01-01Epub Date: 2025-11-24DOI: 10.1177/23312165251397699
Qi Gao, Lena L N Wong, Fei Chen
This study investigated the effect of temporal misalignment between acoustic and simulated electric signals on the ability to process fast speech in normal-hearing listeners. The within-ear integration of acoustic and electric hearing was simulated, mimicking the electric-acoustic stimulation (EAS) condition, where cochlear implant users receive acoustic input at low frequencies and electric stimulation at high frequencies in the same ear. Time-compression thresholds (TCTs), defined as the 50% correct performance for time-compressed sentences, were adaptively measured in quiet and in speech-spectrum noise (SSN) as well as amplitude-modulated noise (AMN) at 4 dB and 10 dB signal-to-noise ratio (SNR). Temporal misalignment was introduced by delaying the acoustic or the simulated electric signals, which were generated using a low-pass filter (cutoff frequency: 600 Hz) and a five-channel noise vocoder, respectively. Listeners showed significant benefits from the addition of low-frequency acoustic signals in terms of TCTs, regardless of temporal misalignment. Within the range from 0 ms to ±30 ms, temporal misalignment decreased listeners' TCTs, and its effect interacted with SNR such that the adverse impact of misalignment was more pronounced at higher SNR levels. When misalignment was limited to within ±7 ms, which is closer to the clinically relevant range, its effect disappeared. In conclusion, while temporal misalignment negatively affects the ability of listeners with simulated EAS hearing to process fast sentences in Mandarin, its effect is negligible when it is close to a clinically relevant range. Future research should validate these findings in real EAS users.
{"title":"The Effect of Temporal Misalignment Between Acoustic and Simulated Electric Signals on the Time Compression Thresholds of Normal-Hearing Listeners.","authors":"Qi Gao, Lena L N Wong, Fei Chen","doi":"10.1177/23312165251397699","DOIUrl":"10.1177/23312165251397699","url":null,"abstract":"<p><p>This study investigated the effect of temporal misalignment between acoustic and simulated electric signals on the ability to process fast speech in normal-hearing listeners. The within-ear integration of acoustic and electric hearing was simulated, mimicking the electric-acoustic stimulation (EAS) condition, where cochlear implant users receive acoustic input at low frequencies and electric stimulation at high frequencies in the same ear. Time-compression thresholds (TCTs), defined as the 50% correct performance for time-compressed sentences, were adaptively measured in quiet and in speech-spectrum noise (SSN) as well as amplitude-modulated noise (AMN) at 4 dB and 10 dB signal-to-noise ratio (SNR). Temporal misalignment was introduced by delaying the acoustic or the simulated electric signals, which were generated using a low-pass filter (cutoff frequency: 600 Hz) and a five-channel noise vocoder, respectively. Listeners showed significant benefits from the addition of low-frequency acoustic signals in terms of TCTs, regardless of temporal misalignment. Within the range from 0 ms to ±30 ms, temporal misalignment decreased listeners' TCTs, and its effect interacted with SNR such that the adverse impact of misalignment was more pronounced at higher SNR levels. When misalignment was limited to within ±7 ms, which is closer to the clinically relevant range, its effect disappeared. In conclusion, while temporal misalignment negatively affects the ability of listeners with simulated EAS hearing to process fast sentences in Mandarin, its effect is negligible when it is close to a clinically relevant range. Future research should validate these findings in real EAS users.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251397699"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12644445/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145597739","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-08-25DOI: 10.1177/23312165251368669
Robin Hake, Michel Bürgel, Christophe Lesimple, Matthias Vormann, Kirsten C Wagener, Volker Kuehnel, Kai Siedenburg
Hearing aids have traditionally been designed to facilitate speech perception. With regards to music perception, previous work indicates that hearing aid users frequently complain about music sound quality. Yet, the effects of hearing aid amplification on musical perception abilities are largely unknown. This study aimed to investigate the effects of hearing aid amplification and dynamic range compression (DRC) settings on musical scene analysis (MSA) abilities and sound quality ratings (SQR) using polyphonic music recordings. Additionally, speech reception thresholds in noise (SRT) were measured. Thirty-three hearing aid users with moderate to severe hearing loss participated in three conditions: unaided, and aided with either slow or fast DRC settings. Overall, MSA abilities, SQR and SRT significantly improved with the use of hearing aids compared to the unaided condition. Yet, differences were observed regarding the choice of compression settings. Fast DRC led to better MSA performance, reflecting enhanced selective listening in musical mixtures, while slow DRC elicited more favorable SQR. Despite these improvements, variability in amplification benefit across DRC settings and tasks remained considerable, with some individuals showing limited or no improvement. These findings highlight a trade-off between scene transparency (indexed by MSA) and perceived sound quality, with individual differences emerging as a key factor in shaping amplification outcomes. Our results underscore the potential benefits of hearing aids for music perception and indicate the need for personalized fitting strategies tailored to task-specific demands.
{"title":"Perception of Recorded Music With Hearing Aids: Compression Differentially Affects Musical Scene Analysis and Musical Sound Quality.","authors":"Robin Hake, Michel Bürgel, Christophe Lesimple, Matthias Vormann, Kirsten C Wagener, Volker Kuehnel, Kai Siedenburg","doi":"10.1177/23312165251368669","DOIUrl":"https://doi.org/10.1177/23312165251368669","url":null,"abstract":"<p><p>Hearing aids have traditionally been designed to facilitate speech perception. With regards to music perception, previous work indicates that hearing aid users frequently complain about music sound quality. Yet, the effects of hearing aid amplification on musical perception abilities are largely unknown. This study aimed to investigate the effects of hearing aid amplification and dynamic range compression (DRC) settings on musical scene analysis (MSA) abilities and sound quality ratings (SQR) using polyphonic music recordings. Additionally, speech reception thresholds in noise (SRT) were measured. Thirty-three hearing aid users with moderate to severe hearing loss participated in three conditions: unaided, and aided with either slow or fast DRC settings. Overall, MSA abilities, SQR and SRT significantly improved with the use of hearing aids compared to the unaided condition. Yet, differences were observed regarding the choice of compression settings. Fast DRC led to better MSA performance, reflecting enhanced selective listening in musical mixtures, while slow DRC elicited more favorable SQR. Despite these improvements, variability in amplification benefit across DRC settings and tasks remained considerable, with some individuals showing limited or no improvement. These findings highlight a trade-off between scene transparency (indexed by MSA) and perceived sound quality, with individual differences emerging as a key factor in shaping amplification outcomes. Our results underscore the potential benefits of hearing aids for music perception and indicate the need for personalized fitting strategies tailored to task-specific demands.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251368669"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12378302/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144975114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01DOI: 10.1177/23312165241312449
Aaron C Moberly, Liping Du, Terrin N Tamati
When listening to speech under adverse conditions, listeners compensate using neurocognitive resources. A clinically relevant form of adverse listening is listening through a cochlear implant (CI), which provides a spectrally degraded signal. CI listening is often simulated through noise-vocoding. This study investigated the neurocognitive mechanisms supporting recognition of spectrally degraded speech in adult CI users and normal-hearing (NH) peers listening to noise-vocoded speech, with the hypothesis that an overlapping set of neurocognitive functions would contribute to speech recognition in both groups. Ninety-seven adults with either a CI (54 CI individuals, mean age 66.6 years, range 45-87 years) or age-normal hearing (43 NH individuals, mean age 66.8 years, range 50-81 years) participated. Listeners heard materials varying in linguistic complexity consisting of isolated words, meaningful sentences, anomalous sentences, high-variability sentences, and audiovisually (AV) presented sentences. Participants were also tested for vocabulary knowledge, nonverbal reasoning, working memory capacity, inhibition-concentration, and speed of lexical and phonological access. Linear regression analyses with robust standard errors were performed for speech recognition tasks on neurocognitive functions. Nonverbal reasoning contributed to meaningful sentence recognition in NH peers and anomalous sentence recognition in CI users. Speed of lexical access contributed to performance on most speech tasks for CI users but not for NH peers. Finally, inhibition-concentration and vocabulary knowledge contributed to AV sentence recognition in NH listeners alone. Findings suggest that the complexity of speech materials may determine the particular contributions of neurocognitive skills, and that NH processing of noise-vocoded speech may not represent how CI listeners process speech.
{"title":"Individual Differences in the Recognition of Spectrally Degraded Speech: Associations With Neurocognitive Functions in Adult Cochlear Implant Users and With Noise-Vocoded Simulations.","authors":"Aaron C Moberly, Liping Du, Terrin N Tamati","doi":"10.1177/23312165241312449","DOIUrl":"10.1177/23312165241312449","url":null,"abstract":"<p><p>When listening to speech under adverse conditions, listeners compensate using neurocognitive resources. A clinically relevant form of adverse listening is listening through a cochlear implant (CI), which provides a spectrally degraded signal. CI listening is often simulated through noise-vocoding. This study investigated the neurocognitive mechanisms supporting recognition of spectrally degraded speech in adult CI users and normal-hearing (NH) peers listening to noise-vocoded speech, with the hypothesis that an overlapping set of neurocognitive functions would contribute to speech recognition in both groups. Ninety-seven adults with either a CI (54 CI individuals, mean age 66.6 years, range 45-87 years) or age-normal hearing (43 NH individuals, mean age 66.8 years, range 50-81 years) participated. Listeners heard materials varying in linguistic complexity consisting of isolated words, meaningful sentences, anomalous sentences, high-variability sentences, and audiovisually (AV) presented sentences. Participants were also tested for vocabulary knowledge, nonverbal reasoning, working memory capacity, inhibition-concentration, and speed of lexical and phonological access. Linear regression analyses with robust standard errors were performed for speech recognition tasks on neurocognitive functions. Nonverbal reasoning contributed to meaningful sentence recognition in NH peers and anomalous sentence recognition in CI users. Speed of lexical access contributed to performance on most speech tasks for CI users but not for NH peers. Finally, inhibition-concentration and vocabulary knowledge contributed to AV sentence recognition in NH listeners alone. Findings suggest that the complexity of speech materials may determine the particular contributions of neurocognitive skills, and that NH processing of noise-vocoded speech may not represent how CI listeners process speech.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165241312449"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11742172/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143014599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-03-18DOI: 10.1177/23312165251317027
Pedro Lladó, Piotr Majdak, Roberto Barumerli, Robert Baumgartner
Localization of sound sources in sagittal planes significantly relies on monaural spectral cues. These cues are primarily derived from the direction-specific filtering of the pinnae. The contribution of specific frequency regions to the cue evaluation has not been fully clarified. To this end, we analyzed how different spectral weighting schemes contribute to the explanatory power of a sagittal-plane localization model in response to wideband, flat-spectrum stimuli. Each weighting scheme emphasized the contribution of spectral cues within well-defined frequency bands, enabling us to assess their impact on the predictions of individual patterns of localization responses. By means of Bayesian model selection, we compared five model variants representing various spectral weights. Our results indicate a preference for the weighting schemes emphasizing the contribution of frequencies above 8 kHz, suggesting that, in the auditory system, spectral cue evaluation is upweighted in that frequency region. While various potential explanations are discussed, we conclude that special attention should be put on this high-frequency region in spatial-audio applications aiming at the best localization performance.
{"title":"Spectral Weighting of Monaural Cues for Auditory Localization in Sagittal Planes.","authors":"Pedro Lladó, Piotr Majdak, Roberto Barumerli, Robert Baumgartner","doi":"10.1177/23312165251317027","DOIUrl":"10.1177/23312165251317027","url":null,"abstract":"<p><p>Localization of sound sources in sagittal planes significantly relies on monaural spectral cues. These cues are primarily derived from the direction-specific filtering of the pinnae. The contribution of specific frequency regions to the cue evaluation has not been fully clarified. To this end, we analyzed how different spectral weighting schemes contribute to the explanatory power of a sagittal-plane localization model in response to wideband, flat-spectrum stimuli. Each weighting scheme emphasized the contribution of spectral cues within well-defined frequency bands, enabling us to assess their impact on the predictions of individual patterns of localization responses. By means of Bayesian model selection, we compared five model variants representing various spectral weights. Our results indicate a preference for the weighting schemes emphasizing the contribution of frequencies above 8 kHz, suggesting that, in the auditory system, spectral cue evaluation is upweighted in that frequency region. While various potential explanations are discussed, we conclude that special attention should be put on this high-frequency region in spatial-audio applications aiming at the best localization performance.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251317027"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11920987/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143659047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-12-08DOI: 10.1177/23312165251401318
Freja Baxter, Harriet J Smith, Emma Holmes
Understanding speech among competing speech poses a substantial challenge. In these environments, familiar voices-including naturally familiar (e.g., friends, partners) and lab-trained voices-are more intelligible than unfamiliar voices. Yet, whether familiar voices also require less effort to understand is currently unknown. We trained 20 participants to become familiar with three voices, then tested listening effort during a speech intelligibility task. During familiarization and training, participants were exposed to three talkers for different lengths of time, either speaking 88, 166, or 478 sentences ("Least Familiar," "Moderately Familiar," or "Most Familiar" voice, respectively). During each trial of the speech intelligibility task, two competing sentences were presented at a target-to-masker ratio (TMR) of -6 or +3 dB. Participants reported target sentences that were spoken by trained or by novel, unfamiliar talkers. We assessed effort using self-reported ratings and physiologically, using pupil dilation. We found that self-report scores were more sensitive than pupil dilation to differences in TMR, with lower self-reported effort at +3 than -6 dB TMR. The two measures may also be differentially sensitive to the extent of training. We found lower self-reported effort for all three trained voices over unfamiliar voices, with no differences among the trained voices, whereas pupil dilation was only lower for the voice that had been trained for the longest. Thus, both self-report scores and pupil dilation showed advantages for the voice that was trained for the longest (∼1 h), but self-report scores additionally showed reduced effort even following relatively short durations of training (<10 min).
理解相互竞争的语言构成了巨大的挑战。在这些环境中,熟悉的声音——包括自然熟悉的声音(如朋友、伙伴)和实验室训练的声音——比不熟悉的声音更容易理解。然而,熟悉的声音是否也需要更少的努力来理解,目前尚不清楚。我们训练了20名参与者熟悉三种声音,然后在语音清晰度任务中测试听力的努力程度。在熟悉和训练期间,参与者被暴露在三个说话者的不同时间长度下,分别说88、166或478个句子(分别是“最不熟悉”、“一般熟悉”或“最熟悉”的声音)。在语音可理解性任务的每次试验中,以-6或+3 dB的目标与掩蔽比(TMR)呈现两个相互竞争的句子。参与者报告的目标句子是由训练有素的或陌生的说话者说的。我们用自我报告的评分来评估努力程度,生理学上用瞳孔扩张来评估努力程度。我们发现自我报告得分比瞳孔扩张对TMR差异更敏感,+3 dB TMR时自我报告的努力程度低于-6 dB TMR。这两种方法对训练程度的敏感程度也可能不同。我们发现,与不熟悉的声音相比,所有三种经过训练的声音的自我报告努力程度都较低,而经过训练的声音之间没有差异,而瞳孔扩张仅在训练时间最长的声音中较低。因此,自我报告分数和瞳孔扩张都显示了训练时间最长(约1小时)的声音的优势,但自我报告分数还显示,即使在相对较短的训练时间后,努力程度也会降低(
{"title":"Voice Familiarization Training Improves Speech Intelligibility and Reduces Listening Effort.","authors":"Freja Baxter, Harriet J Smith, Emma Holmes","doi":"10.1177/23312165251401318","DOIUrl":"10.1177/23312165251401318","url":null,"abstract":"<p><p>Understanding speech among competing speech poses a substantial challenge. In these environments, familiar voices-including naturally familiar (e.g., friends, partners) and lab-trained voices-are more intelligible than unfamiliar voices. Yet, whether familiar voices also require less effort to understand is currently unknown. We trained 20 participants to become familiar with three voices, then tested listening effort during a speech intelligibility task. During familiarization and training, participants were exposed to three talkers for different lengths of time, either speaking 88, 166, or 478 sentences (\"Least Familiar,\" \"Moderately Familiar,\" or \"Most Familiar\" voice, respectively). During each trial of the speech intelligibility task, two competing sentences were presented at a target-to-masker ratio (TMR) of -6 or +3 dB. Participants reported target sentences that were spoken by trained or by novel, unfamiliar talkers. We assessed effort using self-reported ratings and physiologically, using pupil dilation. We found that self-report scores were more sensitive than pupil dilation to differences in TMR, with lower self-reported effort at +3 than -6 dB TMR. The two measures may also be differentially sensitive to the extent of training. We found lower self-reported effort for all three trained voices over unfamiliar voices, with no differences among the trained voices, whereas pupil dilation was only lower for the voice that had been trained for the longest. Thus, both self-report scores and pupil dilation showed advantages for the voice that was trained for the longest (∼1 h), but self-report scores additionally showed reduced effort even following relatively short durations of training (<10 min).</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251401318"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12686366/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145702706","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-07-29DOI: 10.1177/23312165251364675
Courtney Coburn Glavin, Sumitrajit Dhar
Age-related hearing loss (ARHL) currently affects over 20 million adults in the U.S. and its prevalence is expected to increase as the population ages. However, little is known about the earliest manifestations of ARHL, including its influence on auditory function beyond the threshold of sensation. This work explores the effects of early aging on frequency selectivity (i.e., "tuning"), a critical feature of normal hearing function. Tuning is estimated using both behavioral and physiological measures-fast psychophysical tuning curves (fPTC), distortion product otoacoustic emission level ratio functions (DPOAE LRFs), and stimulus-frequency OAE (SFOAE) phase gradient delay. All three measures were selected because they have high potential for clinical translation but have not been compared directly in the same sample of ears. Results indicate that there may be subtle changes in tuning during early aging, even in ears with clinically normal audiometric thresholds. Additionally, there are notable differences in tuning estimates derived from the three measures. Psychophysical tuning estimates are highly variable and statistically significantly different from OAE-derived tuning estimates, suggesting that behavioral tuning is uniquely influenced by factors not affecting OAE-based tuning. Across all measures, there is considerable individual variability that warrants future investigation. Collectively, this work suggests that age-related auditory decline begins in relatively young ears (<60 years) and in the absence of traditionally defined "hearing loss." These findings suggest the potential benefit of characterizing ARHL beyond threshold and establishing a gold standard for measuring frequency selectivity in humans.
{"title":"Cochlear Tuning in Early Aging Estimated with Three Methods.","authors":"Courtney Coburn Glavin, Sumitrajit Dhar","doi":"10.1177/23312165251364675","DOIUrl":"10.1177/23312165251364675","url":null,"abstract":"<p><p>Age-related hearing loss (ARHL) currently affects over 20 million adults in the U.S. and its prevalence is expected to increase as the population ages. However, little is known about the earliest manifestations of ARHL, including its influence on auditory function beyond the threshold of sensation. This work explores the effects of early aging on frequency selectivity (i.e., \"tuning\"), a critical feature of normal hearing function. Tuning is estimated using both behavioral and physiological measures-fast psychophysical tuning curves (fPTC), distortion product otoacoustic emission level ratio functions (DPOAE LRFs), and stimulus-frequency OAE (SFOAE) phase gradient delay. All three measures were selected because they have high potential for clinical translation but have not been compared directly in the same sample of ears. Results indicate that there may be subtle changes in tuning during early aging, even in ears with clinically normal audiometric thresholds. Additionally, there are notable differences in tuning estimates derived from the three measures. Psychophysical tuning estimates are highly variable and statistically significantly different from OAE-derived tuning estimates, suggesting that behavioral tuning is uniquely influenced by factors not affecting OAE-based tuning. Across all measures, there is considerable individual variability that warrants future investigation. Collectively, this work suggests that age-related auditory decline begins in relatively young ears (<60 years) and in the absence of traditionally defined \"hearing loss.\" These findings suggest the potential benefit of characterizing ARHL beyond threshold and establishing a gold standard for measuring frequency selectivity in humans.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251364675"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12317184/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144745544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-08-18DOI: 10.1177/23312165251370006
Hendrik Husstedt, Jennifer Schmidt, Luca Wiederschein, Robert Wiedenbeck, Markus Kemper, Florian Denk
In addition to speech intelligibility, listening effort has emerged as a critical indicator of hearing performance. It can be defined as the effort experienced or invested in solving an auditory task. Subjective, behavioral, and physiological methods have been employed to assess listening effort. While previous studies have focused predominantly evaluated listening effort at clearly audible levels, such as in speech-in-noise conditions, we present findings from a study investigating listening effort for soft speech in quiet. Twenty young adults with normal hearing participated in speech intelligibility testing (OLSA), adaptive listening effort scaling (ACALES), and pupillometry. Experienced effort decreased with increasing speech level and "no effort" was reached at 40 dB sound pressure level (SPL). The difference between levels rated with "extreme effort" and "no effort" was, on average, 20.6 dB SPL. Thus, speech must be presented well above the speech-recognition threshold in quiet to achieve effortless listening. These results prompted a follow-up experiment involving 18 additional participants, who completed OLSA and ACALES tests with hearing threshold-simulating noise at conversational levels. Comparing the results of the main and follow-up experiments suggests that the observations in quiet cannot be fully attributed to the masking effects of internal noise but likely also reflect cognitive processes that are not yet fully understood. These findings have important implications, particularly regarding the benefits of amplification for soft sounds. We propose that the concept of a threshold for effortless listening has been overlooked and should be prioritized in future research, especially in the context of soft speech in quiet environments.
除了言语可理解性,听力努力也成为听力表现的一个重要指标。它可以被定义为在解决听觉任务中所经历或投入的努力。主观的、行为的和生理的方法被用来评估听力努力。虽然以前的研究主要集中在评估清晰可听水平下的听力努力,例如在噪音条件下的语音,但我们的研究结果来自于一项研究,调查了安静环境下软语的听力努力。20名听力正常的年轻人参加了语音清晰度测试(OLSA)、自适应听力努力量表(ACALES)和瞳孔测量。声压级(SPL)为40 dB时达到“不费力”;被评为“极度努力”和“不努力”的水平之间的差异平均为20.6 dB SPL。因此,语音必须在安静的情况下远高于语音识别阈值,以实现轻松的聆听。这些结果促使另外18名参与者进行了后续实验,他们在会话水平的听力阈值模拟噪音下完成了OLSA和ACALES测试。比较主要实验和后续实验的结果表明,安静环境下的观察结果不能完全归因于内部噪音的掩蔽效应,而可能也反映了尚未完全理解的认知过程。这些发现具有重要的意义,特别是关于对柔和声音的放大的好处。我们认为,在未来的研究中,特别是在安静环境中的软语环境中,容易倾听的阈值概念被忽视了,应该优先考虑。
{"title":"Listening Effort for Soft Speech in Quiet.","authors":"Hendrik Husstedt, Jennifer Schmidt, Luca Wiederschein, Robert Wiedenbeck, Markus Kemper, Florian Denk","doi":"10.1177/23312165251370006","DOIUrl":"10.1177/23312165251370006","url":null,"abstract":"<p><p>In addition to speech intelligibility, listening effort has emerged as a critical indicator of hearing performance. It can be defined as the effort experienced or invested in solving an auditory task. Subjective, behavioral, and physiological methods have been employed to assess listening effort. While previous studies have focused predominantly evaluated listening effort at clearly audible levels, such as in speech-in-noise conditions, we present findings from a study investigating listening effort for soft speech in quiet. Twenty young adults with normal hearing participated in speech intelligibility testing (OLSA), adaptive listening effort scaling (ACALES), and pupillometry. Experienced effort decreased with increasing speech level and \"no effort\" was reached at 40 dB sound pressure level (SPL). The difference between levels rated with \"extreme effort\" and \"no effort\" was, on average, 20.6 dB SPL. Thus, speech must be presented well above the speech-recognition threshold in quiet to achieve effortless listening. These results prompted a follow-up experiment involving 18 additional participants, who completed OLSA and ACALES tests with hearing threshold-simulating noise at conversational levels. Comparing the results of the main and follow-up experiments suggests that the observations in quiet cannot be fully attributed to the masking effects of internal noise but likely also reflect cognitive processes that are not yet fully understood. These findings have important implications, particularly regarding the benefits of amplification for soft sounds. We propose that the concept of a threshold for effortless listening has been overlooked and should be prioritized in future research, especially in the context of soft speech in quiet environments.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251370006"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12365469/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144876067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01DOI: 10.1177/23312165241309301
Huiyong Zhang, Brian C J Moore, Feng Jiang, Mingfang Diao, Fei Ji, Xiaodong Li, Chengshi Zheng
Wide dynamic range compression (WDRC) and noise reduction both play important roles in hearing aids. WDRC provides level-dependent amplification so that the level of sound produced by the hearing aid falls between the hearing threshold and the highest comfortable level of the listener, while noise reduction reduces ambient noise with the goal of improving intelligibility and listening comfort and reducing effort. In most current hearing aids, noise reduction and WDRC are implemented sequentially, but this may lead to distortion of the amplitude modulation patterns of both the speech and the noise. This paper describes a deep learning method, called Neural-WDRC, for implementing both noise reduction and WDRC, employing a two-stage low-complexity network. The network initially estimates the noise alone and the speech alone. Fast-acting compression is applied to the estimated speech and slow-acting compression to the estimated noise, but with a controllable residual noise level to help the user to perceive natural environmental sounds. Neural-WDRC is frame-based, and the output of the current frame is determined only by the current and preceding frames. Neural-WDRC was compared with conventional slow- and fast-acting compression and with signal-to-noise ratio (SNR)-aware compression using objective measures and listening tests based on normal-hearing participants listening to signals processed to simulate the effects of hearing loss and hearing-impaired participants. The objective measures demonstrated that Neural-WDRC effectively reduced negative interactions of speech and noise in highly non-stationary noise scenarios. The listening tests showed that Neural-WDRC was preferred over the other compression methods for speech in non-stationary noises.
{"title":"Neural-WDRC: A Deep Learning Wide Dynamic Range Compression Method Combined With Controllable Noise Reduction for Hearing Aids.","authors":"Huiyong Zhang, Brian C J Moore, Feng Jiang, Mingfang Diao, Fei Ji, Xiaodong Li, Chengshi Zheng","doi":"10.1177/23312165241309301","DOIUrl":"10.1177/23312165241309301","url":null,"abstract":"<p><p>Wide dynamic range compression (WDRC) and noise reduction both play important roles in hearing aids. WDRC provides level-dependent amplification so that the level of sound produced by the hearing aid falls between the hearing threshold and the highest comfortable level of the listener, while noise reduction reduces ambient noise with the goal of improving intelligibility and listening comfort and reducing effort. In most current hearing aids, noise reduction and WDRC are implemented sequentially, but this may lead to distortion of the amplitude modulation patterns of both the speech and the noise. This paper describes a deep learning method, called Neural-WDRC, for implementing both noise reduction and WDRC, employing a two-stage low-complexity network. The network initially estimates the noise alone and the speech alone. Fast-acting compression is applied to the estimated speech and slow-acting compression to the estimated noise, but with a controllable residual noise level to help the user to perceive natural environmental sounds. Neural-WDRC is frame-based, and the output of the current frame is determined only by the current and preceding frames. Neural-WDRC was compared with conventional slow- and fast-acting compression and with signal-to-noise ratio (SNR)-aware compression using objective measures and listening tests based on normal-hearing participants listening to signals processed to simulate the effects of hearing loss and hearing-impaired participants. The objective measures demonstrated that Neural-WDRC effectively reduced negative interactions of speech and noise in highly non-stationary noise scenarios. The listening tests showed that Neural-WDRC was preferred over the other compression methods for speech in non-stationary noises.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165241309301"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11770718/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143048166","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-06-27DOI: 10.1177/23312165251347138
Pam Dawson, Amanda Fullerton, Harish Krishnamoorthi, Kerrie Plant, Robert Cowan, Nadine Buczak, Christopher Long, Chris J James, Fergio Sismono, Andreas Büchner
This study investigated which of a range of factors could explain performance in two distinct groups of experienced, adult cochlear implant recipients differentiated by performance on words in quiet: 72 with poorer word scores versus 77 with better word scores. Tests measured the potential contribution of sound processor mapping, electrode placement, neural health, impedance, cognitive, and patient-related factors in predicting performance. A systematically measured sound processor MAP was compared to the subject's walk-in MAP. Electrode placement included modiolar distance, basal and apical insertion angle, and presence of scalar translocation. Neural health measurements included bipolar thresholds, polarity effect using asymmetrical pulses, and evoked compound action potential (ECAP) measures such as the interphase gap (IPG) effect, total refractory time, and panoramic ECAP. Impedance measurements included trans impedance matrix and four-point impedance. Cognitive tests comprised vocabulary ability, the Stroop test, and the Symbol Digits Modality Test. Performance was measured with words in quiet and sentence in noise tests and basic auditory sensitivity measures including phoneme discrimination in noise and quiet, amplitude modulation detection thresholds and quick spectral modulation detection. A range of predictor variables accounted for between 33% and 60% of the variability in performance outcomes. Multivariable regression analyses showed four key factors that were consistently predictive of poorer performance across several outcomes: substantially underfitted sound processor MAP thresholds, higher average bipolar thresholds, greater total refractory time, and greater IPG offset. Scalar translocation, cognitive variables, and other patient related factors were also significant predictors across more than one performance outcome.
{"title":"A Prospective, Multicentre Case-Control Trial Examining Factors That Explain Variable Clinical Performance in Post Lingual Adult CI Recipients.","authors":"Pam Dawson, Amanda Fullerton, Harish Krishnamoorthi, Kerrie Plant, Robert Cowan, Nadine Buczak, Christopher Long, Chris J James, Fergio Sismono, Andreas Büchner","doi":"10.1177/23312165251347138","DOIUrl":"10.1177/23312165251347138","url":null,"abstract":"<p><p>This study investigated which of a range of factors could explain performance in two distinct groups of experienced, adult cochlear implant recipients differentiated by performance on words in quiet: 72 with poorer word scores versus 77 with better word scores. Tests measured the potential contribution of sound processor mapping, electrode placement, neural health, impedance, cognitive, and patient-related factors in predicting performance. A systematically measured sound processor MAP was compared to the subject's walk-in MAP. Electrode placement included modiolar distance, basal and apical insertion angle, and presence of scalar translocation. Neural health measurements included bipolar thresholds, polarity effect using asymmetrical pulses, and evoked compound action potential (ECAP) measures such as the interphase gap (IPG) effect, total refractory time, and panoramic ECAP. Impedance measurements included trans impedance matrix and four-point impedance. Cognitive tests comprised vocabulary ability, the Stroop test, and the Symbol Digits Modality Test. Performance was measured with words in quiet and sentence in noise tests and basic auditory sensitivity measures including phoneme discrimination in noise and quiet, amplitude modulation detection thresholds and quick spectral modulation detection. A range of predictor variables accounted for between 33% and 60% of the variability in performance outcomes. Multivariable regression analyses showed four key factors that were consistently predictive of poorer performance across several outcomes: substantially underfitted sound processor MAP thresholds, higher average bipolar thresholds, greater total refractory time, and greater IPG offset. Scalar translocation, cognitive variables, and other patient related factors were also significant predictors across more than one performance outcome.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251347138"},"PeriodicalIF":2.6,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12205208/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144508936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-01-01Epub Date: 2025-09-19DOI: 10.1177/23312165251375892
Ruijing Ning, Carine Signoret, Emil Holmer, Henrik Danielsson
This study investigates the impact of hearing aid (HA) use on visual lexical decision (LD) performance in individuals with hearing loss. We hypothesize that HA use benefits phonological processing and leads to faster and more accurate visual LD. We compared the visual LD performance among three groups: 92 short-term HA users (<5 years), 98 long-term HA users, and 55 nonusers, while controlling for hearing level, age, and years of education. Results showed that, compared with non-HA users, HA users showed significantly faster reaction times in visual LD, specifically, long-term HA use was associated with smaller difference in reaction time for pseudowords compared to nonwords. These results suggest that HA use is associated with faster visual word recognition, potentially reflecting enhanced cognitive functions beyond auditory processing. These findings point to possible cognitive advantages linked to HA use.
{"title":"Hearing Aid Use is Associated with Faster Visual Lexical Decision.","authors":"Ruijing Ning, Carine Signoret, Emil Holmer, Henrik Danielsson","doi":"10.1177/23312165251375892","DOIUrl":"10.1177/23312165251375892","url":null,"abstract":"<p><p>This study investigates the impact of hearing aid (HA) use on visual lexical decision (LD) performance in individuals with hearing loss. We hypothesize that HA use benefits phonological processing and leads to faster and more accurate visual LD. We compared the visual LD performance among three groups: 92 short-term HA users (<5 years), 98 long-term HA users, and 55 nonusers, while controlling for hearing level, age, and years of education. Results showed that, compared with non-HA users, HA users showed significantly faster reaction times in visual LD, specifically, long-term HA use was associated with smaller difference in reaction time for pseudowords compared to nonwords. These results suggest that HA use is associated with faster visual word recognition, potentially reflecting enhanced cognitive functions beyond auditory processing. These findings point to possible cognitive advantages linked to HA use.</p>","PeriodicalId":48678,"journal":{"name":"Trends in Hearing","volume":"29 ","pages":"23312165251375892"},"PeriodicalIF":3.0,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12449647/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145092808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}