电编码声音的频谱分组可预测人工耳蜗植入者的噪音中语音表现

Inyong Choi, Phillip E. Gander, Joel I. Berger, Jihwan Woo, Matthew H. Choy, Jean Hong, Sarah Colby, Bob McMurray, Timothy D. Griffiths
{"title":"电编码声音的频谱分组可预测人工耳蜗植入者的噪音中语音表现","authors":"Inyong Choi, Phillip E. Gander, Joel I. Berger, Jihwan Woo, Matthew H. Choy, Jean Hong, Sarah Colby, Bob McMurray, Timothy D. Griffiths","doi":"10.1007/s10162-023-00918-x","DOIUrl":null,"url":null,"abstract":"<h3 data-test=\"abstract-sub-heading\">Objectives</h3><p>Cochlear implant (CI) users exhibit large variability in understanding speech in noise. Past work in CI users found that spectral and temporal resolution correlates with speech-in-noise ability, but a large portion of variance remains unexplained. Recent work on normal-hearing listeners showed that the ability to group temporally and spectrally coherent tones in a complex auditory scene predicts speech-in-noise ability independently of the audiogram, highlighting a central mechanism for auditory scene analysis that contributes to speech-in-noise. The current study examined whether the auditory grouping ability also contributes to speech-in-noise understanding in CI users.</p><h3 data-test=\"abstract-sub-heading\">Design</h3><p>Forty-seven post-lingually deafened CI users were tested with psychophysical measures of spectral and temporal resolution, a stochastic figure-ground task that depends on the detection of a figure by grouping multiple fixed frequency elements against a random background, and a sentence-in-noise measure. Multiple linear regression was used to predict sentence-in-noise performance from the other tasks.</p><h3 data-test=\"abstract-sub-heading\">Results</h3><p>No co-linearity was found between any predictor variables. All three predictors (spectral and temporal resolution plus the figure-ground task) exhibited significant contribution in the multiple linear regression model, indicating that the auditory grouping ability in a complex auditory scene explains a further proportion of variance in CI users’ speech-in-noise performance that was not explained by spectral and temporal resolution.</p><h3 data-test=\"abstract-sub-heading\">Conclusion</h3><p>Measures of cross-frequency grouping reflect an auditory cognitive mechanism that determines speech-in-noise understanding independently of cochlear function. Such measures are easily implemented clinically as predictors of CI success and suggest potential strategies for rehabilitation based on training with non-speech stimuli.</p>","PeriodicalId":17236,"journal":{"name":"Journal of the Association for Research in Otolaryngology","volume":"17 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Spectral Grouping of Electrically Encoded Sound Predicts Speech-in-Noise Performance in Cochlear Implantees\",\"authors\":\"Inyong Choi, Phillip E. Gander, Joel I. Berger, Jihwan Woo, Matthew H. Choy, Jean Hong, Sarah Colby, Bob McMurray, Timothy D. Griffiths\",\"doi\":\"10.1007/s10162-023-00918-x\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<h3 data-test=\\\"abstract-sub-heading\\\">Objectives</h3><p>Cochlear implant (CI) users exhibit large variability in understanding speech in noise. Past work in CI users found that spectral and temporal resolution correlates with speech-in-noise ability, but a large portion of variance remains unexplained. Recent work on normal-hearing listeners showed that the ability to group temporally and spectrally coherent tones in a complex auditory scene predicts speech-in-noise ability independently of the audiogram, highlighting a central mechanism for auditory scene analysis that contributes to speech-in-noise. The current study examined whether the auditory grouping ability also contributes to speech-in-noise understanding in CI users.</p><h3 data-test=\\\"abstract-sub-heading\\\">Design</h3><p>Forty-seven post-lingually deafened CI users were tested with psychophysical measures of spectral and temporal resolution, a stochastic figure-ground task that depends on the detection of a figure by grouping multiple fixed frequency elements against a random background, and a sentence-in-noise measure. Multiple linear regression was used to predict sentence-in-noise performance from the other tasks.</p><h3 data-test=\\\"abstract-sub-heading\\\">Results</h3><p>No co-linearity was found between any predictor variables. All three predictors (spectral and temporal resolution plus the figure-ground task) exhibited significant contribution in the multiple linear regression model, indicating that the auditory grouping ability in a complex auditory scene explains a further proportion of variance in CI users’ speech-in-noise performance that was not explained by spectral and temporal resolution.</p><h3 data-test=\\\"abstract-sub-heading\\\">Conclusion</h3><p>Measures of cross-frequency grouping reflect an auditory cognitive mechanism that determines speech-in-noise understanding independently of cochlear function. Such measures are easily implemented clinically as predictors of CI success and suggest potential strategies for rehabilitation based on training with non-speech stimuli.</p>\",\"PeriodicalId\":17236,\"journal\":{\"name\":\"Journal of the Association for Research in Otolaryngology\",\"volume\":\"17 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-12-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of the Association for Research in Otolaryngology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/s10162-023-00918-x\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Association for Research in Otolaryngology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s10162-023-00918-x","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

目标人工耳蜗 (CI) 用户在理解噪声语音方面表现出很大的差异。过去针对 CI 使用者的研究发现,频谱和时间分辨率与噪声中的语音能力相关,但仍有很大一部分差异无法解释。最近对听力正常的听者进行的研究表明,在复杂的听觉场景中将时间和频谱上一致的音调分组的能力可以预测噪声中的语音能力,而与听力图无关,这凸显了有助于噪声中语音的听觉场景分析的核心机制。本研究考察了听觉分组能力是否也有助于 CI 用户的噪声语音理解能力。设计对 47 名耳聋后 CI 用户进行了测试,测试内容包括频谱和时间分辨率的心理物理测量、随机图形-地面任务(该任务取决于在随机背景下通过将多个固定频率元素分组来检测图形)以及噪声句子测量。结果没有发现任何预测变量之间存在共线性。在多元线性回归模型中,所有三个预测变量(频谱和时间分辨率加上图-地任务)都有显著贡献,这表明在复杂的听觉场景中,听觉分组能力可以解释 CI 用户噪声中言语能力差异的另一部分,而这部分差异是频谱和时间分辨率无法解释的。这种测量方法在临床上很容易应用,可作为人工耳蜗成功与否的预测指标,并提出了基于非语音刺激训练的潜在康复策略。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

摘要图片

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Spectral Grouping of Electrically Encoded Sound Predicts Speech-in-Noise Performance in Cochlear Implantees

Objectives

Cochlear implant (CI) users exhibit large variability in understanding speech in noise. Past work in CI users found that spectral and temporal resolution correlates with speech-in-noise ability, but a large portion of variance remains unexplained. Recent work on normal-hearing listeners showed that the ability to group temporally and spectrally coherent tones in a complex auditory scene predicts speech-in-noise ability independently of the audiogram, highlighting a central mechanism for auditory scene analysis that contributes to speech-in-noise. The current study examined whether the auditory grouping ability also contributes to speech-in-noise understanding in CI users.

Design

Forty-seven post-lingually deafened CI users were tested with psychophysical measures of spectral and temporal resolution, a stochastic figure-ground task that depends on the detection of a figure by grouping multiple fixed frequency elements against a random background, and a sentence-in-noise measure. Multiple linear regression was used to predict sentence-in-noise performance from the other tasks.

Results

No co-linearity was found between any predictor variables. All three predictors (spectral and temporal resolution plus the figure-ground task) exhibited significant contribution in the multiple linear regression model, indicating that the auditory grouping ability in a complex auditory scene explains a further proportion of variance in CI users’ speech-in-noise performance that was not explained by spectral and temporal resolution.

Conclusion

Measures of cross-frequency grouping reflect an auditory cognitive mechanism that determines speech-in-noise understanding independently of cochlear function. Such measures are easily implemented clinically as predictors of CI success and suggest potential strategies for rehabilitation based on training with non-speech stimuli.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Cochlear Aqueduct Post-Natal Growth: A Computed Tomography Study Otoacoustic Estimate of Astronauts’ Intracranial Pressure Changes During Spaceflight FDA-Approved Tedizolid Phosphate Prevents Cisplatin-Induced Hearing Loss Without Decreasing Its Anti-tumor Effect Hampshire Sheep as a Large-Animal Model for Cochlear Implantation Echolocating Bats Have Evolved Decreased Susceptibility to Noise-Induced Temporary Hearing Losses
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1