通过对瞥见和掩蔽声音事件的颅内反应进行分类改进听觉注意力解码

Vinay S. Raghavan, James A. O'Sullivan, J. Herrero, Stephan Bickel, A. Mehta, N. Mesgarani
{"title":"通过对瞥见和掩蔽声音事件的颅内反应进行分类改进听觉注意力解码","authors":"Vinay S. Raghavan, James A. O'Sullivan, J. Herrero, Stephan Bickel, A. Mehta, N. Mesgarani","doi":"10.1162/imag_a_00148","DOIUrl":null,"url":null,"abstract":"Abstract Listeners with hearing loss have trouble following a conversation in multitalker environments. While modern hearing aids can generally amplify speech, these devices are unable to tune into a target speaker without first knowing to which speaker a user aims to attend. Brain-controlled hearing aids have been proposed using auditory attention decoding (AAD) methods, but current methods use the same model to compare the speech stimulus and neural response, regardless of the dynamic overlap between talkers which is known to influence neural encoding. Here, we propose a novel framework that directly classifies event-related potentials (ERPs) evoked by glimpsed and masked acoustic events to determine whether the source of the event was attended. We present a system that identifies auditory events using the local maxima in the envelope rate of change, assesses the temporal masking of auditory events relative to competing speakers, and utilizes masking-specific ERP classifiers to determine if the source of the event was attended. Using intracranial electrophysiological recordings, we showed that high gamma ERPs from recording sites in auditory cortex can effectively decode the attention of subjects. This method of AAD provides higher accuracy, shorter switch times, and more stable decoding results compared with traditional correlational methods, permitting the quick and accurate detection of changes in a listener’s attentional focus. This framework also holds unique potential for detecting instances of divided attention and inattention. Overall, we extend the scope of AAD algorithms by introducing the first linear, direct-classification method for determining a listener’s attentional focus that leverages the latest research in multitalker speech perception. This work represents another step toward informing the development of effective and intuitive brain-controlled hearing assistive devices.","PeriodicalId":507939,"journal":{"name":"Imaging Neuroscience","volume":"339 4","pages":"1-19"},"PeriodicalIF":0.0000,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving auditory attention decoding by classifying intracranial responses to glimpsed and masked acoustic events\",\"authors\":\"Vinay S. Raghavan, James A. O'Sullivan, J. Herrero, Stephan Bickel, A. Mehta, N. Mesgarani\",\"doi\":\"10.1162/imag_a_00148\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Abstract Listeners with hearing loss have trouble following a conversation in multitalker environments. While modern hearing aids can generally amplify speech, these devices are unable to tune into a target speaker without first knowing to which speaker a user aims to attend. Brain-controlled hearing aids have been proposed using auditory attention decoding (AAD) methods, but current methods use the same model to compare the speech stimulus and neural response, regardless of the dynamic overlap between talkers which is known to influence neural encoding. Here, we propose a novel framework that directly classifies event-related potentials (ERPs) evoked by glimpsed and masked acoustic events to determine whether the source of the event was attended. We present a system that identifies auditory events using the local maxima in the envelope rate of change, assesses the temporal masking of auditory events relative to competing speakers, and utilizes masking-specific ERP classifiers to determine if the source of the event was attended. Using intracranial electrophysiological recordings, we showed that high gamma ERPs from recording sites in auditory cortex can effectively decode the attention of subjects. This method of AAD provides higher accuracy, shorter switch times, and more stable decoding results compared with traditional correlational methods, permitting the quick and accurate detection of changes in a listener’s attentional focus. This framework also holds unique potential for detecting instances of divided attention and inattention. Overall, we extend the scope of AAD algorithms by introducing the first linear, direct-classification method for determining a listener’s attentional focus that leverages the latest research in multitalker speech perception. This work represents another step toward informing the development of effective and intuitive brain-controlled hearing assistive devices.\",\"PeriodicalId\":507939,\"journal\":{\"name\":\"Imaging Neuroscience\",\"volume\":\"339 4\",\"pages\":\"1-19\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Imaging Neuroscience\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1162/imag_a_00148\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Imaging Neuroscience","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1162/imag_a_00148","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

摘要 有听力损失的听众在多人交谈环境中很难跟上谈话内容。虽然现代助听器一般都能放大语音,但这些设备无法在不知道用户想要听哪位发言者讲话的情况下调整到目标发言者的位置。大脑控制助听器已被提出使用听觉注意力解码(AAD)方法,但目前的方法使用相同的模型来比较语音刺激和神经反应,而不考虑说话者之间的动态重叠,众所周知,这种重叠会影响神经编码。在这里,我们提出了一个新颖的框架,它能直接对瞥见的和掩蔽的声学事件诱发的事件相关电位(ERPs)进行分类,以确定事件的来源是否被关注。我们提出的系统能利用包络变化率的局部最大值识别听觉事件,评估听觉事件相对于竞争扬声器的时间掩蔽,并利用掩蔽特定的 ERP 分类器确定事件源是否被注意到。通过颅内电生理记录,我们发现来自听觉皮层记录点的高伽马ERP能有效解码受试者的注意力。与传统的相关方法相比,这种 AAD 方法具有更高的准确性、更短的切换时间和更稳定的解码结果,可以快速准确地检测出听者注意力焦点的变化。这一框架还具有检测注意力分散和不集中情况的独特潜力。总之,我们通过引入第一种线性、直接分类方法来确定听者的注意力焦点,并利用多语者语音感知方面的最新研究成果,扩展了 AAD 算法的范围。这项工作标志着我们向开发有效、直观的脑控听力辅助设备又迈进了一步。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Improving auditory attention decoding by classifying intracranial responses to glimpsed and masked acoustic events
Abstract Listeners with hearing loss have trouble following a conversation in multitalker environments. While modern hearing aids can generally amplify speech, these devices are unable to tune into a target speaker without first knowing to which speaker a user aims to attend. Brain-controlled hearing aids have been proposed using auditory attention decoding (AAD) methods, but current methods use the same model to compare the speech stimulus and neural response, regardless of the dynamic overlap between talkers which is known to influence neural encoding. Here, we propose a novel framework that directly classifies event-related potentials (ERPs) evoked by glimpsed and masked acoustic events to determine whether the source of the event was attended. We present a system that identifies auditory events using the local maxima in the envelope rate of change, assesses the temporal masking of auditory events relative to competing speakers, and utilizes masking-specific ERP classifiers to determine if the source of the event was attended. Using intracranial electrophysiological recordings, we showed that high gamma ERPs from recording sites in auditory cortex can effectively decode the attention of subjects. This method of AAD provides higher accuracy, shorter switch times, and more stable decoding results compared with traditional correlational methods, permitting the quick and accurate detection of changes in a listener’s attentional focus. This framework also holds unique potential for detecting instances of divided attention and inattention. Overall, we extend the scope of AAD algorithms by introducing the first linear, direct-classification method for determining a listener’s attentional focus that leverages the latest research in multitalker speech perception. This work represents another step toward informing the development of effective and intuitive brain-controlled hearing assistive devices.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Optimization and validation of multi-echo, multi-contrast SAGE acquisition in fMRI BOLD fMRI responses to amplitude-modulated sounds across age in adult listeners Developmental trajectories of the default mode, frontoparietal, and salience networks from the third trimester through the newborn period GABA levels decline with age: A longitudinal study Unveiling hidden sources of dynamic functional connectome through a novel regularized blind source separation approach
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1