Two measures of auditory spatial resolution, the minimum audible angle and the minimum audible movement angle, have been obtained in a simulated acoustic environment using Ambisonics sound field reproduction. Trajectories were designed to provide no reliable cues for the spatial discrimination task. Larger threshold angles were found in reverberant compared to anechoic conditions, for stimuli on the side compared to the front, and for moving compared to static stimuli. The effect of reverberation appeared to be independent of the position of the sound source (same relative threshold increase) and was independently present for static and moving sound sources.
{"title":"Static and moving minimum audible angle: Independent contributions of reverberation and position.","authors":"Anna Dietze, Samuel W Clapp, Bernhard U Seeber","doi":"10.1121/10.0025992","DOIUrl":"https://doi.org/10.1121/10.0025992","url":null,"abstract":"<p><p>Two measures of auditory spatial resolution, the minimum audible angle and the minimum audible movement angle, have been obtained in a simulated acoustic environment using Ambisonics sound field reproduction. Trajectories were designed to provide no reliable cues for the spatial discrimination task. Larger threshold angles were found in reverberant compared to anechoic conditions, for stimuli on the side compared to the front, and for moving compared to static stimuli. The effect of reverberation appeared to be independent of the position of the sound source (same relative threshold increase) and was independently present for static and moving sound sources.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140923846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.
{"title":"Neural correlates of tonal loudness, intensity discrimination, and duration discrimination.","authors":"Shigeyuki Kuwada, Constantine Trahiotis","doi":"10.1121/10.0025874","DOIUrl":"https://doi.org/10.1121/10.0025874","url":null,"abstract":"<p><p>A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Predicting acoustic transmission loss in the SOFAR channel faces challenges, such as excessively complex algorithms and computationally intensive calculations in classical methods. To address these challenges, a deep learning-based underwater acoustic transmission loss prediction method is proposed. By properly training a U-net-type convolutional neural network, the method can provide an accurate mapping between ray trajectories and the transmission loss over the problem domain. Verifications are performed in a SOFAR channel with Munk's sound speed profile. The results suggest that the method has potential to be used as a fast predicting model without sacrificing accuracy.
预测 SOFAR 信道中的声波传输损耗面临着各种挑战,例如传统方法中过于复杂的算法和计算密集型计算。为了应对这些挑战,本文提出了一种基于深度学习的水下声波传输损耗预测方法。通过适当训练 U 网型卷积神经网络,该方法可提供射线轨迹与问题域传输损耗之间的精确映射。在具有 Munk 声速剖面的 SOFAR 信道中进行了验证。结果表明,该方法具有作为快速预测模型的潜力,同时不会牺牲精度。
{"title":"Predicting underwater acoustic transmission loss in the SOFAR channel from ray trajectories via deep learning.","authors":"Haitao Wang, Shiwei Peng, Qunyi He, Xiangyang Zeng","doi":"10.1121/10.0025976","DOIUrl":"https://doi.org/10.1121/10.0025976","url":null,"abstract":"<p><p>Predicting acoustic transmission loss in the SOFAR channel faces challenges, such as excessively complex algorithms and computationally intensive calculations in classical methods. To address these challenges, a deep learning-based underwater acoustic transmission loss prediction method is proposed. By properly training a U-net-type convolutional neural network, the method can provide an accurate mapping between ray trajectories and the transmission loss over the problem domain. Verifications are performed in a SOFAR channel with Munk's sound speed profile. The results suggest that the method has potential to be used as a fast predicting model without sacrificing accuracy.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima
The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.
{"title":"Impact of speech rate on perception of vowel and consonant duration by bilinguals and monolinguals.","authors":"Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima","doi":"10.1121/10.0025862","DOIUrl":"https://doi.org/10.1121/10.0025862","url":null,"abstract":"<p><p>The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya
Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.
{"title":"Reduced digit spans and ear dominance using dichotic digits in bimodal cochlear-implant users.","authors":"Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya","doi":"10.1121/10.0025977","DOIUrl":"https://doi.org/10.1121/10.0025977","url":null,"abstract":"<p><p>Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140899058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Machine learning enabled auscultating diagnosis can provide promising solutions especially for prescreening purposes. The bottleneck for its potential success is that high-quality datasets for training are still scarce. An open auscultation dataset that consists of samples and annotations from patients and healthy individuals is established in this work for the respiratory diagnosis studies with machine learning, which is of both scientific importance and practical potential. A machine learning approach is examined to showcase the use of this new dataset for lung sound classifications with different diseases. The open dataset is available to the public online.
{"title":"An open auscultation dataset for machine learning-based respiratory diagnosis studies.","authors":"Guanyu Zhou, Chengjian Liu, Xiaoguang Li, Sicong Liang, Ruichen Wang, Xun Huang","doi":"10.1121/10.0025851","DOIUrl":"https://doi.org/10.1121/10.0025851","url":null,"abstract":"<p><p>Machine learning enabled auscultating diagnosis can provide promising solutions especially for prescreening purposes. The bottleneck for its potential success is that high-quality datasets for training are still scarce. An open auscultation dataset that consists of samples and annotations from patients and healthy individuals is established in this work for the respiratory diagnosis studies with machine learning, which is of both scientific importance and practical potential. A machine learning approach is examined to showcase the use of this new dataset for lung sound classifications with different diseases. The open dataset is available to the public online.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Adding to limited research on clear speech in tone languages, productions of Mandarin lexical tones were examined in pentasyllabic sentences. Fourteen participants read sentences imagining a hard-of-hearing addressee or a friend in a casual social setting. Tones produced in clear speech had longer duration, higher intensity, and larger F0 values. This style effect was rarely modulated by tone, preceding tonal context, or syllable position, consistent with an overall signal enhancement strategy. Possible evidence for tone enhancement was observed only in one set of analysis for F0 minimum and F0 range, contrasting tones with low targets and tones with high targets.
{"title":"Clear speech effects in production of sentence-medial Mandarin lexical tonesa).","authors":"Jack Rittenberry, Irina A Shport","doi":"10.1121/10.0025991","DOIUrl":"https://doi.org/10.1121/10.0025991","url":null,"abstract":"<p><p>Adding to limited research on clear speech in tone languages, productions of Mandarin lexical tones were examined in pentasyllabic sentences. Fourteen participants read sentences imagining a hard-of-hearing addressee or a friend in a casual social setting. Tones produced in clear speech had longer duration, higher intensity, and larger F0 values. This style effect was rarely modulated by tone, preceding tonal context, or syllable position, consistent with an overall signal enhancement strategy. Possible evidence for tone enhancement was observed only in one set of analysis for F0 minimum and F0 range, contrasting tones with low targets and tones with high targets.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141158923","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Amara C Ezenwa, Matthew J Goupell, Sandra Gordon-Salant
This study evaluated whether adaptive training with time-compressed speech produces an age-dependent improvement in speech recognition in 14 adult cochlear-implant users. The protocol consisted of a pretest, 5 h of training, and a posttest using time-compressed speech and an adaptive procedure. There were significant improvements in time-compressed speech recognition at the posttest session following training (>5% in the average time-compressed speech recognition threshold) but no effects of age. These results are promising for the use of adaptive training in aural rehabilitation strategies for cochlear-implant users across the adult lifespan and possibly using speech signals, such as time-compressed speech, to train temporal processing.
{"title":"Cochlear-implant listeners benefit from training with time-compressed speech, even at advanced ages.","authors":"Amara C Ezenwa, Matthew J Goupell, Sandra Gordon-Salant","doi":"10.1121/10.0025431","DOIUrl":"10.1121/10.0025431","url":null,"abstract":"<p><p>This study evaluated whether adaptive training with time-compressed speech produces an age-dependent improvement in speech recognition in 14 adult cochlear-implant users. The protocol consisted of a pretest, 5 h of training, and a posttest using time-compressed speech and an adaptive procedure. There were significant improvements in time-compressed speech recognition at the posttest session following training (>5% in the average time-compressed speech recognition threshold) but no effects of age. These results are promising for the use of adaptive training in aural rehabilitation strategies for cochlear-implant users across the adult lifespan and possibly using speech signals, such as time-compressed speech, to train temporal processing.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11075136/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878103","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The present study examined English vowel recognition in multi-talker babbles (MTBs) in 20 normal-hearing, native-English-speaking adult listeners. Twelve vowels, embedded in the h-V-d structure, were presented in MTBs consisting of 1, 2, 4, 6, 8, 10, and 12 talkers (numbers of talkers [N]) and a speech-shaped noise at signal-to-noise ratios of -12, -6, and 0 dB. Results showed that vowel recognition performance was a non-monotonic function of N when signal-to-noise ratios were less favorable. The masking effects of MTBs on vowel recognition were most similar to consonant recognition but less so to word and sentence recognition reported in previous studies.
{"title":"English vowel recognition in multi-talker babbles mixed with different numbers of talkersa).","authors":"Xianhui Wang, Li Xu","doi":"10.1121/10.0025616","DOIUrl":"https://doi.org/10.1121/10.0025616","url":null,"abstract":"The present study examined English vowel recognition in multi-talker babbles (MTBs) in 20 normal-hearing, native-English-speaking adult listeners. Twelve vowels, embedded in the h-V-d structure, were presented in MTBs consisting of 1, 2, 4, 6, 8, 10, and 12 talkers (numbers of talkers [N]) and a speech-shaped noise at signal-to-noise ratios of -12, -6, and 0 dB. Results showed that vowel recognition performance was a non-monotonic function of N when signal-to-noise ratios were less favorable. The masking effects of MTBs on vowel recognition were most similar to consonant recognition but less so to word and sentence recognition reported in previous studies.","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140788246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A noise-insensitive cost function was developed for estimating the speed of harmonic acoustic sources in uniform linear motion. This function weighs and integrates the energy distribution of received tones in the time-frequency plane to enhance the robustness of parameter estimation under low signal-to-noise ratio conditions, where weight values are intentionally combined with the law of observed instantaneous frequency. As the cost function is differentiable, the procedure of parameter estimations also has high computing efficiency. Processing data of SWellEx-96 experiments with real ocean noise confirmed the anti-noise capabilities of this cost function to conventional processing methods.
{"title":"Robust speed estimation for a moving harmonic acoustic source with a single stationary sensor.","authors":"Yixin Yang, Ningning Liang, Jianbo Zhou","doi":"10.1121/10.0025508","DOIUrl":"https://doi.org/10.1121/10.0025508","url":null,"abstract":"A noise-insensitive cost function was developed for estimating the speed of harmonic acoustic sources in uniform linear motion. This function weighs and integrates the energy distribution of received tones in the time-frequency plane to enhance the robustness of parameter estimation under low signal-to-noise ratio conditions, where weight values are intentionally combined with the law of observed instantaneous frequency. As the cost function is differentiable, the procedure of parameter estimations also has high computing efficiency. Processing data of SWellEx-96 experiments with real ocean noise confirmed the anti-noise capabilities of this cost function to conventional processing methods.","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140791512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}