Large surface wave breaking events in deep water are acoustically detectable by beamforming at 5-6 kHz with a mid-frequency planar array located 130 m below the surface. Due to the array's depth and modest 1 m horizontal aperture, wave breaking events cannot be tracked accurately by beamforming alone. Their trajectories are estimated instead by splitting the array into sub-arrays, beamforming each sub-array toward the source, and computing the temporal cross-correlation of the sub-array beams. Source tracks estimated from sub-array cross-correlations match the trajectories of breaking waves that are visible in aerial images of the ocean surface above the array.
{"title":"Mid-frequency acoustic tracking of breaking waves.","authors":"Ryan Saenger, Luc Lenain, William S Hodgkiss","doi":"10.1121/10.0026149","DOIUrl":"https://doi.org/10.1121/10.0026149","url":null,"abstract":"<p><p>Large surface wave breaking events in deep water are acoustically detectable by beamforming at 5-6 kHz with a mid-frequency planar array located 130 m below the surface. Due to the array's depth and modest 1 m horizontal aperture, wave breaking events cannot be tracked accurately by beamforming alone. Their trajectories are estimated instead by splitting the array into sub-arrays, beamforming each sub-array toward the source, and computing the temporal cross-correlation of the sub-array beams. Source tracks estimated from sub-array cross-correlations match the trajectories of breaking waves that are visible in aerial images of the ocean surface above the array.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141201678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Brian S Miller, Cara Masere, Mark Milnes, Jaimie Cleeland, Timothy Lamb, Dale Maschette, Dirk Welsford
An underwater acoustic recorder was moored off Heard Island from September 2017 through March 2018 to listen for marine mammals. Analysis of data was initially conducted by visual inspection of long-term spectral averages to reveal sounds from sperm whales, Antarctic and pygmy blue whales, fin whales, minke whales, odontocete whistles, and noise from nearby ships. Automated detection of sperm whale clicks revealed they were seldom detected from September through January (n = 35 h) but were detected nearly every day of February and March (n = 684 h). Additional analysis of these detections revealed further diel and demographic patterns.
{"title":"Heard off Heard: Acoustic detections of sperm whales (Physeter macrocephalus) and other cetaceans off Heard Island.","authors":"Brian S Miller, Cara Masere, Mark Milnes, Jaimie Cleeland, Timothy Lamb, Dale Maschette, Dirk Welsford","doi":"10.1121/10.0026242","DOIUrl":"10.1121/10.0026242","url":null,"abstract":"<p><p>An underwater acoustic recorder was moored off Heard Island from September 2017 through March 2018 to listen for marine mammals. Analysis of data was initially conducted by visual inspection of long-term spectral averages to reveal sounds from sperm whales, Antarctic and pygmy blue whales, fin whales, minke whales, odontocete whistles, and noise from nearby ships. Automated detection of sperm whale clicks revealed they were seldom detected from September through January (n = 35 h) but were detected nearly every day of February and March (n = 684 h). Additional analysis of these detections revealed further diel and demographic patterns.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 6","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141452336","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Xueli Sheng, Dewen Li, Ran Cao, Xuan Zhou, Jiarui Yin
Passive detection of target-of-interest (TOI) within strong interferences poses a challenge. This paper introduces an adaptive interference suppression based on an invariant subspace of matrix matching. Assume that the TOI-bearing intervals are known. We define a correlation ratio for each eigenvector to obtain the highest one. Then, we use invariant subspace of matrix matching to measure the distance between two invariant projection matrices of this eigenvector. This identifies and removes the eigenvectors associated with TOI. Finally, the remaining eigenvectors are subtracted from the sample covariance matrix to suppress interference and noise. The viability of the proposed method is demonstrated experimentally.
在强干扰下被动探测感兴趣目标(TOI)是一项挑战。本文介绍了一种基于矩阵匹配不变子空间的自适应干扰抑制方法。假设含有兴趣目标的区间是已知的。我们为每个特征向量定义一个相关比,以获得最高的相关比。然后,我们使用矩阵匹配的不变子空间来测量该特征向量的两个不变投影矩阵之间的距离。这样就能识别并移除与 TOI 相关的特征向量。最后,从样本协方差矩阵中减去剩余的特征向量,以抑制干扰和噪声。实验证明了所提方法的可行性。
{"title":"Adaptive interference suppression based on an invariant subspace of matrices matching for a horizontal array in underwater acoustics.","authors":"Xueli Sheng, Dewen Li, Ran Cao, Xuan Zhou, Jiarui Yin","doi":"10.1121/10.0026373","DOIUrl":"https://doi.org/10.1121/10.0026373","url":null,"abstract":"<p><p>Passive detection of target-of-interest (TOI) within strong interferences poses a challenge. This paper introduces an adaptive interference suppression based on an invariant subspace of matrix matching. Assume that the TOI-bearing intervals are known. We define a correlation ratio for each eigenvector to obtain the highest one. Then, we use invariant subspace of matrix matching to measure the distance between two invariant projection matrices of this eigenvector. This identifies and removes the eigenvectors associated with TOI. Finally, the remaining eigenvectors are subtracted from the sample covariance matrix to suppress interference and noise. The viability of the proposed method is demonstrated experimentally.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141312447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This Letter proposes a low-complexity joint equalization and decoding reception scheme based on super-trellis per-survivor processing, making it possible to apply maximum likelihood sequence estimation in high-order underwater acoustic communications under fast time-varying channels. The technique combines trellis-coded modulation states and intersymbol interference states and uses per-survivor processing to track channel parameters. Furthermore, a general trellis configuration for arbitrary order quadrature amplitude modulation signal is provided when truncate the channel is used to describe the intersymbol interference state to 1. Sea trials results show that the performance of proposed method can be more than 1.4 dB superiority than conventional schemes.
本信提出了一种基于超trellis每存活者处理的低复杂度联合均衡和解码接收方案,使得在快速时变信道下的高阶水下声学通信中应用最大似然序列估计成为可能。该技术结合了trellis编码调制状态和符号间干扰状态,并使用每存活者处理来跟踪信道参数。海试结果表明,所提方法的性能比传统方案优越 1.4 dB 以上。
{"title":"Joint equalization and decoding with per-survivor processing based on super-trellis in time-varying underwater acoustic channels.","authors":"Xu Kou, Yanbo Wu, Min Zhu","doi":"10.1121/10.0026372","DOIUrl":"https://doi.org/10.1121/10.0026372","url":null,"abstract":"<p><p>This Letter proposes a low-complexity joint equalization and decoding reception scheme based on super-trellis per-survivor processing, making it possible to apply maximum likelihood sequence estimation in high-order underwater acoustic communications under fast time-varying channels. The technique combines trellis-coded modulation states and intersymbol interference states and uses per-survivor processing to track channel parameters. Furthermore, a general trellis configuration for arbitrary order quadrature amplitude modulation signal is provided when truncate the channel is used to describe the intersymbol interference state to 1. Sea trials results show that the performance of proposed method can be more than 1.4 dB superiority than conventional schemes.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 6","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141319167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.
{"title":"Neural correlates of tonal loudness, intensity discrimination, and duration discrimination.","authors":"Shigeyuki Kuwada, Constantine Trahiotis","doi":"10.1121/10.0025874","DOIUrl":"10.1121/10.0025874","url":null,"abstract":"<p><p>A long-standing quest in audition concerns understanding relations between behavioral measures and neural representations of changes in sound intensity. Here, we examined relations between aspects of intensity perception and central neural responses within the inferior colliculus of unanesthetized rabbits (by averaging the population's spike count/level functions). We found parallels between the population's neural output and: (1) how loudness grows with intensity; (2) how loudness grows with duration; (3) how discrimination of intensity improves with increasing sound level; (4) findings that intensity discrimination does not depend on duration; and (5) findings that duration discrimination is a constant fraction of base duration.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Two measures of auditory spatial resolution, the minimum audible angle and the minimum audible movement angle, have been obtained in a simulated acoustic environment using Ambisonics sound field reproduction. Trajectories were designed to provide no reliable cues for the spatial discrimination task. Larger threshold angles were found in reverberant compared to anechoic conditions, for stimuli on the side compared to the front, and for moving compared to static stimuli. The effect of reverberation appeared to be independent of the position of the sound source (same relative threshold increase) and was independently present for static and moving sound sources.
{"title":"Static and moving minimum audible angle: Independent contributions of reverberation and position.","authors":"Anna Dietze, Samuel W Clapp, Bernhard U Seeber","doi":"10.1121/10.0025992","DOIUrl":"https://doi.org/10.1121/10.0025992","url":null,"abstract":"<p><p>Two measures of auditory spatial resolution, the minimum audible angle and the minimum audible movement angle, have been obtained in a simulated acoustic environment using Ambisonics sound field reproduction. Trajectories were designed to provide no reliable cues for the spatial discrimination task. Larger threshold angles were found in reverberant compared to anechoic conditions, for stimuli on the side compared to the front, and for moving compared to static stimuli. The effect of reverberation appeared to be independent of the position of the sound source (same relative threshold increase) and was independently present for static and moving sound sources.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140923846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima
The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.
{"title":"Impact of speech rate on perception of vowel and consonant duration by bilinguals and monolinguals.","authors":"Miwako Hisagi, Eve Higby, Mike Zandona, Annett P Acosta, Justin Kent, Keiichi Tajima","doi":"10.1121/10.0025862","DOIUrl":"10.1121/10.0025862","url":null,"abstract":"<p><p>The perceptual boundary between short and long categories depends on speech rate. We investigated the influence of speech rate on perceptual boundaries for short and long vowel and consonant contrasts by Spanish-English bilingual listeners and English monolinguals. Listeners tended to adapt their perceptual boundaries to speech rates, but the strategy differed between groups, especially for consonants. Understanding the factors that influence auditory processing in this population is essential for developing appropriate assessments of auditory comprehension. These findings have implications for the clinical care of older populations whose ability to rely on spectral and/or temporal information in the auditory signal may decline.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Predicting acoustic transmission loss in the SOFAR channel faces challenges, such as excessively complex algorithms and computationally intensive calculations in classical methods. To address these challenges, a deep learning-based underwater acoustic transmission loss prediction method is proposed. By properly training a U-net-type convolutional neural network, the method can provide an accurate mapping between ray trajectories and the transmission loss over the problem domain. Verifications are performed in a SOFAR channel with Munk's sound speed profile. The results suggest that the method has potential to be used as a fast predicting model without sacrificing accuracy.
预测 SOFAR 信道中的声波传输损耗面临着各种挑战,例如传统方法中过于复杂的算法和计算密集型计算。为了应对这些挑战,本文提出了一种基于深度学习的水下声波传输损耗预测方法。通过适当训练 U 网型卷积神经网络,该方法可提供射线轨迹与问题域传输损耗之间的精确映射。在具有 Munk 声速剖面的 SOFAR 信道中进行了验证。结果表明,该方法具有作为快速预测模型的潜力,同时不会牺牲精度。
{"title":"Predicting underwater acoustic transmission loss in the SOFAR channel from ray trajectories via deep learning.","authors":"Haitao Wang, Shiwei Peng, Qunyi He, Xiangyang Zeng","doi":"10.1121/10.0025976","DOIUrl":"https://doi.org/10.1121/10.0025976","url":null,"abstract":"<p><p>Predicting acoustic transmission loss in the SOFAR channel faces challenges, such as excessively complex algorithms and computationally intensive calculations in classical methods. To address these challenges, a deep learning-based underwater acoustic transmission loss prediction method is proposed. By properly training a U-net-type convolutional neural network, the method can provide an accurate mapping between ray trajectories and the transmission loss over the problem domain. Verifications are performed in a SOFAR channel with Munk's sound speed profile. The results suggest that the method has potential to be used as a fast predicting model without sacrificing accuracy.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya
Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.
{"title":"Reduced digit spans and ear dominance using dichotic digits in bimodal cochlear-implant users.","authors":"Allison Blackmon, Matthew J Goupell, Matthew Bakke, Olga Stakhovskaya","doi":"10.1121/10.0025977","DOIUrl":"10.1121/10.0025977","url":null,"abstract":"<p><p>Bimodal stimulation, a cochlear implant (CI) in one ear and a hearing aid (HA) in the other, provides highly asymmetrical inputs. To understand how asymmetry affects perception and memory, forward and backward digit spans were measured in nine bimodal listeners. Spans were unchanged from monotic to diotic presentation; there was an average two-digit decrease for dichotic presentation with some extreme cases of decreases to zero spans. Interaurally asymmetrical decreases were not predicted based on the device or better-functioning ear. Therefore, bimodal listeners can demonstrate a strong ear dominance, diminishing memory recall dichotically even when perception was intact monaurally.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11550484/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140899058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Machine learning enabled auscultating diagnosis can provide promising solutions especially for prescreening purposes. The bottleneck for its potential success is that high-quality datasets for training are still scarce. An open auscultation dataset that consists of samples and annotations from patients and healthy individuals is established in this work for the respiratory diagnosis studies with machine learning, which is of both scientific importance and practical potential. A machine learning approach is examined to showcase the use of this new dataset for lung sound classifications with different diseases. The open dataset is available to the public online.
{"title":"An open auscultation dataset for machine learning-based respiratory diagnosis studies.","authors":"Guanyu Zhou, Chengjian Liu, Xiaoguang Li, Sicong Liang, Ruichen Wang, Xun Huang","doi":"10.1121/10.0025851","DOIUrl":"10.1121/10.0025851","url":null,"abstract":"<p><p>Machine learning enabled auscultating diagnosis can provide promising solutions especially for prescreening purposes. The bottleneck for its potential success is that high-quality datasets for training are still scarce. An open auscultation dataset that consists of samples and annotations from patients and healthy individuals is established in this work for the respiratory diagnosis studies with machine learning, which is of both scientific importance and practical potential. A machine learning approach is examined to showcase the use of this new dataset for lung sound classifications with different diseases. The open dataset is available to the public online.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"4 5","pages":""},"PeriodicalIF":1.2,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878102","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}