首页 > 最新文献

Journal of the Acoustical Society of America最新文献

英文 中文
The temporal effects of auditory and visual immersion on speech level in virtual environments. 虚拟环境中听觉和视觉沉浸对言语水平的时间效应。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042240
Xinyi N Zhang, Arian Shamei, Florian Grond, Ingrid Verduyckt, Rachel E Bouserhal

Speech takes place in physical environments with visual and acoustic properties, yet how these elements and their interaction influence speech production is not fully understood. While a room's appearance can suggest its acoustics, it is unclear whether people adjust their speech based on this visual information. Previous research shows that higher reverberation leads to reduced speech level, but how auditory and visual information interact in this process remains limited. This study examined how audiovisual information affects speech level by immersing participants in virtual environments with varying reverberation and room visuals (hemi-anechoic room, classroom, and gymnasium) while completing speech tasks. Speech level was analyzed using generalized additive mixed-effects modeling to assess temporal changes during utterances across conditions. Results showed that visual information significantly influenced speech level, though not strictly in line with expected acoustics or perceived room size; auditory information had a stronger overall effect than visual information. Visual information had an earlier influence that diminished over time, whereas the auditory effect increased and plateaued. These findings contribute to the understanding of multisensory integration in speech control and have implications in enhancing vocal performance and supporting more naturalistic communication in virtual environments.

语音发生在具有视觉和声学特性的物理环境中,但这些元素及其相互作用如何影响语音产生尚不完全清楚。虽然一个房间的外观可以表明它的声学效果,但人们是否会根据这种视觉信息来调整自己的语言还不清楚。先前的研究表明,较高的混响会导致语音水平降低,但听觉和视觉信息在这一过程中如何相互作用仍然有限。本研究通过将参与者沉浸在具有不同混响和房间视觉效果(半消声室、教室和体育馆)的虚拟环境中,同时完成演讲任务,研究了视听信息如何影响语音水平。使用广义加性混合效应模型分析语音水平,以评估不同条件下话语的时间变化。结果表明,视觉信息显著影响语音水平,尽管与预期的声学或感知的房间大小不完全一致;听觉信息的整体效果强于视觉信息。视觉信息的早期影响随着时间的推移而减弱,而听觉的影响则增加并趋于稳定。这些发现有助于理解语音控制中的多感觉整合,并对提高语音表现和支持虚拟环境中更自然的交流具有重要意义。
{"title":"The temporal effects of auditory and visual immersion on speech level in virtual environments.","authors":"Xinyi N Zhang, Arian Shamei, Florian Grond, Ingrid Verduyckt, Rachel E Bouserhal","doi":"10.1121/10.0042240","DOIUrl":"https://doi.org/10.1121/10.0042240","url":null,"abstract":"<p><p>Speech takes place in physical environments with visual and acoustic properties, yet how these elements and their interaction influence speech production is not fully understood. While a room's appearance can suggest its acoustics, it is unclear whether people adjust their speech based on this visual information. Previous research shows that higher reverberation leads to reduced speech level, but how auditory and visual information interact in this process remains limited. This study examined how audiovisual information affects speech level by immersing participants in virtual environments with varying reverberation and room visuals (hemi-anechoic room, classroom, and gymnasium) while completing speech tasks. Speech level was analyzed using generalized additive mixed-effects modeling to assess temporal changes during utterances across conditions. Results showed that visual information significantly influenced speech level, though not strictly in line with expected acoustics or perceived room size; auditory information had a stronger overall effect than visual information. Visual information had an earlier influence that diminished over time, whereas the auditory effect increased and plateaued. These findings contribute to the understanding of multisensory integration in speech control and have implications in enhancing vocal performance and supporting more naturalistic communication in virtual environments.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"384-397"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PAMGuard: Application software for passive acoustic detection, classification, and localisation of animal sounds. PAMGuard:用于被动声学检测、分类和动物声音定位的应用软件。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042245
Douglas Gillespie, Jamie Macaulay, Michael Oswald, Marie Roch

Detection, classification, and localisation of animal sounds are essential in many ecological studies, including density estimation and behavioural studies. Real-time acoustic processing can also be used in mitigation exercises, with the possibility of curtailing harmful human activities when animals are present. Animal vocalisations vary widely, and there is no single detection algorithm that can robustly detect all sound types. Human-in-the loop analysis is often required to validate algorithm performance and deal with unexpected noise sources such as are often encountered in real-world situations. The PAMGuard software combines advanced automatic analysis algorithms, including AI methods, with interactive visual tools allowing users to develop efficient workflows for both real-time use and for processing archived datasets. A modular framework enables users to configure multiple detectors, classifiers, and localisers suitable for the equipment and species of interest in a particular application. Multiple detectors for different sound types can be run concurrently on the same data. An extensible "plug-in" interface also makes it possible for third parties to independently develop new modules to run within the software framework. Here, we describe the software's core functionality, illustrated using workflows for both real-time and offline use, and present an update on the latest features.

动物声音的检测、分类和定位在许多生态学研究中是必不可少的,包括密度估计和行为研究。实时声学处理也可用于缓解活动,有可能在有动物在场时减少有害的人类活动。动物的声音变化很大,没有单一的检测算法可以检测所有的声音类型。在验证算法性能和处理意外噪声源(如在现实世界中经常遇到的噪声源)时,通常需要人在循环分析。PAMGuard软件结合了先进的自动分析算法,包括人工智能方法,以及交互式可视化工具,允许用户开发实时使用和处理存档数据集的高效工作流程。模块化框架使用户能够配置多个检测器、分类器和本地化器,适合特定应用程序中感兴趣的设备和物种。不同声音类型的多个检测器可以在同一数据上并发运行。可扩展的“插件”接口也使第三方能够独立开发在软件框架内运行的新模块。在这里,我们描述了软件的核心功能,使用实时和离线使用的工作流进行说明,并介绍了最新功能的更新。
{"title":"PAMGuard: Application software for passive acoustic detection, classification, and localisation of animal sounds.","authors":"Douglas Gillespie, Jamie Macaulay, Michael Oswald, Marie Roch","doi":"10.1121/10.0042245","DOIUrl":"https://doi.org/10.1121/10.0042245","url":null,"abstract":"<p><p>Detection, classification, and localisation of animal sounds are essential in many ecological studies, including density estimation and behavioural studies. Real-time acoustic processing can also be used in mitigation exercises, with the possibility of curtailing harmful human activities when animals are present. Animal vocalisations vary widely, and there is no single detection algorithm that can robustly detect all sound types. Human-in-the loop analysis is often required to validate algorithm performance and deal with unexpected noise sources such as are often encountered in real-world situations. The PAMGuard software combines advanced automatic analysis algorithms, including AI methods, with interactive visual tools allowing users to develop efficient workflows for both real-time use and for processing archived datasets. A modular framework enables users to configure multiple detectors, classifiers, and localisers suitable for the equipment and species of interest in a particular application. Multiple detectors for different sound types can be run concurrently on the same data. An extensible \"plug-in\" interface also makes it possible for third parties to independently develop new modules to run within the software framework. Here, we describe the software's core functionality, illustrated using workflows for both real-time and offline use, and present an update on the latest features.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"437-443"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145985116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The study on the design and performance analysis of acoustic metamaterial lens. 声学超材料透镜的设计与性能分析研究。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042194
Hua-Wei Ji, Li-Ming Lin, Jiang-Hai Wang, Di-Wei Xiong, Chong-Jin Du

Acoustic lens focusing is a commonly used method in high-intensity focused ultrasound (HIFU). However, traditional acoustic lens focusing suffers from low focusing efficiency and excessive sidelobes, which affect the efficacy and safety of HIFU treatment. To address this issue, this paper designs a periodic trapezoidal‑groove acoustic metasurface lens by leveraging the extraordinary acoustic transmission effect. Subsequently, its focal sound‑pressure level is calculated through theoretical analysis and finite‑element simulation, and is further validated experimentally. Finally, the influence of structural parameters-such as the period, center width, depth, and taper angle of the trapezoidal groove, as well as the amplitude of the excitation source-on the focusing performance of the acoustic metasurface lens is systematically analyzed. The results demonstrate that the periodic trapezoidal‑groove acoustic metasurface lens can further enhance focusing and suppress sidelobes within a specific frequency range; the frequency corresponding to the maximum sound pressure is determined by the period of the trapezoidal groove; and the shift of Wood's anomaly frequency is primarily governed by the groove depth. This study provides insights for the development of high‑performance acoustic‑lens-focused ultrasound transducers.

声透镜聚焦是高强度聚焦超声(HIFU)中常用的一种方法。然而,传统声透镜聚焦存在聚焦效率低、副瓣过多等问题,影响了HIFU治疗的有效性和安全性。为了解决这一问题,本文利用超常的声传输效应,设计了一种周期梯形槽声学超表面透镜。随后,通过理论分析和有限元模拟计算了其震源声压级,并进行了实验验证。最后,系统分析了梯形槽周期、中心宽度、深度、锥角以及激励源振幅等结构参数对声超表面透镜聚焦性能的影响。结果表明:周期梯形槽声学超表面透镜在一定频率范围内可以进一步增强聚焦,抑制副瓣;最大声压所对应的频率由梯形槽的周期决定;Wood异常频率的移动主要受槽深的控制。该研究为高性能声透镜聚焦超声换能器的开发提供了见解。
{"title":"The study on the design and performance analysis of acoustic metamaterial lens.","authors":"Hua-Wei Ji, Li-Ming Lin, Jiang-Hai Wang, Di-Wei Xiong, Chong-Jin Du","doi":"10.1121/10.0042194","DOIUrl":"https://doi.org/10.1121/10.0042194","url":null,"abstract":"<p><p>Acoustic lens focusing is a commonly used method in high-intensity focused ultrasound (HIFU). However, traditional acoustic lens focusing suffers from low focusing efficiency and excessive sidelobes, which affect the efficacy and safety of HIFU treatment. To address this issue, this paper designs a periodic trapezoidal‑groove acoustic metasurface lens by leveraging the extraordinary acoustic transmission effect. Subsequently, its focal sound‑pressure level is calculated through theoretical analysis and finite‑element simulation, and is further validated experimentally. Finally, the influence of structural parameters-such as the period, center width, depth, and taper angle of the trapezoidal groove, as well as the amplitude of the excitation source-on the focusing performance of the acoustic metasurface lens is systematically analyzed. The results demonstrate that the periodic trapezoidal‑groove acoustic metasurface lens can further enhance focusing and suppress sidelobes within a specific frequency range; the frequency corresponding to the maximum sound pressure is determined by the period of the trapezoidal groove; and the shift of Wood's anomaly frequency is primarily governed by the groove depth. This study provides insights for the development of high‑performance acoustic‑lens-focused ultrasound transducers.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"234-246"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145933942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparisons of air and bone conduction transfer properties utilizing stimulus frequency otoacoustic emissions. 利用刺激频率耳声发射的空气和骨传导转移特性的比较。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042219
Jie Wang, Zhuoran Shi, Shengjian Wu, Stefan Stenfelt, Jinqiu Sang, Xiaodong Li, Chengshi Zheng

Otoacoustic emissions represent cochlear responses to auditory stimuli, enabling the investigation of air conduction (AC) and bone conduction (BC) transmission. This study developed and validated a non-invasive, objective method for measuring the sensitivity difference between AC and BC transmission, here termed bone-air difference transfer property (BADTP), using stimulus frequency otoacoustic emission (SFOAE). The BADTP was defined as the difference between the AC transfer property and the BC transfer property. To cross-validate the objective approach, BADTP was compared with subjectively obtained hearing thresholds. Measurements were conducted across frequencies from 1000 to 4000 Hz in ten individuals with normal hearing. Results revealed that the mean differences between the two methods were within 2 dB at frequencies from 1000 to 1600 Hz, while both methods showed similar trends from 1850 to 4000 Hz. The proposed SFOAE-based method for measuring provides valuable insight into BC transmission, with potential applications for objective assessment of BC function in research settings.

耳声发射代表耳蜗对听觉刺激的反应,使研究空气传导(AC)和骨传导(BC)传输成为可能。本研究开发并验证了一种非侵入性、客观的方法,用于测量交流和BC传输之间的灵敏度差异,这里称为骨-气差异传输特性(BADTP),使用刺激频率耳声发射(SFOAE)。BADTP被定义为AC传输属性和BC传输属性之间的差值。为了交叉验证客观方法,将BADTP与主观获得的听力阈值进行比较。对10名听力正常的人进行了1000到4000赫兹的频率测量。结果表明,在1000 ~ 1600 Hz范围内,两种方法的平均差异在2 dB以内,而在1850 ~ 4000 Hz范围内,两种方法的趋势相似。提出的基于sfoae的测量方法提供了对BC传播的有价值的见解,具有在研究环境中客观评估BC功能的潜在应用。
{"title":"Comparisons of air and bone conduction transfer properties utilizing stimulus frequency otoacoustic emissions.","authors":"Jie Wang, Zhuoran Shi, Shengjian Wu, Stefan Stenfelt, Jinqiu Sang, Xiaodong Li, Chengshi Zheng","doi":"10.1121/10.0042219","DOIUrl":"https://doi.org/10.1121/10.0042219","url":null,"abstract":"<p><p>Otoacoustic emissions represent cochlear responses to auditory stimuli, enabling the investigation of air conduction (AC) and bone conduction (BC) transmission. This study developed and validated a non-invasive, objective method for measuring the sensitivity difference between AC and BC transmission, here termed bone-air difference transfer property (BADTP), using stimulus frequency otoacoustic emission (SFOAE). The BADTP was defined as the difference between the AC transfer property and the BC transfer property. To cross-validate the objective approach, BADTP was compared with subjectively obtained hearing thresholds. Measurements were conducted across frequencies from 1000 to 4000 Hz in ten individuals with normal hearing. Results revealed that the mean differences between the two methods were within 2 dB at frequencies from 1000 to 1600 Hz, while both methods showed similar trends from 1850 to 4000 Hz. The proposed SFOAE-based method for measuring provides valuable insight into BC transmission, with potential applications for objective assessment of BC function in research settings.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"315-326"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-stage representation learning for blind Room-Acoustic parameter estimation with uncertainty quantification. 基于多阶段表征学习的盲室声学参数估计。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042193
Philipp Götz, Cagdas Tuna, Andreas Brendel, Andreas Walther, Emanuël A P Habets

The ability to infer a general representation of the acoustic environment from a reverberant recording is a key objective in numerous applications. We propose a multi-stage approach that integrates task-agnostic representation learning with uncertainty quantification. Leveraging the conformal prediction framework, our method models the error incurred in the estimation of the acoustic environment embedded in a reverberant recording, which reflects the ambiguity inherent in distinguishing between an unknown source signal and the induced reverberation. Although our approach is flexible and agnostic to specific downstream objectives, experiments on real-world data demonstrate competitive performance on established parameter estimation tasks when compared to baselines trained end-to-end or with contrastive losses. Furthermore, a latent disentanglement analysis reveals the interpretability of the learned representations, which effectively capture distinct factors of variation within the acoustic environment.

从混响录音中推断声环境的一般表示的能力在许多应用中是一个关键目标。我们提出了一种将任务不可知表征学习与不确定性量化相结合的多阶段方法。利用保形预测框架,我们的方法模拟了混响记录中嵌入的声环境估计中产生的误差,这反映了在区分未知源信号和诱导混响时固有的模糊性。尽管我们的方法是灵活的,对特定的下游目标是不可知的,但在现实世界数据上的实验表明,与端到端训练基线或对比损失相比,在已建立的参数估计任务上具有竞争力。此外,潜在解纠缠分析揭示了学习表征的可解释性,它有效地捕获了声环境中不同的变化因素。
{"title":"Multi-stage representation learning for blind Room-Acoustic parameter estimation with uncertainty quantification.","authors":"Philipp Götz, Cagdas Tuna, Andreas Brendel, Andreas Walther, Emanuël A P Habets","doi":"10.1121/10.0042193","DOIUrl":"https://doi.org/10.1121/10.0042193","url":null,"abstract":"<p><p>The ability to infer a general representation of the acoustic environment from a reverberant recording is a key objective in numerous applications. We propose a multi-stage approach that integrates task-agnostic representation learning with uncertainty quantification. Leveraging the conformal prediction framework, our method models the error incurred in the estimation of the acoustic environment embedded in a reverberant recording, which reflects the ambiguity inherent in distinguishing between an unknown source signal and the induced reverberation. Although our approach is flexible and agnostic to specific downstream objectives, experiments on real-world data demonstrate competitive performance on established parameter estimation tasks when compared to baselines trained end-to-end or with contrastive losses. Furthermore, a latent disentanglement analysis reveals the interpretability of the learned representations, which effectively capture distinct factors of variation within the acoustic environment.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"247-259"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145952471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of eco-friendly spiral-shaped sound absorber made from handcrafted fibrous paper enhanced with spent coffee waste for broadband noise control. 用咖啡渣增强手工纤维纸制成的环保螺旋形吸声器,用于宽带噪声控制。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0041885
Jie Jin, Yunle Cao, Haipeng Hao, Yecheng Feng, Daitong Wei, Zhuqing Zhang

To address the non-degradability and toxicity of conventional acoustic materials, this study proposes a sustainable spiral-shaped sound absorber composed of plant fiber-based fibrous paper and recycled coffee waste (CW). The strong mechanical bonding between CW and Kozo fibrous paper in this composite acoustic material was observed using metallurgical microscopy, resulting in an environmentally friendly structure capable of controlling broadband noise. A prediction model based on parallel-slit theory was developed to evaluate the influence of key structural parameters-CW layer mass density, fibrous paper length, and absorber width-on sound absorption coefficients. Optimization reveals that wide spiral-shaped geometry paired with a high-density CW layer (0.04-0.05 kg/m2) enhances low-frequency noise reduction (<1000 Hz), whereas narrow configurations with a medium-density CW layer (0.03-0.04 kg/m2) improves high-frequency attenuation (>2000 Hz). The sound absorption coefficients of five prepared samples were measured using the two-microphone impedance tube method. The sound absorption coefficient showed significant improvement with the addition of an appropriate amount of CW in the mid- and high-frequency range. This work advances the development of lightweight, efficient, and sustainable acoustic solutions, providing a scalable strategy for the next generation of eco-friendly materials in line with circular economy principles and low-carbon manufacturing practices.

为了解决传统声学材料的不可降解性和毒性问题,本研究提出了一种可持续的螺旋形吸声器,该吸声器由植物纤维基纤维纸和再生咖啡废料(CW)组成。通过金相显微镜观察到这种复合声学材料中CW和Kozo纤维纸之间的强机械结合,从而形成了一种能够控制宽带噪声的环保结构。建立了基于平行缝理论的连续波层质量密度、纤维纸长度和吸声器宽度等关键结构参数对吸声系数的影响预测模型。优化结果表明,宽螺旋形几何结构与高密度连续波层(0.04-0.05 kg/m2)相匹配,可以增强低频降噪(2000 Hz)。采用双传声器阻抗管法测量了5种制备样品的吸声系数。在中频和高频范围内加入适量的连续波,吸声系数有显著提高。这项工作推动了轻质、高效和可持续声学解决方案的发展,为符合循环经济原则和低碳制造实践的下一代环保材料提供了可扩展的策略。
{"title":"Development of eco-friendly spiral-shaped sound absorber made from handcrafted fibrous paper enhanced with spent coffee waste for broadband noise control.","authors":"Jie Jin, Yunle Cao, Haipeng Hao, Yecheng Feng, Daitong Wei, Zhuqing Zhang","doi":"10.1121/10.0041885","DOIUrl":"https://doi.org/10.1121/10.0041885","url":null,"abstract":"<p><p>To address the non-degradability and toxicity of conventional acoustic materials, this study proposes a sustainable spiral-shaped sound absorber composed of plant fiber-based fibrous paper and recycled coffee waste (CW). The strong mechanical bonding between CW and Kozo fibrous paper in this composite acoustic material was observed using metallurgical microscopy, resulting in an environmentally friendly structure capable of controlling broadband noise. A prediction model based on parallel-slit theory was developed to evaluate the influence of key structural parameters-CW layer mass density, fibrous paper length, and absorber width-on sound absorption coefficients. Optimization reveals that wide spiral-shaped geometry paired with a high-density CW layer (0.04-0.05 kg/m2) enhances low-frequency noise reduction (<1000 Hz), whereas narrow configurations with a medium-density CW layer (0.03-0.04 kg/m2) improves high-frequency attenuation (>2000 Hz). The sound absorption coefficients of five prepared samples were measured using the two-microphone impedance tube method. The sound absorption coefficient showed significant improvement with the addition of an appropriate amount of CW in the mid- and high-frequency range. This work advances the development of lightweight, efficient, and sustainable acoustic solutions, providing a scalable strategy for the next generation of eco-friendly materials in line with circular economy principles and low-carbon manufacturing practices.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"260-271"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145952421","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Motion-aware sonar denoising for autonomous underwater vehicles self-noise using a speed-conditioned U-Net-transformer dual-branch conditional generative adversarial network. 基于速度条件u - net变压器双分支条件生成对抗网络的自主水下航行器运动感知声呐自噪声降噪研究。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042221
Yufei Wang, Yu Tian, Shilong Li, Jie Sun, Jiancheng Yu

Passive sonar surveillance by autonomous underwater vehicles (AUVs) is often hindered by non-stationary, nonlinear speed-dependent self-noise. To address this, we propose Speed-UT2-CGAN, a motion-aware sonar denoising framework utilizing a dual-branch conditional generative adversarial network that combines a U-Net convolutional branch for local feature extraction from time-domain audio sequences and a transformer-based attention branch for long-range temporal dependencies. The architecture incorporates AUV speed as an additional conditioning input to dynamically adapt to speed-dependent noise characteristics, and is trained with a combination of adversarial, time-domain, and frequency-domain loss functions to ensure accurate spectral and temporal reconstruction. Experiments on synthetic mixtures combining real AUV self-noise recordings from lake trials with ShipsEar vessel signals demonstrate that Speed-UT2-CGAN significantly outperforms traditional methods, speech enhancement generative adversarial network, and dual-path recurrent neural network, for a single AUV in shallow-water lake trials at 0, 2, and 3 knots, achieving an output average signal-to-noise ratio of 6.6 at -5 dB input and an average correlation coefficient of 0.87. These results confirm the effectiveness of motion-aware speed conditioning for passive sonar enhancement in single-sensor AUV systems, under controlled synthetic-data conditions representative of AUV constant depth, speed, and heading in shallow-water lake environments.

自主水下航行器(auv)的被动声呐监视经常受到非平稳、非线性速度相关自噪声的阻碍。为了解决这个问题,我们提出了Speed-UT2-CGAN,这是一个运动感知声纳去噪框架,利用双分支条件生成对抗网络,该网络结合了U-Net卷积分支,用于从时域音频序列中提取局部特征,以及基于变压器的注意分支,用于远程时间依赖性。该架构将水下航行器的速度作为额外的调节输入,以动态适应与速度相关的噪声特性,并结合对抗、时域和频域损失函数进行训练,以确保准确的频谱和时间重建。将湖泊试验中真实AUV自噪声记录与ShipsEar船舶信号相结合的合成混合实验表明,对于单个AUV在0、2和3节的浅水湖泊试验中,速度- ut2 - cgan显著优于传统方法、语音增强生成对抗网络和双路径递推神经网络,在-5 dB输入下,输出平均信噪比为6.6,平均相关系数为0.87。这些结果证实了运动感知速度调节对单传感器AUV系统被动声纳增强的有效性,在控制的合成数据条件下,代表浅水湖泊环境中AUV的恒定深度、速度和航向。
{"title":"Motion-aware sonar denoising for autonomous underwater vehicles self-noise using a speed-conditioned U-Net-transformer dual-branch conditional generative adversarial network.","authors":"Yufei Wang, Yu Tian, Shilong Li, Jie Sun, Jiancheng Yu","doi":"10.1121/10.0042221","DOIUrl":"https://doi.org/10.1121/10.0042221","url":null,"abstract":"<p><p>Passive sonar surveillance by autonomous underwater vehicles (AUVs) is often hindered by non-stationary, nonlinear speed-dependent self-noise. To address this, we propose Speed-UT2-CGAN, a motion-aware sonar denoising framework utilizing a dual-branch conditional generative adversarial network that combines a U-Net convolutional branch for local feature extraction from time-domain audio sequences and a transformer-based attention branch for long-range temporal dependencies. The architecture incorporates AUV speed as an additional conditioning input to dynamically adapt to speed-dependent noise characteristics, and is trained with a combination of adversarial, time-domain, and frequency-domain loss functions to ensure accurate spectral and temporal reconstruction. Experiments on synthetic mixtures combining real AUV self-noise recordings from lake trials with ShipsEar vessel signals demonstrate that Speed-UT2-CGAN significantly outperforms traditional methods, speech enhancement generative adversarial network, and dual-path recurrent neural network, for a single AUV in shallow-water lake trials at 0, 2, and 3 knots, achieving an output average signal-to-noise ratio of 6.6 at -5 dB input and an average correlation coefficient of 0.87. These results confirm the effectiveness of motion-aware speed conditioning for passive sonar enhancement in single-sensor AUV systems, under controlled synthetic-data conditions representative of AUV constant depth, speed, and heading in shallow-water lake environments.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"327-342"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating sediment properties using a new source level function for wind-driven underwater sound derived from long-term archival data. 基于长期档案数据的风动水声声源级函数估算沉积物特性。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042217
S Bruce Martin, Martin Siderius

Wind-driven breaking waves generate the background sound throughout the ocean. An accurate source level for wind-driven breaking waves is needed for estimating the ambient sound levels needed for sound exposure modeling, environmental assessments, and assessing the detection performance of sonars. Previous models applied a constant roll-off of sound levels at -16 dB/decade at all wind speeds, and these models' source levels were flat at frequencies below ∼1000 Hz due to a lack of measurements. Here, we analyzed 16 long-term archival datasets with limited anthropogenic sound sources to estimate the wind-driven source level down to 100 Hz. We estimated the site-specific areic propagation loss (APL) using a ray-based model and then added the APL to the median received levels at each wind speed to obtain the source level. An equation for the areic dipole source level is provided that increases as wind speed cubed, like most other air-ocean coupling processes. The model may be used to estimate sediment properties (given a wind speed history and measured sound levels) or to estimate wind speeds (given the sediment type and measured sound levels). It is well suited for estimating ambient sound levels from wind for soundscape modeling. An open-source implementation is available.

风驱动的破碎海浪在整个海洋中产生背景声。为了估计声音暴露建模、环境评估和评估声纳探测性能所需的环境声级,需要精确的风力破碎波声源级。以前的模型在所有风速下都以-16 dB/ 10年的速度恒定滚动声级,由于缺乏测量,这些模型的声源级在低于~ 1000 Hz的频率下是平坦的。在此,我们分析了16个长期档案数据集,这些数据集具有有限的人为声源,以估计低至100 Hz的风力声源水平。我们使用基于射线的模型估计了特定站点的面积传播损耗(APL),然后将APL与每个风速下的中位数接收电平相加,得到源电平。提供了一个随风速立方增加的区域偶极源水平的方程,就像大多数其他空气-海洋耦合过程一样。该模型可用于估计沉积物特性(给定风速历史和测量声级)或估计风速(给定沉积物类型和测量声级)。它非常适合于估计来自风的声景建模的环境声级。一个开源实现是可用的。
{"title":"Estimating sediment properties using a new source level function for wind-driven underwater sound derived from long-term archival data.","authors":"S Bruce Martin, Martin Siderius","doi":"10.1121/10.0042217","DOIUrl":"https://doi.org/10.1121/10.0042217","url":null,"abstract":"<p><p>Wind-driven breaking waves generate the background sound throughout the ocean. An accurate source level for wind-driven breaking waves is needed for estimating the ambient sound levels needed for sound exposure modeling, environmental assessments, and assessing the detection performance of sonars. Previous models applied a constant roll-off of sound levels at -16 dB/decade at all wind speeds, and these models' source levels were flat at frequencies below ∼1000 Hz due to a lack of measurements. Here, we analyzed 16 long-term archival datasets with limited anthropogenic sound sources to estimate the wind-driven source level down to 100 Hz. We estimated the site-specific areic propagation loss (APL) using a ray-based model and then added the APL to the median received levels at each wind speed to obtain the source level. An equation for the areic dipole source level is provided that increases as wind speed cubed, like most other air-ocean coupling processes. The model may be used to estimate sediment properties (given a wind speed history and measured sound levels) or to estimate wind speeds (given the sediment type and measured sound levels). It is well suited for estimating ambient sound levels from wind for soundscape modeling. An open-source implementation is available.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"300-314"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable machine learning models for outdoor exceedance level prediction based on geospatial variables. 基于地理空间变量的室外超标水平预测的可解释机器学习模型。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0042225
Ciro Régulo Martínez, Débora Pollicelli, Juan Bajo, Sharolyn J Anderson, Claudio Delrieux

Sound level modeling has emerged as an essential tool for predicting acoustic environments. We present the development and analysis of models using a dataset previously applied for sound exceedance level modeling in the contiguous United States. This dataset comprises acoustic exceedance levels measured in diverse locations including National Park Service sites and urban environments. We applied advanced python libraries to train Random Forest regression models to predict exceedance levels from 99 geospatial variables. In total, 3 general and 5 ancillary fully data-driven models (not modeling actual physics of sound propagation) were developed, and the particular performance and limitations of each model is discussed. Results show promising predictive power, with R2 between 0.54 and 0.91 and root mean squared error between 1.77 and 5.97 dB, where models incorporating more urban information performed better. These results highlight the strength of the models, with performance variability primarily attributed to the limited coverage of diverse natural and urban environments in the current dataset. Results are accessible via an interactive online dashboard, allowing users without machine learning expertise to analyze different aspects of the models. This platform supports broader accessibility, encouraging a wider audience to engage with outdoor sound level modeling and its applications.

声级建模已成为预测声环境的重要工具。我们使用先前应用于美国相邻地区声音超标水平建模的数据集,提出了模型的开发和分析。该数据集包括在不同地点测量的声学超标水平,包括国家公园管理局网站和城市环境。我们使用先进的python库来训练随机森林回归模型,以预测99个地理空间变量的超标水平。总共建立了3个通用模型和5个辅助的完全数据驱动模型(不模拟声音传播的实际物理),并讨论了每个模型的特定性能和局限性。结果显示出良好的预测能力,R2在0.54 ~ 0.91之间,均方根误差在1.77 ~ 5.97 dB之间,其中包含更多城市信息的模型表现更好。这些结果突出了模型的强度,其性能变化主要归因于当前数据集中不同自然和城市环境的有限覆盖。结果可以通过交互式在线仪表板访问,允许没有机器学习专业知识的用户分析模型的不同方面。该平台支持更广泛的可访问性,鼓励更广泛的受众参与户外声级建模及其应用。
{"title":"Explainable machine learning models for outdoor exceedance level prediction based on geospatial variables.","authors":"Ciro Régulo Martínez, Débora Pollicelli, Juan Bajo, Sharolyn J Anderson, Claudio Delrieux","doi":"10.1121/10.0042225","DOIUrl":"https://doi.org/10.1121/10.0042225","url":null,"abstract":"<p><p>Sound level modeling has emerged as an essential tool for predicting acoustic environments. We present the development and analysis of models using a dataset previously applied for sound exceedance level modeling in the contiguous United States. This dataset comprises acoustic exceedance levels measured in diverse locations including National Park Service sites and urban environments. We applied advanced python libraries to train Random Forest regression models to predict exceedance levels from 99 geospatial variables. In total, 3 general and 5 ancillary fully data-driven models (not modeling actual physics of sound propagation) were developed, and the particular performance and limitations of each model is discussed. Results show promising predictive power, with R2 between 0.54 and 0.91 and root mean squared error between 1.77 and 5.97 dB, where models incorporating more urban information performed better. These results highlight the strength of the models, with performance variability primarily attributed to the limited coverage of diverse natural and urban environments in the current dataset. Results are accessible via an interactive online dashboard, allowing users without machine learning expertise to analyze different aspects of the models. This platform supports broader accessibility, encouraging a wider audience to engage with outdoor sound level modeling and its applications.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"459-469"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145989765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Acoustic-phonetic masking in Spanish vowel recognition by native English- and Spanish-speaking subjects. 母语为英语和西班牙语的受试者在西班牙语元音识别中的声学-语音掩蔽。
IF 2.3 2区 物理与天体物理 Q2 ACOUSTICS Pub Date : 2026-01-01 DOI: 10.1121/10.0041884
Mark Gibson, Marcel Schlechtweg, Xianhui Wang, Judit Ayala Alcalde, Mark Arvidson, Li Xu

We report results for a psycho-acoustic experiment examining Spanish vowel ([a,e,i,o,u]). recognition in speech-shaped noise (SSN) and background babble (1-16 talkers) by two listening groups: native Spanish speakers (SP group) and native English speakers (EN group). The motivation for the current study is to investigate acoustic-phonetic and informational masking (APM and IM, respectively) effects (1) on segment/phoneme recognition, and (2) by participants who do not speak the language of the target or masker (as well as native speakers of Spanish) in order to disambiguate the effects of APM and IM. For the tests, background noise, both SSN and background babble, were presented at three signal-to-noise ratios (at 0, -6, and -12 dB) while a target containing one of the five Spanish vowels was presented in the syllables [da, de, di, do, du]. Inter-group differences in response accuracy point to significant effects of APM as listening conditions erode, and minimal effects due to higher-order factors based on masker meaningfulness, semantic content, and language familiarity.

我们报告了一项检测西班牙元音([a,e,i,o,u])的心理声学实验的结果。两组以西班牙语为母语的人(SP组)和以英语为母语的人(EN组)对语音形状噪声(SSN)和背景胡言乱语(1-16个说话者)的识别。本研究的动机是调查声学-语音掩蔽和信息掩蔽(分别为APM和IM)对片段/音素识别的影响(1),以及(2)不讲目标语言或掩蔽者语言的参与者(以及母语为西班牙语的人),以消除APM和IM的影响。在测试中,背景噪声(SSN和背景牙牙学语)以三种信噪比(0,-6和-12 dB)呈现,同时在音节[da, de, di, do, du]中呈现包含五个西班牙语元音之一的目标。组间反应准确性的差异表明,随着听力条件的侵蚀,APM的影响显著,而基于掩模意义、语义内容和语言熟悉度的高阶因素的影响最小。
{"title":"Acoustic-phonetic masking in Spanish vowel recognition by native English- and Spanish-speaking subjects.","authors":"Mark Gibson, Marcel Schlechtweg, Xianhui Wang, Judit Ayala Alcalde, Mark Arvidson, Li Xu","doi":"10.1121/10.0041884","DOIUrl":"10.1121/10.0041884","url":null,"abstract":"<p><p>We report results for a psycho-acoustic experiment examining Spanish vowel ([a,e,i,o,u]). recognition in speech-shaped noise (SSN) and background babble (1-16 talkers) by two listening groups: native Spanish speakers (SP group) and native English speakers (EN group). The motivation for the current study is to investigate acoustic-phonetic and informational masking (APM and IM, respectively) effects (1) on segment/phoneme recognition, and (2) by participants who do not speak the language of the target or masker (as well as native speakers of Spanish) in order to disambiguate the effects of APM and IM. For the tests, background noise, both SSN and background babble, were presented at three signal-to-noise ratios (at 0, -6, and -12 dB) while a target containing one of the five Spanish vowels was presented in the syllables [da, de, di, do, du]. Inter-group differences in response accuracy point to significant effects of APM as listening conditions erode, and minimal effects due to higher-order factors based on masker meaningfulness, semantic content, and language familiarity.</p>","PeriodicalId":17168,"journal":{"name":"Journal of the Acoustical Society of America","volume":"159 1","pages":"105-116"},"PeriodicalIF":2.3,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145889486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of the Acoustical Society of America
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1