A general theory for obtaining spherical wave function expansion coefficients for a sound beam transmitted by a planar velocity source is presented. By neglecting evanescent wave components and thus making the proposed method only approximate in the near field, it is shown that these coefficients can be obtained for any expansion point using just the normal velocity condition in the source plane. Additionally, simplifications are presented for axisymmetric sources. Results are compared with direct numerical evaluation of the Rayleigh integral for source conditions corresponding to a circular and rectangular piston. The present theory can be used in calculations of acoustic scattering and radiation force for spherical objects, or in spherical acoustical holography applications.
{"title":"Spherical wave expansion coefficients for radiation from a planar velocity source.","authors":"Blake E Simon, Mark F Hamilton","doi":"10.1121/10.0039687","DOIUrl":"https://doi.org/10.1121/10.0039687","url":null,"abstract":"<p><p>A general theory for obtaining spherical wave function expansion coefficients for a sound beam transmitted by a planar velocity source is presented. By neglecting evanescent wave components and thus making the proposed method only approximate in the near field, it is shown that these coefficients can be obtained for any expansion point using just the normal velocity condition in the source plane. Additionally, simplifications are presented for axisymmetric sources. Results are compared with direct numerical evaluation of the Rayleigh integral for source conditions corresponding to a circular and rectangular piston. The present theory can be used in calculations of acoustic scattering and radiation force for spherical objects, or in spherical acoustical holography applications.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 10","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145402493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To accurately characterize the non-radial motion of the source relative to the receiver, a three-dimensional (3D) model is essential. The extended Kalman filter (EKF) state matrix is employed to characterize the source's 3D motion. The measurement input for the EKF is the time delay between the direct and surface-reflected arrivals. The differences in the partial derivatives of the distance component have been identified and discussed. Through iterative filtering, a reliable estimate of the source's position in 3D space is obtained. Both simulations and experiments validate the effectiveness of the method, with experimental depth estimation errors within 1.5%.
{"title":"Three-dimensional acoustic source localization based on multipath time delay in deep ocean.","authors":"Zhen Zhang, Haigang Zhang, Jinshan Fu","doi":"10.1121/10.0039398","DOIUrl":"https://doi.org/10.1121/10.0039398","url":null,"abstract":"<p><p>To accurately characterize the non-radial motion of the source relative to the receiver, a three-dimensional (3D) model is essential. The extended Kalman filter (EKF) state matrix is employed to characterize the source's 3D motion. The measurement input for the EKF is the time delay between the direct and surface-reflected arrivals. The differences in the partial derivatives of the distance component have been identified and discussed. Through iterative filtering, a reliable estimate of the source's position in 3D space is obtained. Both simulations and experiments validate the effectiveness of the method, with experimental depth estimation errors within 1.5%.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 10","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145214632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Katelyn A Berg, Hugh M Birky, Victoria A Sevich, Aaron C Moberly, Terrin N Tamati
This study examined relationships between cochlear implant-aided speech recognition, sound quality ratings, and quality of life outcomes in 41 adult cochlear implant users. Participants completed word and sentence recognition tasks, the Speech, Spatial and Qualities (SSQ) questionnaire, and the Cochlear Implant Quality of Life (CIQOL) questionnaire. Sound quality ratings on SSQ showed a weak correlation with sentence recognition in noise but strong positive correlations with CIQOL scores. Sound quality independently predicted quality of life, explaining 32% of global score variance, while speech recognition measures showed no effects. These findings highlight sound quality as a distinct and meaningful outcome metric.
{"title":"Sound quality, not speech recognition, explains cochlear implant-related quality of life outcomes.","authors":"Katelyn A Berg, Hugh M Birky, Victoria A Sevich, Aaron C Moberly, Terrin N Tamati","doi":"10.1121/10.0039069","DOIUrl":"10.1121/10.0039069","url":null,"abstract":"<p><p>This study examined relationships between cochlear implant-aided speech recognition, sound quality ratings, and quality of life outcomes in 41 adult cochlear implant users. Participants completed word and sentence recognition tasks, the Speech, Spatial and Qualities (SSQ) questionnaire, and the Cochlear Implant Quality of Life (CIQOL) questionnaire. Sound quality ratings on SSQ showed a weak correlation with sentence recognition in noise but strong positive correlations with CIQOL scores. Sound quality independently predicted quality of life, explaining 32% of global score variance, while speech recognition measures showed no effects. These findings highlight sound quality as a distinct and meaningful outcome metric.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 10","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145287858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Spatially separating target and masker talkers improves speech perception in noise, an effect known as spatial release from masking (SRM). Independently, the perceived location of a sound can erroneously shift towards an associated but spatially displaced visual stimulus (the "ventriloquist effect"). This study investigated whether SRM can be induced by spatially separating visual stimuli associated with a target and masker without separating the sound sources themselves. Results showed that SRM was not induced by spatially separated visual stimuli, but collocated visual stimuli reduced the benefit of auditory SRM. There was no influence of individual differences in auditory localization ability on effects related to the visual stimuli.
{"title":"Visual induction of spatial release from masking during speech perception in noise.","authors":"Sarah Knight, Charlotte Levy, Sven Mattys","doi":"10.1121/10.0039627","DOIUrl":"https://doi.org/10.1121/10.0039627","url":null,"abstract":"<p><p>Spatially separating target and masker talkers improves speech perception in noise, an effect known as spatial release from masking (SRM). Independently, the perceived location of a sound can erroneously shift towards an associated but spatially displaced visual stimulus (the \"ventriloquist effect\"). This study investigated whether SRM can be induced by spatially separating visual stimuli associated with a target and masker without separating the sound sources themselves. Results showed that SRM was not induced by spatially separated visual stimuli, but collocated visual stimuli reduced the benefit of auditory SRM. There was no influence of individual differences in auditory localization ability on effects related to the visual stimuli.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 10","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145380017","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dame Fall, Marc Duquennoy, Nikolay Smagin, Zakariae Oumekloul, Mohammadi Ouaftouh
This study presents a non-destructive method for estimating surface acoustic wave attenuation, which is highly sensitive to microstructural features, especially at high frequencies. The method uses a single wideband dispersive interdigital transducer (IDT) that remotely emits acoustic waves at the sample's edge. Chirp compression of the temporal displacement response is achieved by correlating the excitation signal with the spatial configuration of the IDT's electrodes. This technique generates high-amplitude pulses with a sufficient signal-to-noise ratio, critical for enabling accurate attenuation estimation over a frequency range (15-70 MHz). Results from nickel and aluminum demonstrate the method's effectiveness for rapid material characterization.
{"title":"Broadband surface acoustic wave attenuation in metals using chirp compression and dispersive interdigital transducers.","authors":"Dame Fall, Marc Duquennoy, Nikolay Smagin, Zakariae Oumekloul, Mohammadi Ouaftouh","doi":"10.1121/10.0039237","DOIUrl":"https://doi.org/10.1121/10.0039237","url":null,"abstract":"<p><p>This study presents a non-destructive method for estimating surface acoustic wave attenuation, which is highly sensitive to microstructural features, especially at high frequencies. The method uses a single wideband dispersive interdigital transducer (IDT) that remotely emits acoustic waves at the sample's edge. Chirp compression of the temporal displacement response is achieved by correlating the excitation signal with the spatial configuration of the IDT's electrodes. This technique generates high-amplitude pulses with a sufficient signal-to-noise ratio, critical for enabling accurate attenuation estimation over a frequency range (15-70 MHz). Results from nickel and aluminum demonstrate the method's effectiveness for rapid material characterization.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 9","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145002179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hearing aid (HA) processing can affect acoustic features linked with emotions, potentially making them less distinguishable. This study investigated whether HA processing, with both standard and short processing delays, affects emotion prediction from a set of acoustic features associated with speech emotions and how well these predictions align with perceived emotions. The findings indicated that anger and sadness are the easiest emotions to predict from acoustic features, while happiness and fear are the most accurately perceived emotions by listeners with normal hearing. HA processing, regardless of delay, does not seem to impair the predictability of emotions from acoustic features or the perception of these emotions.
{"title":"Influence of hearing aid processing on acoustic features associated with emotional speech: Acoustic analyses and perception by listeners with normal hearing.","authors":"Frederic Marmel, Dina Lelic","doi":"10.1121/10.0039220","DOIUrl":"https://doi.org/10.1121/10.0039220","url":null,"abstract":"<p><p>Hearing aid (HA) processing can affect acoustic features linked with emotions, potentially making them less distinguishable. This study investigated whether HA processing, with both standard and short processing delays, affects emotion prediction from a set of acoustic features associated with speech emotions and how well these predictions align with perceived emotions. The findings indicated that anger and sadness are the easiest emotions to predict from acoustic features, while happiness and fear are the most accurately perceived emotions by listeners with normal hearing. HA processing, regardless of delay, does not seem to impair the predictability of emotions from acoustic features or the perception of these emotions.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 9","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145031224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Benjamin J Kirby, Alaina Cunningham, Olivia Montou Zant
Misophonia is a condition characterized by intense negative emotional reactions to trigger sounds and related stimuli. In this study, adult listeners (N = 15) with a self-reported history of misophonia symptoms and a control group without misophonia (N = 15) completed listening judgements of recorded misophonia trigger stimuli using a standard scale. Participants also completed an established questionnaire of misophonia symptoms, the Misophonia Questionnaire (MQ). Summed scores of the listening task were significantly correlated with overall MQ score. The misophonia group had significantly higher listening scores and MQ scores compared to controls. These findings indicate applications for psychoacoustic methods in the assessment of misophonia.
{"title":"Psychoacoustic assessment of misophonia.","authors":"Benjamin J Kirby, Alaina Cunningham, Olivia Montou Zant","doi":"10.1121/10.0039238","DOIUrl":"https://doi.org/10.1121/10.0039238","url":null,"abstract":"<p><p>Misophonia is a condition characterized by intense negative emotional reactions to trigger sounds and related stimuli. In this study, adult listeners (N = 15) with a self-reported history of misophonia symptoms and a control group without misophonia (N = 15) completed listening judgements of recorded misophonia trigger stimuli using a standard scale. Participants also completed an established questionnaire of misophonia symptoms, the Misophonia Questionnaire (MQ). Summed scores of the listening task were significantly correlated with overall MQ score. The misophonia group had significantly higher listening scores and MQ scores compared to controls. These findings indicate applications for psychoacoustic methods in the assessment of misophonia.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 9","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145024843","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jenna L Cramer, Ashley Reynard, Vanessa Torres, Jeremy J Donai
Identifying speakers of interest in an auditory scene is a fundamental task that facilitates effective communication. Little is known about the specific contributions of spectral and temporal detail required for identifying a specific speaker of interest by human listeners. This study investigated the relative contributions of spectral and temporal detail for identifying a speaker of interest and perceived effort in doing so. Results showed significant improvements in speaker identification and decreased effort ratings as spectral channels increased. Improved speaker identification performance with increased temporal filter cutoff from 20 Hz to 800 Hz was observed. These results have implications for speech signal processing by amplification devices and automated speaker recognition systems.
{"title":"Spectral and temporal information and presentation mode effects on individual speaker identification and listening effort.","authors":"Jenna L Cramer, Ashley Reynard, Vanessa Torres, Jeremy J Donai","doi":"10.1121/10.0039369","DOIUrl":"https://doi.org/10.1121/10.0039369","url":null,"abstract":"<p><p>Identifying speakers of interest in an auditory scene is a fundamental task that facilitates effective communication. Little is known about the specific contributions of spectral and temporal detail required for identifying a specific speaker of interest by human listeners. This study investigated the relative contributions of spectral and temporal detail for identifying a speaker of interest and perceived effort in doing so. Results showed significant improvements in speaker identification and decreased effort ratings as spectral channels increased. Improved speaker identification performance with increased temporal filter cutoff from 20 Hz to 800 Hz was observed. These results have implications for speech signal processing by amplification devices and automated speaker recognition systems.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 9","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145082696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A prior solution for the scattering of traveling wave sound by concentric fluid spheres is recast using complex unimodular s-function notation, which is convenient for expressing partial wave amplitudes and radiation forces on spheres in standing waves. Viscous and thermal energy dissipation are neglected. The fluid core affects the low-frequency dynamics of the fluid shell. The lowest monopole mode of air-filled liquid shells in air is considered. The frequency is approximated by generalizing the analysis of the Minnaert resonance of an air bubble in water. This analysis is relevant to the acoustical scattering by and conditions for trapping of compound drops.
{"title":"Concentric fluid spheres: Scattering and radiation forces and the lowest monopole resonance of bubble shells.","authors":"Philip L Marston","doi":"10.1121/10.0039423","DOIUrl":"https://doi.org/10.1121/10.0039423","url":null,"abstract":"<p><p>A prior solution for the scattering of traveling wave sound by concentric fluid spheres is recast using complex unimodular s-function notation, which is convenient for expressing partial wave amplitudes and radiation forces on spheres in standing waves. Viscous and thermal energy dissipation are neglected. The fluid core affects the low-frequency dynamics of the fluid shell. The lowest monopole mode of air-filled liquid shells in air is considered. The frequency is approximated by generalizing the analysis of the Minnaert resonance of an air bubble in water. This analysis is relevant to the acoustical scattering by and conditions for trapping of compound drops.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 9","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145126593","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ny Tsiky Rakotomalala, Pierre Baraduc, Pascal Perrier
We present a speech motor control model that integrates optimal feedback control (OFC) for movement planning and execution with a biomechanical model of the vocal tract. The OFC model was designed to optimize a cost function that combines motor effort and the achievement of multisensory goal zones. We show that the model can account for various aspects of speech production: kinematic properties, coarticulation, and sensorimotor integration. Furthermore, we provide evidence that hearing, proprioception, and tactile feedback may play distinct roles in shaping speech trajectories.
{"title":"GEPPETO-OFC: An optimal feedback speech motor control model integrating biomechanical constraints and multisensory goal specification.","authors":"Ny Tsiky Rakotomalala, Pierre Baraduc, Pascal Perrier","doi":"10.1121/10.0039197","DOIUrl":"https://doi.org/10.1121/10.0039197","url":null,"abstract":"<p><p>We present a speech motor control model that integrates optimal feedback control (OFC) for movement planning and execution with a biomechanical model of the vocal tract. The OFC model was designed to optimize a cost function that combines motor effort and the achievement of multisensory goal zones. We show that the model can account for various aspects of speech production: kinematic properties, coarticulation, and sensorimotor integration. Furthermore, we provide evidence that hearing, proprioception, and tactile feedback may play distinct roles in shaping speech trajectories.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":"5 9","pages":""},"PeriodicalIF":1.4,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145002173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}