Christophe Lesimple, Volker Kuehnel, Kai Siedenburg
Music is an important signal class for hearing aids, and musical genre is often used as a descriptor for stimulus selection. However, little research has systematically investigated the acoustical properties of musical genres with respect to hearing aid amplification. Here, extracts from a combination of two comprehensive music databases were acoustically analyzed. Considerable overlap in acoustic descriptor space between genres emerged. By simulating hearing aid processing, it was shown that effects of amplification regarding dynamic range compression and spectral weighting differed across musical genres, underlining the critical role of systematic stimulus selection for research on music and hearing aids.
{"title":"Hearing aid evaluation for music: Accounting for acoustical variability of music stimuli.","authors":"Christophe Lesimple, Volker Kuehnel, Kai Siedenburg","doi":"10.1121/10.0028397","DOIUrl":"https://doi.org/10.1121/10.0028397","url":null,"abstract":"<p><p>Music is an important signal class for hearing aids, and musical genre is often used as a descriptor for stimulus selection. However, little research has systematically investigated the acoustical properties of musical genres with respect to hearing aid amplification. Here, extracts from a combination of two comprehensive music databases were acoustically analyzed. Considerable overlap in acoustic descriptor space between genres emerged. By simulating hearing aid processing, it was shown that effects of amplification regarding dynamic range compression and spectral weighting differed across musical genres, underlining the critical role of systematic stimulus selection for research on music and hearing aids.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In an attempt to develop tests of auditory temporal resolution using gap detection, we conducted computer simulations of Zippy Estimation by Sequential Testing (ZEST), an adaptive Bayesian threshold estimation procedure, for measuring gap detection thresholds. The results showed that the measures of efficiency and precision of ZEST changed with the mean and standard deviation (SD) of the initial probability density function implemented in ZEST. Appropriate combinations of mean and SD values led to efficient ZEST performance; i.e., the threshold estimates converged to their true values after 10 to 15 trials.
{"title":"Tests of human auditory temporal resolution: Simulations of Bayesian threshold estimation for auditory gap detection.","authors":"Shuji Mori, Yuto Murata, Takashi Morimoto, Yasuhide Okamoto, Sho Kanzaki","doi":"10.1121/10.0028501","DOIUrl":"https://doi.org/10.1121/10.0028501","url":null,"abstract":"<p><p>In an attempt to develop tests of auditory temporal resolution using gap detection, we conducted computer simulations of Zippy Estimation by Sequential Testing (ZEST), an adaptive Bayesian threshold estimation procedure, for measuring gap detection thresholds. The results showed that the measures of efficiency and precision of ZEST changed with the mean and standard deviation (SD) of the initial probability density function implemented in ZEST. Appropriate combinations of mean and SD values led to efficient ZEST performance; i.e., the threshold estimates converged to their true values after 10 to 15 trials.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142121270","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A test is proposed to characterize the performance of speech recognition systems. The QuickSIN test is used by audiologists to measure the ability of humans to recognize continuous speech in noise. This test yields the signal-to-noise ratio at which individuals can correctly recognize 50% of the keywords in low-context sentences. It is argued that a metric for automatic speech recognizers will ground the performance of automatic speech-in-noise recognizers to human abilities. Here, it is demonstrated that the performance of modern recognizers, built using millions of hours of unsupervised training data, is anywhere from normal to mildly impaired in noise compared to human participants.
{"title":"Comparing human and machine speech recognition in noise with QuickSIN.","authors":"Malcolm Slaney, Matthew B Fitzgerald","doi":"10.1121/10.0028612","DOIUrl":"https://doi.org/10.1121/10.0028612","url":null,"abstract":"<p><p>A test is proposed to characterize the performance of speech recognition systems. The QuickSIN test is used by audiologists to measure the ability of humans to recognize continuous speech in noise. This test yields the signal-to-noise ratio at which individuals can correctly recognize 50% of the keywords in low-context sentences. It is argued that a metric for automatic speech recognizers will ground the performance of automatic speech-in-noise recognizers to human abilities. Here, it is demonstrated that the performance of modern recognizers, built using millions of hours of unsupervised training data, is anywhere from normal to mildly impaired in noise compared to human participants.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142156849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To calculate a grand average waveform for a set of auditory brain-stem responses (ABRs), no generally accepted method exists. Here, we evaluate a new method using temporal adjustment of the underlying ABRs. Compared to a method without temporal adjustment, the new method results in higher amplitudes of the individual waves in the grand average. The grand average produced by the method better represents the group mean wave-amplitudes because it reduces smearing of the individual waves caused by inter-subject latency variability.
要计算一组听觉脑干反应(ABR)的总平均波形,目前还没有公认的方法。在此,我们评估了一种使用时间调整基础 ABR 的新方法。与未进行时间调整的方法相比,新方法可使总平均值中的单个波振幅更高。该方法产生的总平均值能更好地代表群体的平均波幅,因为它减少了受试者之间的潜伏期差异对单个波幅造成的涂抹。
{"title":"A method for calculating the grand average of a set of auditory brain-stem responses.","authors":"Sinnet G B Kristensen, Claus Elberling","doi":"10.1121/10.0028320","DOIUrl":"https://doi.org/10.1121/10.0028320","url":null,"abstract":"<p><p>To calculate a grand average waveform for a set of auditory brain-stem responses (ABRs), no generally accepted method exists. Here, we evaluate a new method using temporal adjustment of the underlying ABRs. Compared to a method without temporal adjustment, the new method results in higher amplitudes of the individual waves in the grand average. The grand average produced by the method better represents the group mean wave-amplitudes because it reduces smearing of the individual waves caused by inter-subject latency variability.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Hyeonjeong Park, Fangxu Xing, Maureen Stone, Hahn Kang, Xiaofeng Liu, Jiachen Zhuo, Sidney Fels, Timothy G Reese, Van J Wedeen, Georges El Fakhri, Jerry L Prince, Jonghye Woo
The human tongue exhibits an orchestrated arrangement of internal muscles, working in sequential order to execute tongue movements. Understanding the muscle coordination patterns involved in tongue protrusive motion is crucial for advancing knowledge of tongue structure and function. To achieve this, this work focuses on five muscles known to contribute to protrusive motion. Tagged and diffusion MRI data are collected for analysis of muscle fiber geometry and motion patterns. Lagrangian strain measurements are derived, and Granger causal analysis is carried out to assess predictive information among the muscles. Experimental results suggest sequential muscle coordination of protrusive motion among distinct muscle groups.
{"title":"Investigating muscle coordination patterns with Granger causality analysis in protrusive motion from tagged and diffusion MRI.","authors":"Hyeonjeong Park, Fangxu Xing, Maureen Stone, Hahn Kang, Xiaofeng Liu, Jiachen Zhuo, Sidney Fels, Timothy G Reese, Van J Wedeen, Georges El Fakhri, Jerry L Prince, Jonghye Woo","doi":"10.1121/10.0028500","DOIUrl":"10.1121/10.0028500","url":null,"abstract":"<p><p>The human tongue exhibits an orchestrated arrangement of internal muscles, working in sequential order to execute tongue movements. Understanding the muscle coordination patterns involved in tongue protrusive motion is crucial for advancing knowledge of tongue structure and function. To achieve this, this work focuses on five muscles known to contribute to protrusive motion. Tagged and diffusion MRI data are collected for analysis of muscle fiber geometry and motion patterns. Lagrangian strain measurements are derived, and Granger causal analysis is carried out to assess predictive information among the muscles. Experimental results suggest sequential muscle coordination of protrusive motion among distinct muscle groups.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11384280/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142141922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The relevance of comodulation and interaural phase difference for speech perception is still unclear. We used speech-like stimuli to link spectro-temporal properties of formants with masking release. The stimuli comprised a tone and three masker bands centered at formant frequencies F1, F2, and F3 derived from a consonant-vowel. The target was a diotic or dichotic frequency-modulated tone following F2 trajectories. Results showed a small comodulation masking release, while the binaural masking level difference was comparable to previous findings. The data suggest that factors other than comodulation may play a dominant role in grouping frequency components in speech.
{"title":"Monaural and binaural masking release with speech-like stimuli.","authors":"Hyojin Kim, Viktorija Ratkute, Bastian Epp","doi":"10.1121/10.0028736","DOIUrl":"https://doi.org/10.1121/10.0028736","url":null,"abstract":"<p><p>The relevance of comodulation and interaural phase difference for speech perception is still unclear. We used speech-like stimuli to link spectro-temporal properties of formants with masking release. The stimuli comprised a tone and three masker bands centered at formant frequencies F1, F2, and F3 derived from a consonant-vowel. The target was a diotic or dichotic frequency-modulated tone following F2 trajectories. Results showed a small comodulation masking release, while the binaural masking level difference was comparable to previous findings. The data suggest that factors other than comodulation may play a dominant role in grouping frequency components in speech.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Joshua G W Bernstein, Julianna Voelker, Sandeep A Phatak
Psychoacoustic stimulus presentation to the cochlear implant via direct audio input (DAI) is no longer possible for many newer sound processors (SPs). This study assessed the feasibility of placing circumaural headphones over the SP. Calibration spectra for loudspeaker, DAI, and headphone modalities were estimated by measuring cochlear-implant electrical output levels for tones presented to SPs on an acoustic manikin. Differences in calibration spectra between modalities arose mainly from microphone-response characteristics (high-frequency differences between DAI and the other modalities) or a proximity effect (low-frequency differences between headphones and loudspeaker). Calibration tables are provided to adjust for differences between the three modalities.
{"title":"Headphones over the cochlear-implant sound processor to replace direct audio input.","authors":"Joshua G W Bernstein, Julianna Voelker, Sandeep A Phatak","doi":"10.1121/10.0028737","DOIUrl":"https://doi.org/10.1121/10.0028737","url":null,"abstract":"<p><p>Psychoacoustic stimulus presentation to the cochlear implant via direct audio input (DAI) is no longer possible for many newer sound processors (SPs). This study assessed the feasibility of placing circumaural headphones over the SP. Calibration spectra for loudspeaker, DAI, and headphone modalities were estimated by measuring cochlear-implant electrical output levels for tones presented to SPs on an acoustic manikin. Differences in calibration spectra between modalities arose mainly from microphone-response characteristics (high-frequency differences between DAI and the other modalities) or a proximity effect (low-frequency differences between headphones and loudspeaker). Calibration tables are provided to adjust for differences between the three modalities.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142309278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Input/output (I/O) functions of distortion-product otoacoustic emissions (DPOAEs) may contain sudden amplitude minima (notches) although they are measured in animals with a negligible reflection source. We measured DPOAEs in humans and analyzed the data by time-frequency filtering to decompose the nonlinear-distortion and coherent-reflection components of DPOAE. The presented I/O functions of the nonlinear-distortion component contain notches. We suggest that because these notches are present only in the nonlinear-distortion component, they result from destructive interference between distortion-product wavelets coming from the primary generation region. Simulations conducted with a nonlinear cochlear model showed qualitative similarities with the presented experimental results.
{"title":"Minima in cubic distortion-product otoacoustic emission input/output functions due to distributed primary sources.","authors":"Václav Vencovský, Aleš Vetešník","doi":"10.1121/10.0028386","DOIUrl":"https://doi.org/10.1121/10.0028386","url":null,"abstract":"<p><p>Input/output (I/O) functions of distortion-product otoacoustic emissions (DPOAEs) may contain sudden amplitude minima (notches) although they are measured in animals with a negligible reflection source. We measured DPOAEs in humans and analyzed the data by time-frequency filtering to decompose the nonlinear-distortion and coherent-reflection components of DPOAE. The presented I/O functions of the nonlinear-distortion component contain notches. We suggest that because these notches are present only in the nonlinear-distortion component, they result from destructive interference between distortion-product wavelets coming from the primary generation region. Simulations conducted with a nonlinear cochlear model showed qualitative similarities with the presented experimental results.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142134645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Greek uses H*, L + H*, and H* + L, all followed by L-L% edge tones, as nuclear pitch accents in statements. A previous analysis demonstrated that these accents are distinguished by F0 scaling and contour shape. This study expands the earlier investigation by exploring additional cues, namely, voice quality, amplitude, and duration, in distinguishing the pitch accents, and investigating individual variability in the selection of both F0 and non-F0 cues. Bayesian multivariate analysis and hierarchical clustering demonstrate that the accents are distinguished not only by F0 but also by additional cues at the group level, with individual variability in cue selection.
{"title":"Individual variability in the use of tonal and non-tonal cues in intonationa).","authors":"Na Hu, Amalia Arvaniti","doi":"10.1121/10.0028613","DOIUrl":"https://doi.org/10.1121/10.0028613","url":null,"abstract":"<p><p>Greek uses H*, L + H*, and H* + L, all followed by L-L% edge tones, as nuclear pitch accents in statements. A previous analysis demonstrated that these accents are distinguished by F0 scaling and contour shape. This study expands the earlier investigation by exploring additional cues, namely, voice quality, amplitude, and duration, in distinguishing the pitch accents, and investigating individual variability in the selection of both F0 and non-F0 cues. Bayesian multivariate analysis and hierarchical clustering demonstrate that the accents are distinguished not only by F0 but also by additional cues at the group level, with individual variability in cue selection.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142302633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This study examined the influence of stimulus properties on sound externalization when listening with hearing aids. Normally hearing listeners were presented with broadband "tokens" (environmental sounds and speech) from loudspeakers, and rated externalization using a continuous scale. In separate blocks, they listened unaided or while wearing behind-the-ear hearing aids with closed domes and low gain (linear or compressive). There was a significant influence of token on ratings, even for unaided listening, and the effect of hearing aids depended on token. An acoustic analysis indicated that hearing aids were more likely to disrupt externalization for peakier sounds with a low-frequency emphasis.
{"title":"Effect of hearing aids on the externalization of everyday sounds.","authors":"Elin Roverud, Virginia Best","doi":"10.1121/10.0028381","DOIUrl":"10.1121/10.0028381","url":null,"abstract":"<p><p>This study examined the influence of stimulus properties on sound externalization when listening with hearing aids. Normally hearing listeners were presented with broadband \"tokens\" (environmental sounds and speech) from loudspeakers, and rated externalization using a continuous scale. In separate blocks, they listened unaided or while wearing behind-the-ear hearing aids with closed domes and low gain (linear or compressive). There was a significant influence of token on ratings, even for unaided listening, and the effect of hearing aids depended on token. An acoustic analysis indicated that hearing aids were more likely to disrupt externalization for peakier sounds with a low-frequency emphasis.</p>","PeriodicalId":73538,"journal":{"name":"JASA express letters","volume":null,"pages":null},"PeriodicalIF":1.2,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11382285/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142121269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}