Pub Date : 2025-11-19DOI: 10.1016/j.heares.2025.109479
Linwei Zhang , Xikun Lu , Yangyang Zheng , Ruohan Na , Jinqiu Sang , Lei Jin
Eustachian tube dysfunction (ETD) is a critical factor for otitis media with effusion (OME) in children, yet there is currently no gold standard for its direct clinical measurement. Sonotubometry serves as a physiological diagnosis method for ETD, assessing sound transmission from the nasopharynx to the ear canal during swallowing. However, its usability is limited by challenges in interpreting the recordings. In this work, a machine learning (ML) model is applied to analyze the audio features from sonotubometry for the detection and classification of ETD. Various audio feature extraction techniques were employed, with the Mel-frequency cepstral coefficients (MFCC) features yielding the best results. Specifically, when combined with the convolutional neural network (CNN) model, MFCC achieved a sensitivity of 0.975 (95 % CI: 0.906, 1.000), which significantly outperformed the traditional threshold-based method 0.645 (95 % CI: 0.293, 0.997). Through feature heatmaps generated via masking, it was found that classification of normal ET opening primarily relies on the acoustic response of 6 to 8 kHz. This work demonstrates the potential of ML-based sonotubometry to provide an objective, non-invasive, and efficient diagnostic tool for ETD.
{"title":"Multi-feature machine learning classification of sonotubometry for eustachian tube dysfunction assessment","authors":"Linwei Zhang , Xikun Lu , Yangyang Zheng , Ruohan Na , Jinqiu Sang , Lei Jin","doi":"10.1016/j.heares.2025.109479","DOIUrl":"10.1016/j.heares.2025.109479","url":null,"abstract":"<div><div>Eustachian tube dysfunction (ETD) is a critical factor for otitis media with effusion (OME) in children, yet there is currently no gold standard for its direct clinical measurement. Sonotubometry serves as a physiological diagnosis method for ETD, assessing sound transmission from the nasopharynx to the ear canal during swallowing. However, its usability is limited by challenges in interpreting the recordings. In this work, a machine learning (ML) model is applied to analyze the audio features from sonotubometry for the detection and classification of ETD. Various audio feature extraction techniques were employed, with the Mel-frequency cepstral coefficients (MFCC) features yielding the best results. Specifically, when combined with the convolutional neural network (CNN) model, MFCC achieved a sensitivity of 0.975 (95 % CI: 0.906, 1.000), which significantly outperformed the traditional threshold-based method 0.645 (95 % CI: 0.293, 0.997). Through feature heatmaps generated via masking, it was found that classification of normal ET opening primarily relies on the acoustic response of 6 to 8 kHz. This work demonstrates the potential of ML-based sonotubometry to provide an objective, non-invasive, and efficient diagnostic tool for ETD.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109479"},"PeriodicalIF":2.5,"publicationDate":"2025-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145587357","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-17DOI: 10.1016/j.heares.2025.109478
Alex C. Clonan , Ian H. Stevenson , Monty A. Escabí
Human hearing is critical to everyday communication and the perception of natural auditory scenes. For individuals with misophonia, sounds commonly experienced in daily life can evoke severe discomfort and distress. Aversion is often described in terms of broad sound categories, such as bodily sounds, but what acoustic features cause specific sounds to be aversive or not, within the same category or across different individuals, remains unclear. Here, we explore whether bottom-up statistical sound features processed in the auditory periphery and midbrain can explain aversion to sounds. Using the Free Open-Access Misophonia Stimuli (FOAMS) dataset and a hierarchical model of the auditory system, we find that sound summary statistics can predict discomfort ratings in participants with misophonia. For each listener, the model produces individualized transfer functions that pinpoint the specific spectrotemporal modulations that contribute towards sound aversion. Overall, the model explains 76% of the variance in discomfort ratings, and we find substantial differences across participants in which sound features drive aversion. A major advantage of the modeling approach here is that it is sound-computable – perceptual ratings can be fit from or predicted for any sound. To illustrate applications of sound-computable models, we consider 1) extrapolation of participants' ratings to a large set of untested environmental sounds and develop 2) personalized trigger detection that uses the listener’s acoustic feature preferences to identify potential triggers in continuous audio. Model predictions identify many untested sound categories, not in the original FOAMS set, that may also be aversive and suggest that there may be substantial heterogeneity in how aversive specific sounds are within some sound categories. In continuous audio, we show how sound-computable models can identify the timing of potential triggers from sound mixtures. Altogether, our results suggest that acoustic features – spectrotemporal modulations, in particular – can practically be used to characterize the individualized patterns of aversion in participants with misophonia. Future perceptual studies using synthetic sounds and sound sets with more diverse acoustics will allow model predictions to be tested more broadly; however, sound-computable models may already have applications in precision diagnosis and management of misophonia.
{"title":"Identifying the acoustic fingerprints of trigger sounds and predicting discomfort for misophonia","authors":"Alex C. Clonan , Ian H. Stevenson , Monty A. Escabí","doi":"10.1016/j.heares.2025.109478","DOIUrl":"10.1016/j.heares.2025.109478","url":null,"abstract":"<div><div>Human hearing is critical to everyday communication and the perception of natural auditory scenes. For individuals with misophonia, sounds commonly experienced in daily life can evoke severe discomfort and distress. Aversion is often described in terms of broad sound categories, such as bodily sounds, but what acoustic features cause specific sounds to be aversive or not, within the same category or across different individuals, remains unclear. Here, we explore whether bottom-up statistical sound features processed in the auditory periphery and midbrain can explain aversion to sounds. Using the Free Open-Access Misophonia Stimuli (FOAMS) dataset and a hierarchical model of the auditory system, we find that sound summary statistics can predict discomfort ratings in participants with misophonia. For each listener, the model produces individualized transfer functions that pinpoint the specific spectrotemporal modulations that contribute towards sound aversion. Overall, the model explains 76% of the variance in discomfort ratings, and we find substantial differences across participants in which sound features drive aversion. A major advantage of the modeling approach here is that it is sound-computable – perceptual ratings can be fit from or predicted for any sound. To illustrate applications of sound-computable models, we consider 1) extrapolation of participants' ratings to a large set of untested environmental sounds and develop 2) personalized trigger detection that uses the listener’s acoustic feature preferences to identify potential triggers in continuous audio. Model predictions identify many untested sound categories, not in the original FOAMS set, that may also be aversive and suggest that there may be substantial heterogeneity in how aversive specific sounds are within some sound categories. In continuous audio, we show how sound-computable models can identify the timing of potential triggers from sound mixtures. Altogether, our results suggest that acoustic features – spectrotemporal modulations, in particular – can practically be used to characterize the individualized patterns of aversion in participants with misophonia. Future perceptual studies using synthetic sounds and sound sets with more diverse acoustics will allow model predictions to be tested more broadly; however, sound-computable models may already have applications in precision diagnosis and management of misophonia.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"470 ","pages":"Article 109478"},"PeriodicalIF":2.5,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145749655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-15DOI: 10.1016/j.heares.2025.109476
Irina Wils, Tobias Mair, Ivo Dobrev, Christof Röösli
Effective mechanical coupling between bone conduction implants and the skull is essential for optimal auditory stimulation. The Bonebridge™ (MED-EL, Innsbruck, Austria) uses cortical screws to secure the transducer to the mastoid bone. However, the biomechanical consequences of varying the attachment quality of the screws on cochlear activation remain unclear.
Five Bonebridge implantations were performed, and five coupling conditions were evaluated: (1) both screws fixed, (2) only the superior screw fixed, (3) only the inferior screw fixed, (4) skin tension without screws, and (5) two screws, but neither fixed tightly. Cochlear activation was assessed by measuring pressure changes within the scala vestibuli and scala tympani using custom intracochlear acoustic receivers. Additionally, 3D velocity measurements via laser Doppler vibrometry were performed at multiple locations: the Bonebridge surface (4 points), the skull surface (4 points), and the cochlear promontory (1 point). A 1D scanning vibrometer further recorded vibratory responses on the implant and skull surfaces around the implant (450-500 points).
Intracochlear pressure, promontory, and skull motion decrease when screws are removed, especially the top screw, with reductions of 5–25 dB above 3 kHz. Total harmonic distortions occur mostly for frequencies below 2 kHz, and increase from 2 % when both screws are fixed to 18 % when neither of the screws is fixed.
The reduction of screw fixation weakens mechanical coupling, leading to reduced cochlear activation and worse audio quality. Thus, suboptimal coupling strategies may diminish auditory benefit for Bonebridge recipients, and during surgery, a firm attachment of both screws should be aimed for.
{"title":"Influence of Bonebridge screw fixation on the cochlear activation","authors":"Irina Wils, Tobias Mair, Ivo Dobrev, Christof Röösli","doi":"10.1016/j.heares.2025.109476","DOIUrl":"10.1016/j.heares.2025.109476","url":null,"abstract":"<div><div>Effective mechanical coupling between bone conduction implants and the skull is essential for optimal auditory stimulation. The Bonebridge™ (MED-EL, Innsbruck, Austria) uses cortical screws to secure the transducer to the mastoid bone. However, the biomechanical consequences of varying the attachment quality of the screws on cochlear activation remain unclear.</div><div>Five Bonebridge implantations were performed, and five coupling conditions were evaluated: (1) both screws fixed, (2) only the superior screw fixed, (3) only the inferior screw fixed, (4) skin tension without screws, and (5) two screws, but neither fixed tightly. Cochlear activation was assessed by measuring pressure changes within the scala vestibuli and scala tympani using custom intracochlear acoustic receivers. Additionally, 3D velocity measurements via laser Doppler vibrometry were performed at multiple locations: the Bonebridge surface (4 points), the skull surface (4 points), and the cochlear promontory (1 point). A 1D scanning vibrometer further recorded vibratory responses on the implant and skull surfaces around the implant (450-500 points).</div><div>Intracochlear pressure, promontory, and skull motion decrease when screws are removed, especially the top screw, with reductions of 5–25 dB above 3 kHz. Total harmonic distortions occur mostly for frequencies below 2 kHz, and increase from 2 % when both screws are fixed to 18 % when neither of the screws is fixed.</div><div>The reduction of screw fixation weakens mechanical coupling, leading to reduced cochlear activation and worse audio quality. Thus, suboptimal coupling strategies may diminish auditory benefit for Bonebridge recipients, and during surgery, a firm attachment of both screws should be aimed for.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109476"},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145573416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-15DOI: 10.1016/j.heares.2025.109475
Jessica Ky-Lee Choong , Dongcheng Zhang , Amy Judith Hampson , Ellie Cho , Amanda Edgley , Tayla Razmovski , Stephen John O’Leary , Kate Maree Brody
Introduction
The tissue response to cochlear implantation is thought to be involved in the loss of residual hearing, and its extent has been found to affect speech perception. This study investigated tissue response formation and maturation, focusing on both wound contraction using immunofluorescence markers and collagen deposition by multiphoton microscopy, in the implanted guinea pig cochlea.
Methods
Cochlear implantation was undertaken in a total of thirty-five guinea pigs, and the cochleae were harvested at four time-points: 1-, 14-, 28- and 84-days post-implantation. Ten animals were designated as controls. The tissue response morphology and collagen deposition were investigated by multiphoton microscopy. The wound contraction was studied using immunofluorescence markers of SMA and vimentin.
Results
Tissue response was located in the scala tympani’s basal-turn, between the lateral wall and basilar membrane. Its area didn’t change; however, its morphology and contents changed overtime after cochlear implantation within the sampled basal-turn field. At the proliferative phase of wound healing (14-days), an increased SMA was observed, with a “bubble-like” space formed between the basilar membrane and tissue response. During the early remodelling phase (28-days), there was increased vimentin expression and collagen deposition within the middle of the tissue response. By the later remodelling phase (84-days), vimentin expression reduced, but collagen density continued to increase.
Conclusion
Cochlear implantation triggers proliferation of the fibroblasts involved in wound contraction; this process is completed by three months. In contrast, collagen density continued to increase during the three-month window indicative of scar formation which may precede bone growth.
{"title":"Tissue response formation and maturation after cochlear implantation","authors":"Jessica Ky-Lee Choong , Dongcheng Zhang , Amy Judith Hampson , Ellie Cho , Amanda Edgley , Tayla Razmovski , Stephen John O’Leary , Kate Maree Brody","doi":"10.1016/j.heares.2025.109475","DOIUrl":"10.1016/j.heares.2025.109475","url":null,"abstract":"<div><h3>Introduction</h3><div>The tissue response to cochlear implantation is thought to be involved in the loss of residual hearing, and its extent has been found to affect speech perception. This study investigated tissue response formation and maturation, focusing on both wound contraction using immunofluorescence markers and collagen deposition by multiphoton microscopy, in the implanted guinea pig cochlea.</div></div><div><h3>Methods</h3><div>Cochlear implantation was undertaken in a total of thirty-five guinea pigs, and the cochleae were harvested at four time-points: 1-, 14-, 28- and 84-days post-implantation. Ten animals were designated as controls. The tissue response morphology and collagen deposition were investigated by multiphoton microscopy. The wound contraction was studied using immunofluorescence markers of SMA and vimentin.</div></div><div><h3>Results</h3><div>Tissue response was located in the scala tympani’s basal-turn, between the lateral wall and basilar membrane. Its area didn’t change; however, its morphology and contents changed overtime after cochlear implantation within the sampled basal-turn field. At the proliferative phase of wound healing (14-days), an increased SMA was observed, with a “bubble-like” space formed between the basilar membrane and tissue response. During the early remodelling phase (28-days), there was increased vimentin expression and collagen deposition within the middle of the tissue response. By the later remodelling phase (84-days), vimentin expression reduced, but collagen density continued to increase.</div></div><div><h3>Conclusion</h3><div>Cochlear implantation triggers proliferation of the fibroblasts involved in wound contraction; this process is completed by three months. In contrast, collagen density continued to increase during the three-month window indicative of scar formation which may precede bone growth.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109475"},"PeriodicalIF":2.5,"publicationDate":"2025-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145603598","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-14DOI: 10.1016/j.heares.2025.109474
Daniel O.J. Reijntjes , Kali Burke , Srijita Paul , Ulrich Mueller , Elisabeth Glowatzki , Amanda M. Lauer
The inner hair cells (IHCs) in the inner ear form synapses with auditory nerve fibers (ANFs) that send sound signals to the brain. ANFs have been grouped by their level of spontaneous firing rates (SRs) into high-, medium-, and low-SR ANFs. Based on their molecular profiles evaluated by RNAseq experiments, ANFs have been divided into three groups (1A, 1B, and 1C) that likely correspond to high-, medium-, and low-SR ANFs, respectively. In guinea pigs, the synapses between IHCs and low-SR ANFs have been shown to be more vulnerable to noise exposure compared to other ANF subtypes, but not in a study performed in CBA/CaJ mice, questioning if these results can be generalized. Here, an LYPD1 reporter mouse model on a C57Bl/6J background with specifically labeled group 1C, low-SR ANFs was used to examine whether LYPD1 positive ANF synapses are more vulnerable to noise exposure. Six-week-old mice were exposed to an 8–16 kHz octave band noise presented at 100 dB-SPL for 2 h. One week later, cochlear tissue was harvested to quantify ANF synapses and compare the percentage of LYPD1 positive ANF synapses in noise-exposed and unexposed animals. Auditory brainstem response measurements were performed to assess hearing function after noise exposure. The number of all ANF synapses and the percentage of LYPD1-positive ANF synapses were reduced following noise exposure, concurrent with increased ABR thresholds and decreased ABR wave 1 amplitudes. The reduction in the percentage of LYPD1-positive ANF synapses specifically indicates greater vulnerability of LYPD1 positive ANF synapses to noise exposure compared to other ANFs in C57Bl/6J mice.
{"title":"Increased vulnerability to noise exposure of low spontaneous rate type 1C spiral ganglion neuron synapses with inner hair cells","authors":"Daniel O.J. Reijntjes , Kali Burke , Srijita Paul , Ulrich Mueller , Elisabeth Glowatzki , Amanda M. Lauer","doi":"10.1016/j.heares.2025.109474","DOIUrl":"10.1016/j.heares.2025.109474","url":null,"abstract":"<div><div>The inner hair cells (IHCs) in the inner ear form synapses with auditory nerve fibers (ANFs) that send sound signals to the brain. ANFs have been grouped by their level of spontaneous firing rates (SRs) into high-, medium-, and low-SR ANFs. Based on their molecular profiles evaluated by RNAseq experiments, ANFs have been divided into three groups (1A, 1B, and 1C) that likely correspond to high-, medium-, and low-SR ANFs, respectively. In guinea pigs, the synapses between IHCs and low-SR ANFs have been shown to be more vulnerable to noise exposure compared to other ANF subtypes, but not in a study performed in CBA/CaJ mice, questioning if these results can be generalized. Here, an <em>LYPD1</em> reporter mouse model on a C57Bl/6J background with specifically labeled group 1C, low-SR ANFs was used to examine whether <em>LYPD1</em> positive ANF synapses are more vulnerable to noise exposure. Six-week-old mice were exposed to an 8–16 kHz octave band noise presented at 100 dB-SPL for 2 h. One week later, cochlear tissue was harvested to quantify ANF synapses and compare the percentage of <em>LYPD1</em> positive ANF synapses in noise-exposed and unexposed animals. Auditory brainstem response measurements were performed to assess hearing function after noise exposure. The number of all ANF synapses and the percentage of <em>LYPD1-</em>positive ANF synapses were reduced following noise exposure, concurrent with increased ABR thresholds and decreased ABR wave 1 amplitudes. The reduction in the percentage of <em>LYPD1-</em>positive ANF synapses specifically indicates greater vulnerability of <em>LYPD1</em> positive ANF synapses to noise exposure compared to other ANFs in C57Bl/6J mice.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109474"},"PeriodicalIF":2.5,"publicationDate":"2025-11-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145573346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-13DOI: 10.1016/j.heares.2025.109473
Katherine Adcock , Elva Arulchelvan , Nathan Shields , Sven Vanneste
Advances in artificial intelligence, particularly machine learning and deep learning, in conjunction with the rise of personalised medicine, can facilitate tailored decision-making for diagnoses, prognoses, and treatment responses based on individual patient data. The multifaceted nature of symptoms and disorders in (neuro)otology, with their diverse aetiologies and subjective characteristics, makes this field an ideal candidate for computational personalised medicine. This narrative review critically synthesises applications of machine learning and deep learning in otology and neurotology published between 2013 and 2025. Relevant studies were identified through targeted searches of PubMed, Scopus, and Google Scholar using combinations of terms related to artificial intelligence, tinnitus, cochlear implants, and otologic or neurotologic disorders. Only peer-reviewed articles focusing on human applications of machine learning or deep learning in these fields were included, excluding theoretical papers or animal studies. Recent breakthroughs, such as the Whisper speech recognition model for cochlear implant simulations and large language models for refining tinnitus subgroup identification and therapy predictions, underscore the transformative potential of AI in improving clinical outcomes. This review is distinct in its emphasis on these emerging technologies and their integration into multimodal datasets, combining imaging, audiometric data, and patient-reported outcomes to refine diagnosis and treatment approaches. However, challenges including the lack of standardisation, limited generalisability of models, and the need for improved frameworks for multimodal data integration impede rigorous and reproducible implementation, topics that are critically explored in this review. Here, we explore the applications of machine learning, deep learning, and large language models in tinnitus, cochlear implants, and (neuro)tology, providing a critical analysis of recent advancements, persistent challenges, and recommendations for future research. By addressing these challenges and implementing recommended strategies, this review outlines a pathway for integrating cutting-edge artificial intelligence tools into clinical practice, underscoring their immense potential to revolutionise precision medicine in otology and neurotology and improve patient outcomes.
{"title":"Towards precision medicine for otology and neurotology: Machine learning applications and challenges","authors":"Katherine Adcock , Elva Arulchelvan , Nathan Shields , Sven Vanneste","doi":"10.1016/j.heares.2025.109473","DOIUrl":"10.1016/j.heares.2025.109473","url":null,"abstract":"<div><div>Advances in artificial intelligence, particularly machine learning and deep learning, in conjunction with the rise of personalised medicine, can facilitate tailored decision-making for diagnoses, prognoses, and treatment responses based on individual patient data. The multifaceted nature of symptoms and disorders in (neuro)otology, with their diverse aetiologies and subjective characteristics, makes this field an ideal candidate for computational personalised medicine. This narrative review critically synthesises applications of machine learning and deep learning in otology and neurotology published between 2013 and 2025. Relevant studies were identified through targeted searches of PubMed, Scopus, and Google Scholar using combinations of terms related to artificial intelligence, tinnitus, cochlear implants, and otologic or neurotologic disorders. Only peer-reviewed articles focusing on human applications of machine learning or deep learning in these fields were included, excluding theoretical papers or animal studies. Recent breakthroughs, such as the Whisper speech recognition model for cochlear implant simulations and large language models for refining tinnitus subgroup identification and therapy predictions, underscore the transformative potential of AI in improving clinical outcomes. This review is distinct in its emphasis on these emerging technologies and their integration into multimodal datasets, combining imaging, audiometric data, and patient-reported outcomes to refine diagnosis and treatment approaches. However, challenges including the lack of standardisation, limited generalisability of models, and the need for improved frameworks for multimodal data integration impede rigorous and reproducible implementation, topics that are critically explored in this review. Here, we explore the applications of machine learning, deep learning, and large language models in tinnitus, cochlear implants, and (neuro)tology, providing a critical analysis of recent advancements, persistent challenges, and recommendations for future research. By addressing these challenges and implementing recommended strategies, this review outlines a pathway for integrating cutting-edge artificial intelligence tools into clinical practice, underscoring their immense potential to revolutionise precision medicine in otology and neurotology and improve patient outcomes.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109473"},"PeriodicalIF":2.5,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145578112","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-13DOI: 10.1016/j.heares.2025.109472
Ava Schwartz , Grace Pulliam , Jacob I. Feldman , Kacie Dunham-Carr , S. Madison Clark , Kelsea McClurkin , Carissa J. Cascio , Bahar Keçeli-Kaysılı , Tiffany Woynaroski
A growing body of research has shown that decreased sound tolerance (DST) is highly prevalent and impacts the mental health of affected individuals. Recent work has shown this is especially true for autistic individuals. The extant literature has been limited, however, by a focus on DST relatively late in life. Consequently, at present we know little about when and how DST emerges and produces cascading effects on mental health. In this pilot study, we prospectively followed infants at high likelihood for autism, and thus hypothetically for DST, based on their status as younger siblings of autistic children (Sibs-Autism) and infants at lower, general population-level likelihood for these conditions (Sibs-NA) to determine whether (a) DST symptomatology differs based on autism likelihood status and/or diagnostic outcome; (b) indices of early resting brain states, specifically gamma power, predict sensory hyperresponsiveness and DST; (c) sensory hyperresponsiveness predicts DST symptomatology; (d) DST symptomatology predicts anxiety; and (e) if the aforementioned associations vary by familial likelihood for autism or later autism status. Preliminary results indicate that DST symptoms are elevated in Sibs-Autism, particularly those who go on to receive a diagnosis of autism, relative to Sibs-NA. Gamma power is not significantly associated with later sensory hyperresponsiveness or DST, but hyperresponsiveness is associated with later DST, which is associated with later anxiety in Sibs-Autism.
{"title":"Conceptualizing the substrates and sequelae of decreased sound tolerance as a developmental cascade: A pilot study","authors":"Ava Schwartz , Grace Pulliam , Jacob I. Feldman , Kacie Dunham-Carr , S. Madison Clark , Kelsea McClurkin , Carissa J. Cascio , Bahar Keçeli-Kaysılı , Tiffany Woynaroski","doi":"10.1016/j.heares.2025.109472","DOIUrl":"10.1016/j.heares.2025.109472","url":null,"abstract":"<div><div>A growing body of research has shown that decreased sound tolerance (DST) is highly prevalent and impacts the mental health of affected individuals. Recent work has shown this is especially true for autistic individuals. The extant literature has been limited, however, by a focus on DST relatively late in life. Consequently, at present we know little about when and how DST emerges and produces cascading effects on mental health. In this pilot study, we prospectively followed infants at high likelihood for autism, and thus hypothetically for DST, based on their status as younger siblings of autistic children (Sibs-Autism) and infants at lower, general population-level likelihood for these conditions (Sibs-NA) to determine whether (a) DST symptomatology differs based on autism likelihood status and/or diagnostic outcome; (b) indices of early resting brain states, specifically gamma power, predict sensory hyperresponsiveness and DST; (c) sensory hyperresponsiveness predicts DST symptomatology; (d) DST symptomatology predicts anxiety; and (e) if the aforementioned associations vary by familial likelihood for autism or later autism status. Preliminary results indicate that DST symptoms are elevated in Sibs-Autism, particularly those who go on to receive a diagnosis of autism, relative to Sibs-NA. Gamma power is not significantly associated with later sensory hyperresponsiveness or DST, but hyperresponsiveness is associated with later DST, which is associated with later anxiety in Sibs-Autism.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109472"},"PeriodicalIF":2.5,"publicationDate":"2025-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145563788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-12DOI: 10.1016/j.heares.2025.109471
Akil Turner , Bo Ding , Xiaoxia Zhu , Tam Nguyen , Parveen Bazard , Robert D. Frisina
Age-related hearing loss (ARHL) is a highly prevalent sensory neurodegenerative disorder that involves various molecular mechanisms. The present study investigated if age-related oxidative stress (OS) induces alterations in mitochondrial DNA (mtDNA) copy number and heteroplasmy, using complementary in vivo and in vitro models. Aged (30-month) CBA/CaJ mice have elevated auditory brainstem response (ABR) thresholds amplitudes vs. young (3-month) CBA/CaJ mice, a key physiological characteristic of ARHL The in vitro experiments employed the strial SV-K1 cochlear cell line treated with hydrogen peroxide (H2O2). Both the aged cochleae and H2O2-treated cells exhibited a significant (∼50 %) reduction in mtDNA copy number compared to their respective controls with enhanced levels of malondialdehyde (MDA). Therefore, in both the in vivo and in vitro models, OS drove mtDNA depletion. High-depth sequencing employing nuclear mtDNA pseudogene (NUMT)-avoidant methodologies revealed non-random distribution of heteroplasmic variants. Mutation hotspots were identified within the mitochondrial genome, particularly in regions encoding cytochrome c oxidase subunit 1 (COX1) and cytochrome b (CYTB) in both models. While H2O2 treatment induced a more widespread expansion of low-frequency variants, aging mice primarily showed shifts in the allele frequency distribution of existing variants rather than an accumulation of novel mutations. These findings demonstrate that OS is also a key factor of mtDNA regional mutational burden in the aging cochlea. The parallel mtDNA alterations observed in aged tissues and OS cells underscore mitochondrial genomic instability as a central mechanism in ARHL, highlighting potential targets for interventions aimed at preserving mitochondrial integrity and therefore auditory function.
{"title":"Age-related hearing loss involves mitochondrial DNA instability and copy number depletion in the cochlea: Insights from in vivo and in vitro models","authors":"Akil Turner , Bo Ding , Xiaoxia Zhu , Tam Nguyen , Parveen Bazard , Robert D. Frisina","doi":"10.1016/j.heares.2025.109471","DOIUrl":"10.1016/j.heares.2025.109471","url":null,"abstract":"<div><div>Age-related hearing loss (ARHL) is a highly prevalent sensory neurodegenerative disorder that involves various molecular mechanisms. The present study investigated if age-related oxidative stress (OS) induces alterations in mitochondrial DNA (mtDNA) copy number and heteroplasmy, using complementary <em>in vivo</em> and <em>in vitro</em> models. Aged (30-month) CBA/CaJ mice have elevated auditory brainstem response (ABR) thresholds amplitudes <em>vs.</em> young (3-month) CBA/CaJ mice, a key physiological characteristic of ARHL The <em>in vitro</em> experiments employed the strial SV-K1 cochlear cell line treated with hydrogen peroxide (H<sub>2</sub>O<sub>2</sub>). Both the aged cochleae and H<sub>2</sub>O<sub>2</sub>-treated cells exhibited a significant (∼50 %) reduction in mtDNA copy number compared to their respective controls with enhanced levels of malondialdehyde (MDA). Therefore, in both the <em>in vivo</em> and <em>in vitro</em> models, OS drove mtDNA depletion. High-depth sequencing employing nuclear mtDNA pseudogene (NUMT)-avoidant methodologies revealed non-random distribution of heteroplasmic variants. Mutation hotspots were identified within the mitochondrial genome, particularly in regions encoding cytochrome c oxidase subunit 1 (COX1) and cytochrome b (CYTB) in both models. While H<sub>2</sub>O<sub>2</sub> treatment induced a more widespread expansion of low-frequency variants, aging mice primarily showed shifts in the allele frequency distribution of existing variants rather than an accumulation of novel mutations. These findings demonstrate that OS is also a key factor of mtDNA regional mutational burden in the aging cochlea. The parallel mtDNA alterations observed in aged tissues and OS cells underscore mitochondrial genomic instability as a central mechanism in ARHL, highlighting potential targets for interventions aimed at preserving mitochondrial integrity and therefore auditory function.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109471"},"PeriodicalIF":2.5,"publicationDate":"2025-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145621689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-11DOI: 10.1016/j.heares.2025.109469
Jacob de Nobel , Jeroen J. Briaire , Thomas H.W. Bäck , Anna V. Kononova , Johan H.M. Frijns
We present NeuroVoc, a flexible model-agnostic vocoder framework that reconstructs acoustic waveforms from simulated neural activity patterns using an inverse Fourier transform. The system applies straightforward signal processing to neurogram representations, time–frequency binned outputs from auditory nerve fiber models. Crucially, the model architecture is modular, allowing for easy substitution or modification of the underlying auditory models. This flexibility eliminates the need for speech-coding-strategy-specific vocoder implementations when simulating auditory perception in cochlear implant (CI) users. It also allows direct comparisons between normal hearing (NH) and electrical hearing (EH) models, as demonstrated in this study. The vocoder preserves distinctive features of each model; for example, the NH model retains harmonic structure more faithfully than the EH model. We evaluated perceptual intelligibility in noise using an online Digits-in-Noise (DIN) test, where participants completed three test conditions: one with standard speech, and two with vocoded speech using the NH and EH models. Both the standard DIN test and the EH-vocoded groups were statistically equivalent to clinically reported data for NH and CI listeners. On average, the NH and EH vocoded groups increased SRT compared to the standard test by 2.4 dB and 7.1 dB, respectively. These findings show that, although some degradation occurs, the vocoder can reconstruct intelligible speech under both hearing models and accurately reflects the reduced speech-in-noise performance experienced by CI users.
{"title":"From spikes to speech: NeuroVoc — A biologically plausible vocoder framework for auditory perception and cochlear implant simulation","authors":"Jacob de Nobel , Jeroen J. Briaire , Thomas H.W. Bäck , Anna V. Kononova , Johan H.M. Frijns","doi":"10.1016/j.heares.2025.109469","DOIUrl":"10.1016/j.heares.2025.109469","url":null,"abstract":"<div><div>We present <em>NeuroVoc</em>, a flexible model-agnostic vocoder framework that reconstructs acoustic waveforms from simulated neural activity patterns using an inverse Fourier transform. The system applies straightforward signal processing to neurogram representations, time–frequency binned outputs from auditory nerve fiber models. Crucially, the model architecture is modular, allowing for easy substitution or modification of the underlying auditory models. This flexibility eliminates the need for speech-coding-strategy-specific vocoder implementations when simulating auditory perception in cochlear implant (CI) users. It also allows direct comparisons between normal hearing (NH) and electrical hearing (EH) models, as demonstrated in this study. The vocoder preserves distinctive features of each model; for example, the NH model retains harmonic structure more faithfully than the EH model. We evaluated perceptual intelligibility in noise using an online Digits-in-Noise (DIN) test, where participants completed three test conditions: one with standard speech, and two with vocoded speech using the NH and EH models. Both the standard DIN test and the EH-vocoded groups were statistically equivalent to clinically reported data for NH and CI listeners. On average, the NH and EH vocoded groups increased SRT compared to the standard test by 2.4 dB and 7.1 dB, respectively. These findings show that, although some degradation occurs, the vocoder can reconstruct intelligible speech under both hearing models and accurately reflects the reduced speech-in-noise performance experienced by CI users.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"469 ","pages":"Article 109469"},"PeriodicalIF":2.5,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145500385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-11-08DOI: 10.1016/j.heares.2025.109470
Sara Cacciato-Salcedo , Ana B. Lao-Rodríguez , Manuel S. Malmierca
Prenatal exposure to valproic acid (VPA) provides a well-established rodent model of autism, yet its effects on auditory brainstem/midbrain processing across sex and development remain elusive. We recorded click-evoked auditory brainstem responses (ABRs) in Long–Evans rats that received prenatal VPA (400 mg/kg, gestational day 12) and in matched controls at prepubertal (postnatal days 30–45) and adult (65–120) stages under urethane anesthesia. We analyzed peak amplitudes, latencies, inter-peak intervals, and amplitude ratios across sound levels. Auditory thresholds remained comparable among groups. In controls, females showed larger amplitudes for waves I–II, shorter latencies for waves I, II, and IV, and steeper amplitude–intensity slopes for waves II, III, and V than males, indicating stronger level-dependent recruitment. Maturation enhanced early brainstem and midbrain responses by increasing amplitude growth (wave II) and shortening latencies (waves II–V), with effects more pronounced in females. Prenatal VPA exposure reduced wave II amplitude and delayed early peaks (I–III) in females, accompanied by elevated amplitude ratios, whereas in males it mainly affected later responses by reducing amplitudes for waves III–V and prolonging inter-peak latencies (I–III, III–V). These findings show that sex, age, and prenatal VPA exposure distinctly shape auditory brainstem/midbrain function.
{"title":"Sex- and age-specific effects on auditory brainstem responses in the valproic acid-induced rat model of autism","authors":"Sara Cacciato-Salcedo , Ana B. Lao-Rodríguez , Manuel S. Malmierca","doi":"10.1016/j.heares.2025.109470","DOIUrl":"10.1016/j.heares.2025.109470","url":null,"abstract":"<div><div>Prenatal exposure to valproic acid (VPA) provides a well-established rodent model of autism, yet its effects on auditory brainstem/midbrain processing across sex and development remain elusive. We recorded click-evoked auditory brainstem responses (ABRs) in Long–Evans rats that received prenatal VPA (400 mg/kg, gestational day 12) and in matched controls at prepubertal (postnatal days 30–45) and adult (65–120) stages under urethane anesthesia. We analyzed peak amplitudes, latencies, inter-peak intervals, and amplitude ratios across sound levels. Auditory thresholds remained comparable among groups. In controls, females showed larger amplitudes for waves I–II, shorter latencies for waves I, II, and IV, and steeper amplitude–intensity slopes for waves II, III, and V than males, indicating stronger level-dependent recruitment. Maturation enhanced early brainstem and midbrain responses by increasing amplitude growth (wave II) and shortening latencies (waves II–V), with effects more pronounced in females. Prenatal VPA exposure reduced wave II amplitude and delayed early peaks (I–III) in females, accompanied by elevated amplitude ratios, whereas in males it mainly affected later responses by reducing amplitudes for waves III–V and prolonging inter-peak latencies (I–III, III–V). These findings show that sex, age, and prenatal VPA exposure distinctly shape auditory brainstem/midbrain function.</div></div>","PeriodicalId":12881,"journal":{"name":"Hearing Research","volume":"468 ","pages":"Article 109470"},"PeriodicalIF":2.5,"publicationDate":"2025-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145512225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}