Background
Sleep breathing-sound analysis offers a non-contact option for Apnea–Hypopnea Index (AHI) estimation and obstructive sleep apnea–hypopnea syndrome (OSAHS) screening. However, labels for audio segments are typically assigned by automatic alignment with polysomnography (PSG) annotations, which can introduce label noise around apnea–hypopnea events and degrade AHI estimation performance. This work proposes an AHI estimation framework that explicitly corrects noisy labels in large-scale breathing-sound datasets.
Methods
Whole-night sleep breathing sounds from the PSG-Audio dataset were divided into fixed-length segments and automatically labeled according to PSG annotations. An ensemble noisy-label classifier based on three ConvNeXt variants was trained to identify and correct mislabeled labels. The corrected labels were then used to train a lightweight model that combines ConvNeXt with Long Short-Term Memory (LSTM) for apnea–hypopnea event detection. Night-level prediction summaries were then mapped to AHI using a robust RANSAC linear regression model.
Results
Approximately 8% of the audio segments had their labels corrected. On a subject-independent test set of 50 subjects, training with corrected labels improved event-detection accuracy by 4.99% and F1-score by 2.3% compared with the raw-label baseline. The estimated AHI achieved a Pearson correlation coefficient of 0.85 with AHI from PSG. For severe OSAHS screening, the system achieved 0.94 sensitivity and 0.86 specificity.
Conclusions
Explicit label-noise correction improves fully non-contact AHI estimation from breathing sounds without additional sensors or substantially increased complexity. The proposed framework supports scalable AHI-based screening and triage and motivates prospective validation in diverse home settings.
扫码关注我们
求助内容:
应助结果提醒方式:
