首页 > 最新文献

IEEE Transactions on Biomedical Engineering最新文献

英文 中文
The Effect of von Willebrand Disease on Platelet Adhesion Dynamics: Correlating a Multiscale Platelet Model to In Vitro Results. 血管性血友病对血小板粘附动力学的影响:多尺度血小板模型与体外结果的相关性
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-26 DOI: 10.1109/TBME.2026.3658253
Peineng Wang, Jawaad Sheriff, Yuefan Deng, Danny Bluestein

Objective: Von Willebrand Disease (VWD), the most common inherited bleeding disorder affecting 0.1% to 1% of the population, causes extensive mucocutaneous bleeding across various clinical contexts. Von Willebrand Factor (vWF) plays a critical role in hemostasis by mediating platelet adhesion under high shear stress conditions. We simulated platelet-vWF interactions to investigate adhesion dynamics in VWD using a multiscale modeling approach combining Dissipative Particle Dynamics (DPD) and Coarse-Grained Molecular Dynamics (CGMD).

Methods: Our platelet model provides high-resolution insights into adhesion mechanics by representing the platelet as a complex, deformable cellular entity comprising intricate membrane and subcellular components that capture the nuanced biomechanical behavior of platelets under flow conditions.

Results: Simulations under 30 dyne/cm2 shear stress revealed a threshold effect: platelets failed to complete flipping and adhesion below 40% vWF density, mirroring Type 1 VWD clinical manifestations. We identified asymmetric platelet flipping dynamics with longer lift-off periods compared to reattachment periods, and revealed a distinct temporal lag between the platelet's vertical positioning and minimum bond force/contact area configurations. In vitro experiments supported these computational findings, demonstrating a significant reduction in platelet residence duration and translocation distance as vWF surface densities decreased.

Conclusions: This work provides quantitative insights into the molecular mechanisms underlying platelet adhesion in VWD through our advanced CGMD model.

Significance: Our findings establish a comprehensive framework for understanding cellular adhesion processes in biofluid environments, potentially informing therapeutic strategies for bleeding disorders and thrombotic conditions.

目的:血管性血友病(VWD)是最常见的遗传性出血性疾病,影响0.1%至1%的人口,在各种临床情况下导致广泛的粘膜皮肤出血。血管性血友病因子(vWF)在高剪切应力条件下通过介导血小板粘附在止血中起关键作用。我们模拟血小板- vwf相互作用,使用多尺度建模方法结合耗散粒子动力学(DPD)和粗粒度分子动力学(CGMD)来研究VWD中的粘附动力学。方法:我们的血小板模型通过将血小板表示为一个复杂的、可变形的细胞实体,包括复杂的膜和亚细胞成分,捕捉血小板在流动条件下细微的生物力学行为,从而提供了对粘附力学的高分辨率见解。结果:30 dyne/cm2剪切应力下的模拟显示阈值效应:血小板在40% vWF密度下无法完成翻转和粘附,反映了1型VWD的临床表现。我们发现了不对称的血小板翻转动力学,与再附着周期相比,其上升周期更长,并揭示了血小板垂直位置和最小结合力/接触面积配置之间存在明显的时间滞后。体外实验支持这些计算结果,表明血小板停留时间和易位距离随着vWF表面密度的降低而显著减少。结论:通过我们先进的CGMD模型,这项工作为VWD中血小板粘附的分子机制提供了定量的见解。意义:我们的研究结果为理解生物流体环境中的细胞粘附过程建立了一个全面的框架,可能为出血性疾病和血栓性疾病的治疗策略提供信息。
{"title":"The Effect of von Willebrand Disease on Platelet Adhesion Dynamics: Correlating a Multiscale Platelet Model to In Vitro Results.","authors":"Peineng Wang, Jawaad Sheriff, Yuefan Deng, Danny Bluestein","doi":"10.1109/TBME.2026.3658253","DOIUrl":"https://doi.org/10.1109/TBME.2026.3658253","url":null,"abstract":"<p><strong>Objective: </strong>Von Willebrand Disease (VWD), the most common inherited bleeding disorder affecting 0.1% to 1% of the population, causes extensive mucocutaneous bleeding across various clinical contexts. Von Willebrand Factor (vWF) plays a critical role in hemostasis by mediating platelet adhesion under high shear stress conditions. We simulated platelet-vWF interactions to investigate adhesion dynamics in VWD using a multiscale modeling approach combining Dissipative Particle Dynamics (DPD) and Coarse-Grained Molecular Dynamics (CGMD).</p><p><strong>Methods: </strong>Our platelet model provides high-resolution insights into adhesion mechanics by representing the platelet as a complex, deformable cellular entity comprising intricate membrane and subcellular components that capture the nuanced biomechanical behavior of platelets under flow conditions.</p><p><strong>Results: </strong>Simulations under 30 dyne/cm2 shear stress revealed a threshold effect: platelets failed to complete flipping and adhesion below 40% vWF density, mirroring Type 1 VWD clinical manifestations. We identified asymmetric platelet flipping dynamics with longer lift-off periods compared to reattachment periods, and revealed a distinct temporal lag between the platelet's vertical positioning and minimum bond force/contact area configurations. In vitro experiments supported these computational findings, demonstrating a significant reduction in platelet residence duration and translocation distance as vWF surface densities decreased.</p><p><strong>Conclusions: </strong>This work provides quantitative insights into the molecular mechanisms underlying platelet adhesion in VWD through our advanced CGMD model.</p><p><strong>Significance: </strong>Our findings establish a comprehensive framework for understanding cellular adhesion processes in biofluid environments, potentially informing therapeutic strategies for bleeding disorders and thrombotic conditions.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146051637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-channel Electromagnetic Interference Elimination for Shielding-free MRI Using Null Operations. 零操作消除无屏蔽MRI多通道电磁干扰。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-23 DOI: 10.1109/TBME.2026.3656493
Jiaqi Wang, Huifang Wang, Linfang Xiao, Mengye Lyu, Yujiao Zhao, Yilong Liu, Ed X Wu

Objective: Emerging technologies for electromagnetic interference (EMI) elimination have enabled radio frequency (RF) shielding-free magnetic resonance imaging (MRI), significantly reduced costs and increased accessibility. Existing methods often rely on multiple external sensors for EMI elimination, which can degrade with fewer sensors. Our goal is to develop a method that robustly eliminates EMI with fewer or no sensors.

Methods: We propose a method for multi-channel electromagnetic interference elimination in shielding-free MRI using null operations (MEENO). This approach fully exploits the inter-channel correlation across all RF receiving and EMI sensing channels. The method was comprehensively evaluated through simulation studies and human brain imaging.

Results: The MEENO approach effectively eliminates EMI artifacts, outperforming existing methods, particularly with a limited number of sensors. It shows superior performance in terms of signal-to-noise ratio and residual EMI levels.

Conclusion: We introduce a method for EMI elimination in multi-channel MRI using null operations, which fully leverages inter-channel correlation and surpasses existing approaches, especially with limited sensors.

Significance: This work offers a solution for EMI elimination with fewer or no external sensors, providing a more cost-effective and robust approach for shielding-free MRI.

目的:消除电磁干扰(EMI)的新兴技术使射频(RF)无屏蔽磁共振成像(MRI)成为可能,大大降低了成本,增加了可及性。现有的方法通常依赖于多个外部传感器来消除电磁干扰,传感器数量较少会导致电磁干扰性能下降。我们的目标是开发一种方法,以更少或没有传感器健壮地消除电磁干扰。方法:提出了一种利用零操作(MEENO)消除无屏蔽MRI多通道电磁干扰的方法。这种方法充分利用了所有射频接收和电磁干扰感知通道间的相关性。通过仿真研究和人脑成像对该方法进行了综合评价。结果:MEENO方法有效地消除了EMI伪影,优于现有方法,特别是在有限数量的传感器下。它在信噪比和残余EMI电平方面表现出优越的性能。结论:我们介绍了一种在多通道MRI中使用零操作消除电磁干扰的方法,该方法充分利用了通道间的相关性,超越了现有的方法,特别是在有限的传感器下。意义:这项工作提供了一种消除电磁干扰的解决方案,使用更少或没有外部传感器,为无屏蔽MRI提供了一种更具成本效益和强大的方法。
{"title":"Multi-channel Electromagnetic Interference Elimination for Shielding-free MRI Using Null Operations.","authors":"Jiaqi Wang, Huifang Wang, Linfang Xiao, Mengye Lyu, Yujiao Zhao, Yilong Liu, Ed X Wu","doi":"10.1109/TBME.2026.3656493","DOIUrl":"https://doi.org/10.1109/TBME.2026.3656493","url":null,"abstract":"<p><strong>Objective: </strong>Emerging technologies for electromagnetic interference (EMI) elimination have enabled radio frequency (RF) shielding-free magnetic resonance imaging (MRI), significantly reduced costs and increased accessibility. Existing methods often rely on multiple external sensors for EMI elimination, which can degrade with fewer sensors. Our goal is to develop a method that robustly eliminates EMI with fewer or no sensors.</p><p><strong>Methods: </strong>We propose a method for multi-channel electromagnetic interference elimination in shielding-free MRI using null operations (MEENO). This approach fully exploits the inter-channel correlation across all RF receiving and EMI sensing channels. The method was comprehensively evaluated through simulation studies and human brain imaging.</p><p><strong>Results: </strong>The MEENO approach effectively eliminates EMI artifacts, outperforming existing methods, particularly with a limited number of sensors. It shows superior performance in terms of signal-to-noise ratio and residual EMI levels.</p><p><strong>Conclusion: </strong>We introduce a method for EMI elimination in multi-channel MRI using null operations, which fully leverages inter-channel correlation and surpasses existing approaches, especially with limited sensors.</p><p><strong>Significance: </strong>This work offers a solution for EMI elimination with fewer or no external sensors, providing a more cost-effective and robust approach for shielding-free MRI.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146040548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain Knowledge is Power: Leveraging Physiological Priors for Self-Supervised Representation Learning in Electrocardiography. 领域知识就是力量:利用生理先验进行心电图的自我监督表示学习。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-22 DOI: 10.1109/TBME.2026.3656904
Nooshin Maghsoodi, Sarah Nassar, Paul F R Wilson, Minh Nguyen Nhat To, Sophia Mannina, Shamel Addas, Stephanie Sibley, David Pichora, David Maslove, Purang Abolmaesumi, Parvin Mousavi

Objective: Electrocardiograms (ECGs) play a crucial role in diagnosing heart conditions; however, the effectiveness of artificial intelligence (AI)-based ECG analysis is often hindered by the limited availability of labeled data. Self-supervised learning (SSL) can address this by leveraging large-scale unlabeled data. We introduce PhysioCLR (Physiology-aware Contrastive Learning Representation for ECG), a physiology-aware contrastive learning framework that incorporates domain-specific priors to enhance the generalizability and clinical relevance of ECG-based arrhythmia classification. Methods: During pretraining, PhysioCLR learns to bring together embeddings of samples that share similar clinically relevant features while pushing apart those that are dissimilar. Unlike existing methods, our method integrates ECG physiological similarity cues into contrastive learning, promoting the learning of clinically meaningful representations. Additionally, we introduce ECG-specific augmentations that preserve the ECG category post-augmentation and propose a hybrid loss function to further refine the quality of learned representations. Results: We evaluate PhysioCLR on two public ECG datasets, Chapman and Georgia, for multilabel ECG diagnoses, as well as a private ICU dataset labeled for binary classification. Across the Chapman, Georgia, and private cohorts, PhysioCLR boosts the mean AUROC by 12% relative to the strongest baseline, underscoring its robust cross-dataset generalization. Conclusion: By embedding physiological knowledge into contrastive learning, PhysioCLR enables the model to learn clinically meaningful and transferable ECG features.  Significance: PhysioCLR demonstrates the potential of physiology-informed SSL to offer a promising path toward more effective and label-efficient ECG diagnostics.

目的:心电图(ECGs)在心脏疾病诊断中的重要作用;然而,基于人工智能(AI)的心电图分析的有效性经常受到标记数据可用性有限的阻碍。自我监督学习(SSL)可以通过利用大规模未标记数据来解决这个问题。我们介绍了PhysioCLR (ECG的生理感知对比学习表征),这是一个生理感知对比学习框架,结合了特定领域的先验,以增强基于ECG的心律失常分类的泛化性和临床相关性。方法:在预训练期间,PhysioCLR学习将具有相似临床相关特征的样本嵌入在一起,同时将不相似的样本分开。与现有方法不同,我们的方法将ECG生理相似性线索整合到对比学习中,促进临床有意义表征的学习。此外,我们引入了保留ECG类别增强后的ECG特定增强,并提出了混合损失函数来进一步改进学习表征的质量。结果:我们在两个公共ECG数据集Chapman和Georgia上评估PhysioCLR,用于多标签ECG诊断,以及一个私有ICU数据集标记为二元分类。在查普曼,乔治亚州和私人队列中,PhysioCLR将平均AUROC提高了12%,相对于最强基线,强调了其强大的跨数据集泛化。结论:通过将生理学知识嵌入到对比学习中,PhysioCLR使模型能够学习具有临床意义和可转移的ECG特征。意义:PhysioCLR显示了生理信息SSL的潜力,为更有效和更高效的ECG诊断提供了一条有希望的途径。
{"title":"Domain Knowledge is Power: Leveraging Physiological Priors for Self-Supervised Representation Learning in Electrocardiography.","authors":"Nooshin Maghsoodi, Sarah Nassar, Paul F R Wilson, Minh Nguyen Nhat To, Sophia Mannina, Shamel Addas, Stephanie Sibley, David Pichora, David Maslove, Purang Abolmaesumi, Parvin Mousavi","doi":"10.1109/TBME.2026.3656904","DOIUrl":"https://doi.org/10.1109/TBME.2026.3656904","url":null,"abstract":"<p><strong>Objective: </strong>Electrocardiograms (ECGs) play a crucial role in diagnosing heart conditions; however, the effectiveness of artificial intelligence (AI)-based ECG analysis is often hindered by the limited availability of labeled data. Self-supervised learning (SSL) can address this by leveraging large-scale unlabeled data. We introduce PhysioCLR (Physiology-aware Contrastive Learning Representation for ECG), a physiology-aware contrastive learning framework that incorporates domain-specific priors to enhance the generalizability and clinical relevance of ECG-based arrhythmia classification. Methods: During pretraining, PhysioCLR learns to bring together embeddings of samples that share similar clinically relevant features while pushing apart those that are dissimilar. Unlike existing methods, our method integrates ECG physiological similarity cues into contrastive learning, promoting the learning of clinically meaningful representations. Additionally, we introduce ECG-specific augmentations that preserve the ECG category post-augmentation and propose a hybrid loss function to further refine the quality of learned representations. Results: We evaluate PhysioCLR on two public ECG datasets, Chapman and Georgia, for multilabel ECG diagnoses, as well as a private ICU dataset labeled for binary classification. Across the Chapman, Georgia, and private cohorts, PhysioCLR boosts the mean AUROC by 12% relative to the strongest baseline, underscoring its robust cross-dataset generalization. Conclusion: By embedding physiological knowledge into contrastive learning, PhysioCLR enables the model to learn clinically meaningful and transferable ECG features.  Significance: PhysioCLR demonstrates the potential of physiology-informed SSL to offer a promising path toward more effective and label-efficient ECG diagnostics.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146029425","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Double Banking on Knowledge: A Unified All-in-One Framework for Unpaired Multi-Modality Semi-supervised Medical Image Segmentation. 知识的双重银行:一种统一的多模态半监督医学图像分割框架。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-21 DOI: 10.1109/TBME.2026.3656540
Yingyu Chen, Ziyuan Yang, Zhongzhou Zhang, Ming Yan, Hui Yu, Yan Liu, Yi Zhang

Multi-modality (MM) semi-supervised learning (SSL) based medical image segmentation has recently gained increasing attention for its ability to utilize MM data and reduce reliance on labeled images. However, current methods face several challenges: (1) Multiple networks assignment hinder scalability to scenarios with more than two modalities. (2) Focusing solely on modality-invariant representation while neglecting modality-specific features, leads to incomplete MM learning. (3) Leveraging unlabeled data with generative methods can be unreliable for SSL. To address these problems, we propose a novel unified all-in-one framework for MM-SSLmedical image segmentation. To address challenge (1), we propose a modality all-in-one segmentation network based on standard U-Net architecture that accepts data from all modalities, removing the limitation on modality count. To address challenge (2), we design two learnable plug-in banks, Modality-Level Modulation bank (MLMB) and Modality-Level Prototype (MLPB) bank, to capture both modality-invariant and modality-specific knowledge. These banks are updated using our proposed Modality Prototype Contrastive Learning (MPCL). Additionally, we design Modality Adaptive Weighting (MAW) to dynamically adjust learning weights for each modality, ensuring balanced MM learning as different modalities learn at different rates. Finally, to address challenge (3), we introduce a Dual Consistency (DC) strategy that enforces consistency at both the image and feature levels without relying on generative methods. We evaluate our method on a 2-to-4 modality segmentation task using three open-source datasets, and extensive experiments show that our method outperforms state-of-the-art approaches. The source code is publicly available at https://github.com/CYYukio/Double-Knowledge-Banking.

基于多模态(MM)半监督学习(SSL)的医学图像分割最近因其利用MM数据和减少对标记图像依赖的能力而受到越来越多的关注。然而,目前的方法面临着一些挑战:(1)多网络分配阻碍了两种以上模式场景的可扩展性。(2)仅关注模态不变表示而忽略模态特定特征,导致MM学习不完整。(3)使用生成方法利用未标记的数据对于SSL来说可能是不可靠的。为了解决这些问题,我们提出了一种新的统一的MM-SSLmedical图像分割框架。为了解决挑战(1),我们提出了一种基于标准U-Net架构的模态一体化分割网络,该网络接受来自所有模态的数据,消除了对模态计数的限制。为了解决挑战(2),我们设计了两个可学习的插件库,即模态级调制库(MLMB)和模态级原型库(MLPB),以捕获模态不变和模态特定的知识。这些库使用我们提出的模态原型对比学习(MPCL)进行更新。此外,我们设计了模态自适应加权(MAW)来动态调整每个模态的学习权重,以确保在不同模态以不同速率学习时均衡的MM学习。最后,为了解决挑战(3),我们引入了双重一致性(DC)策略,该策略在图像和特征级别上强制一致性,而不依赖于生成方法。我们使用三个开源数据集在2到4模态分割任务上评估了我们的方法,大量的实验表明我们的方法优于最先进的方法。源代码可在https://github.com/CYYukio/Double-Knowledge-Banking上公开获得。
{"title":"Double Banking on Knowledge: A Unified All-in-One Framework for Unpaired Multi-Modality Semi-supervised Medical Image Segmentation.","authors":"Yingyu Chen, Ziyuan Yang, Zhongzhou Zhang, Ming Yan, Hui Yu, Yan Liu, Yi Zhang","doi":"10.1109/TBME.2026.3656540","DOIUrl":"https://doi.org/10.1109/TBME.2026.3656540","url":null,"abstract":"<p><p>Multi-modality (MM) semi-supervised learning (SSL) based medical image segmentation has recently gained increasing attention for its ability to utilize MM data and reduce reliance on labeled images. However, current methods face several challenges: (1) Multiple networks assignment hinder scalability to scenarios with more than two modalities. (2) Focusing solely on modality-invariant representation while neglecting modality-specific features, leads to incomplete MM learning. (3) Leveraging unlabeled data with generative methods can be unreliable for SSL. To address these problems, we propose a novel unified all-in-one framework for MM-SSLmedical image segmentation. To address challenge (1), we propose a modality all-in-one segmentation network based on standard U-Net architecture that accepts data from all modalities, removing the limitation on modality count. To address challenge (2), we design two learnable plug-in banks, Modality-Level Modulation bank (MLMB) and Modality-Level Prototype (MLPB) bank, to capture both modality-invariant and modality-specific knowledge. These banks are updated using our proposed Modality Prototype Contrastive Learning (MPCL). Additionally, we design Modality Adaptive Weighting (MAW) to dynamically adjust learning weights for each modality, ensuring balanced MM learning as different modalities learn at different rates. Finally, to address challenge (3), we introduce a Dual Consistency (DC) strategy that enforces consistency at both the image and feature levels without relying on generative methods. We evaluate our method on a 2-to-4 modality segmentation task using three open-source datasets, and extensive experiments show that our method outperforms state-of-the-art approaches. The source code is publicly available at https://github.com/CYYukio/Double-Knowledge-Banking.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146018457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MVICAD2: Multi-View Independent Component Analysis With Delays and Dilations MVICAD2:具有延迟和扩张的多视图独立分量分析
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-21 DOI: 10.1109/TBME.2025.3596500
Ambroise Heurtebise;Omar Chehab;Pierre Ablin;Alexandre Gramfort
Machine learning techniques in multi-view settings face significant challenges, particularly when integrating heterogeneous data, aligning feature spaces, and managing view-specific biases. These issues are prominent in neuroscience, where data from multiple subjects exposed to the same stimuli are analyzed to uncover brain activity dynamics. In magnetoencephalography (MEG), where signals are captured at the scalp level, estimating the brain's underlying sources is crucial, especially in group studies where sources are assumed to be similar for all subjects. Common methods, such as Multi-View Independent Component Analysis (MVICA), assume identical sources across subjects, but this assumption is often too restrictive due to individual variability and age-related changes. Multi-View Independent Component Analysis with Delays (MVICAD) addresses this by allowing sources to differ up to a temporal delay. However, temporal dilation effects, particularly in auditory stimuli, are common in brain dynamics, making the estimation of time delays alone insufficient. To address this, we propose Multi-View Independent Component Analysis with Delays and Dilations (MVICAD$^{2}$), which allows sources to differ across subjects in both temporal delays and dilations. We present a model with identifiable sources, derive an approximation of its likelihood in closed form, and use regularization and optimization techniques to enhance performance. Through simulations, we demonstrate that MVICAD$^{2}$ outperforms existing multi-view ICA methods. We further validate its effectiveness using the Cam-CAN dataset, and showing how delays and dilations are related to aging.
多视图设置中的机器学习技术面临着重大挑战,特别是在集成异构数据、对齐特征空间和管理视图特定偏差时。这些问题在神经科学中很突出,在神经科学中,对暴露于相同刺激的多个受试者的数据进行分析以揭示大脑活动动态。在脑磁图(MEG)中,信号是在头皮层面捕获的,估计大脑的潜在来源是至关重要的,特别是在假定所有受试者的来源相似的群体研究中。常见的方法,如多视图独立成分分析(MVICA),假设受试者之间的来源相同,但由于个体差异和年龄相关的变化,这种假设通常过于严格。具有延迟的多视图独立分量分析(MVICAD)通过允许源的差异达到时间延迟来解决这个问题。然而,时间扩张效应,特别是听觉刺激,在大脑动力学中很常见,使得单独估计时间延迟是不够的。为了解决这个问题,我们提出了具有延迟和扩张的多视图独立分量分析(MVICAD$^{2}$),它允许不同受试者的源在时间延迟和扩张上有所不同。我们提出了一个具有可识别源的模型,以封闭形式导出其可能性的近似值,并使用正则化和优化技术来提高性能。通过仿真,我们证明了MVICAD$^{2}$优于现有的多视图ICA方法。我们使用Cam-CAN数据集进一步验证了其有效性,并展示了延迟和膨胀如何与老化相关。
{"title":"MVICAD2: Multi-View Independent Component Analysis With Delays and Dilations","authors":"Ambroise Heurtebise;Omar Chehab;Pierre Ablin;Alexandre Gramfort","doi":"10.1109/TBME.2025.3596500","DOIUrl":"https://doi.org/10.1109/TBME.2025.3596500","url":null,"abstract":"Machine learning techniques in multi-view settings face significant challenges, particularly when integrating heterogeneous data, aligning feature spaces, and managing view-specific biases. These issues are prominent in neuroscience, where data from multiple subjects exposed to the same stimuli are analyzed to uncover brain activity dynamics. In magnetoencephalography (MEG), where signals are captured at the scalp level, estimating the brain's underlying sources is crucial, especially in group studies where sources are assumed to be similar for all subjects. Common methods, such as Multi-View Independent Component Analysis (MVICA), assume identical sources across subjects, but this assumption is often too restrictive due to individual variability and age-related changes. Multi-View Independent Component Analysis with Delays (MVICAD) addresses this by allowing sources to differ up to a temporal delay. However, temporal dilation effects, particularly in auditory stimuli, are common in brain dynamics, making the estimation of time delays alone insufficient. To address this, we propose Multi-View Independent Component Analysis with Delays and Dilations (MVICAD<inline-formula><tex-math>$^{2}$</tex-math></inline-formula>), which allows sources to differ across subjects in both temporal delays and dilations. We present a model with identifiable sources, derive an approximation of its likelihood in closed form, and use regularization and optimization techniques to enhance performance. Through simulations, we demonstrate that MVICAD<inline-formula><tex-math>$^{2}$</tex-math></inline-formula> outperforms existing multi-view ICA methods. We further validate its effectiveness using the Cam-CAN dataset, and showing how delays and dilations are related to aging.","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"73 2","pages":"945-952"},"PeriodicalIF":4.5,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146006905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalised Label-Free Artefact Cleaning for Real-Time Medical Pulsatile Time Series. 实时医疗脉冲时间序列的广义无标签伪影清洗。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-21 DOI: 10.1109/TBME.2026.3655095
Xuhang Chen, Ihsane Olakorede, Stefan Yu Bogli, Wenhao Xu, Erta Beqiri, Xuemeng Li, Chenyu Tang, Zeyu Gao, Shuo Gao, Ari Ercole, Peter Smielewski

Artefacts compromise clinical decision-making in the use of medical time series. Pulsatile waveforms offer opportunities for accurate artefact detection, yet most approaches rely on supervised manners and overlook patient-level distribution shifts. To address these issues, we introduce GenClean, a generalised label-free framework for real-time artefact cleaning, implemented within the ICM+ clinical research monitoring software. Leveraging an in-house dataset of 180,000 ten-second arterial blood pressure (ABP) samples for training, we first investigate patient-level generalisation, demonstrating robust performance under both intra- and inter-patient distribution shifts. As an initial exploration beyond the development cohort, we further validate its effectiveness for ABP through site-level generalisation on the MIMIC-III database. We also provided an extension of our method to photoplethysmography (PPG), highlighting its potential applicability to diverse medical pulsatile signals. The real-time integration and these generalisation studies collectively demonstrate the practical utility of our framework in continuous physiological monitoring and represent a promising step towards improving the reliability of high-resolution medical time series analysis.

在使用医疗时间序列时,人工制品会影响临床决策。脉动波形为精确的伪影检测提供了机会,但大多数方法依赖于监督方式,忽略了患者水平的分布变化。为了解决这些问题,我们引入了GenClean,这是一种通用的无标签框架,用于实时伪迹清洗,在ICM+临床研究监测软件中实施。利用180,000个10秒动脉血压(ABP)样本的内部数据集进行训练,我们首先调查了患者水平的泛化,证明了在患者内部和患者间分布变化下的稳健表现。作为开发队列之外的初步探索,我们通过在MIMIC-III数据库上的站点级推广进一步验证了其对ABP的有效性。我们还提供了我们的方法扩展到光体积脉搏图(PPG),强调其潜在的适用性,以不同的医疗脉搏信号。实时集成和这些推广研究共同证明了我们的框架在连续生理监测中的实际效用,并代表了提高高分辨率医学时间序列分析可靠性的有希望的一步。
{"title":"Generalised Label-Free Artefact Cleaning for Real-Time Medical Pulsatile Time Series.","authors":"Xuhang Chen, Ihsane Olakorede, Stefan Yu Bogli, Wenhao Xu, Erta Beqiri, Xuemeng Li, Chenyu Tang, Zeyu Gao, Shuo Gao, Ari Ercole, Peter Smielewski","doi":"10.1109/TBME.2026.3655095","DOIUrl":"https://doi.org/10.1109/TBME.2026.3655095","url":null,"abstract":"<p><p>Artefacts compromise clinical decision-making in the use of medical time series. Pulsatile waveforms offer opportunities for accurate artefact detection, yet most approaches rely on supervised manners and overlook patient-level distribution shifts. To address these issues, we introduce GenClean, a generalised label-free framework for real-time artefact cleaning, implemented within the ICM+ clinical research monitoring software. Leveraging an in-house dataset of 180,000 ten-second arterial blood pressure (ABP) samples for training, we first investigate patient-level generalisation, demonstrating robust performance under both intra- and inter-patient distribution shifts. As an initial exploration beyond the development cohort, we further validate its effectiveness for ABP through site-level generalisation on the MIMIC-III database. We also provided an extension of our method to photoplethysmography (PPG), highlighting its potential applicability to diverse medical pulsatile signals. The real-time integration and these generalisation studies collectively demonstrate the practical utility of our framework in continuous physiological monitoring and represent a promising step towards improving the reliability of high-resolution medical time series analysis.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146018460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
STF: A Unified Framework for Joint Pixel-Level Segmentation and Tracking of Tissues in Endoscopic Surgery. 内镜手术中关节像素级分割和组织跟踪的统一框架。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-21 DOI: 10.1109/TBME.2026.3656751
Yifan Li, Laura Cruciani, Francesco Alessandro Mistretta, Stefano Luzzago, Giancarlo Ferrigno, Gennaro Musi, Elena De Momi

Endoscopic minimally invasive surgery relies on precise tissue video segmentation to avoid complications such as vascular bleeding or nerve injury. However, existing video segmentation methods often fail to maintain long-term robustness due to target loss and challenging conditions (e.g., occlusion, motion blur), limiting their applicability in prolonged surgical procedures. To address these limitations, we proposed the Unified Framework for Joint Pixel-Level Segmentation and Tracking (STF), it integrates a synergistic segmentation-guided tracking pipeline with an adaptive re-detection mechanism. First, a deep learning-based segmentation network precisely localizes the target tissue. A cost-efficient Hough Voting Network then tracks the segmented region, while a Bayesian refinement module improves compatibility between segmentation and tracking. If tracking reliability drops, an evaluation module triggers re-segmentation, ensuring continuous and stable long-term performance. Extensive experiments confirm that STF achieves superior accuracy and temporal consistency over segmentation networks in long-term surgical video segmentation, particularly under extreme conditions. This automated methodology significantly improves the robustness and re-detection capability for sustained tissue analysis, markedly reducing the dependency on manual intervention prevalent in many model-based tracking solutions.

内镜微创手术依靠精确的组织视频分割来避免血管出血或神经损伤等并发症。然而,由于目标丢失和具有挑战性的条件(例如遮挡、运动模糊),现有的视频分割方法往往不能保持长期的鲁棒性,限制了它们在长时间外科手术中的适用性。为了解决这些限制,我们提出了联合像素级分割和跟踪的统一框架(STF),它集成了一个协同分割引导的跟踪管道和自适应重新检测机制。首先,基于深度学习的分割网络精确定位目标组织。然后,一个经济高效的霍夫投票网络跟踪分割区域,而贝叶斯优化模块提高分割和跟踪之间的兼容性。如果跟踪可靠性下降,评估模块触发重新分割,确保持续稳定的长期性能。大量实验证实,在长期外科手术视频分割中,特别是在极端条件下,STF在分割网络上具有优越的准确性和时间一致性。这种自动化方法显著提高了持续组织分析的鲁棒性和重新检测能力,显著减少了许多基于模型的跟踪解决方案对人工干预的依赖。
{"title":"STF: A Unified Framework for Joint Pixel-Level Segmentation and Tracking of Tissues in Endoscopic Surgery.","authors":"Yifan Li, Laura Cruciani, Francesco Alessandro Mistretta, Stefano Luzzago, Giancarlo Ferrigno, Gennaro Musi, Elena De Momi","doi":"10.1109/TBME.2026.3656751","DOIUrl":"https://doi.org/10.1109/TBME.2026.3656751","url":null,"abstract":"<p><p>Endoscopic minimally invasive surgery relies on precise tissue video segmentation to avoid complications such as vascular bleeding or nerve injury. However, existing video segmentation methods often fail to maintain long-term robustness due to target loss and challenging conditions (e.g., occlusion, motion blur), limiting their applicability in prolonged surgical procedures. To address these limitations, we proposed the Unified Framework for Joint Pixel-Level Segmentation and Tracking (STF), it integrates a synergistic segmentation-guided tracking pipeline with an adaptive re-detection mechanism. First, a deep learning-based segmentation network precisely localizes the target tissue. A cost-efficient Hough Voting Network then tracks the segmented region, while a Bayesian refinement module improves compatibility between segmentation and tracking. If tracking reliability drops, an evaluation module triggers re-segmentation, ensuring continuous and stable long-term performance. Extensive experiments confirm that STF achieves superior accuracy and temporal consistency over segmentation networks in long-term surgical video segmentation, particularly under extreme conditions. This automated methodology significantly improves the robustness and re-detection capability for sustained tissue analysis, markedly reducing the dependency on manual intervention prevalent in many model-based tracking solutions.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146018445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
uPVC-Net: A Universal Premature Ventricular Contraction Detection Deep Learning Algorithm. uPVC-Net:一种通用的室性早缩检测深度学习算法。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-20 DOI: 10.1109/TBME.2026.3655531
Hagai Hamami, Yosef Solewicz, Daniel Zur, Yonatan Kleerekoper, Joachim A Behar

Introduction: Premature Ventricular Contractions (PVCs) are common cardiac arrhythmias originating from the ventricles. Accurate detection remains challenging due to variability in electrocardiogram (ECG) waveforms caused by differences in lead placement, recording conditions, and population demographics.

Methods: We developed uPVC-Net, a universal deep learning model to detect PVCs from any single-lead ECG recordings. The model is developed on four independent ECG datasets comprising a total of 8.3 million beats collected from Holter monitors and a modern wearable ECG patch. uPVC-Net employs a custom architecture and a multi-source, multi-lead training strategy. For each experiment, one dataset is held out to evaluate out-of-distribution (OOD) generalization.

Results: uPVC-Net achieved an AUC between 97.8% and 99.1% on the held-out datasets. Notably, performance on wearable single-lead ECG data reached an AUC of 99.1%.

Conclusion: uPVC-Net exhibits strong generalization across diverse lead configurations and populations, highlighting its potential for robust, real-world clinical deployment.

室性早搏是一种常见的心律失常,起源于心室。由于导联位置、记录条件和人口统计数据的差异导致心电图(ECG)波形的变化,准确检测仍然具有挑战性。方法:我们开发了uPVC-Net,这是一种通用的深度学习模型,可以从任何单导联心电图记录中检测室性早搏。该模型是在四个独立的心电图数据集上开发的,这些数据集包括从霍尔特监视器和现代可穿戴心电图贴片收集的总共830万次心跳。uPVC-Net采用自定义架构和多源、多先导的培训策略。对于每个实验,都保留一个数据集来评估分布外(OOD)泛化。结果:uPVC-Net在hold -out数据集上的AUC在97.8%到99.1%之间。值得注意的是,可穿戴单导联心电图数据的AUC达到99.1%。结论:uPVC-Net在不同的引线配置和人群中表现出很强的通用性,突出了其在现实世界临床应用的潜力。
{"title":"uPVC-Net: A Universal Premature Ventricular Contraction Detection Deep Learning Algorithm.","authors":"Hagai Hamami, Yosef Solewicz, Daniel Zur, Yonatan Kleerekoper, Joachim A Behar","doi":"10.1109/TBME.2026.3655531","DOIUrl":"https://doi.org/10.1109/TBME.2026.3655531","url":null,"abstract":"<p><strong>Introduction: </strong>Premature Ventricular Contractions (PVCs) are common cardiac arrhythmias originating from the ventricles. Accurate detection remains challenging due to variability in electrocardiogram (ECG) waveforms caused by differences in lead placement, recording conditions, and population demographics.</p><p><strong>Methods: </strong>We developed uPVC-Net, a universal deep learning model to detect PVCs from any single-lead ECG recordings. The model is developed on four independent ECG datasets comprising a total of 8.3 million beats collected from Holter monitors and a modern wearable ECG patch. uPVC-Net employs a custom architecture and a multi-source, multi-lead training strategy. For each experiment, one dataset is held out to evaluate out-of-distribution (OOD) generalization.</p><p><strong>Results: </strong>uPVC-Net achieved an AUC between 97.8% and 99.1% on the held-out datasets. Notably, performance on wearable single-lead ECG data reached an AUC of 99.1%.</p><p><strong>Conclusion: </strong>uPVC-Net exhibits strong generalization across diverse lead configurations and populations, highlighting its potential for robust, real-world clinical deployment.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146010260","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rectification of Cornea Induced Distortions in Microscopic Images for Assisted Ophthalmic Surgery. 辅助眼科手术中角膜显微图像畸变的矫正。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-20 DOI: 10.1109/TBME.2026.3656209
Rebekka Peter, Erik Oberschulte, Atharva Vaidya, Thomas Lindemeier, Franziska Mathis-Ullrich, Eleonora Tagliabue

Objective: Image distortions induced by the high refractive power of the eye's optical components challenge the accuracy of geometric information derived from intra-operative sensor data in ophthalmic surgery. Correcting these distortions is vital for advancing surgical assistance systems that rely on geometric scene comprehension. In this work, we focus on cornea induced distortions (CIDs) in surgical microscope images of the anterior eye.

Methods: We employ a convolutional neural network (CNN) with stereo fusion layers to predict distortion distribution maps (DDMs) to correct CIDs in stereo images. To enable supervised learning, we introduce CIDCAT, a synthetic surgical microscope dataset generated through a rendering pipeline using a digital eye model. We address the domain gap between the synthetic training data and the unlabeled target domain of real surgical images by employing an auxiliary task of semantic segmentation to regularizes the feature encoder.

Results: Our rectification model reduces the cornea induced pupil radius error from 8.56% to 0.72% and improves the structural similarity by over 9% for synthetic CIDCAT images. Our semantic segmentation driven domain regularization technique enables the translation to real surgical images.

Conclusion: The CIDCAT dataset enables the investigation of CIDs and the implementation of a CID rectification model. Our proposed CID rectification model demonstrate successful minimization of CIDs while preserving image integrity.

Significance: The work represents a significant advancement in research on computer-assistance and robotic solutions for ophthalmic surgery that rely on a distortion-free, three-dimensional understanding of the patient's eye.

目的:眼光学元件的高屈光度引起的图像畸变对眼外科术中传感器数据获取几何信息的准确性提出了挑战。纠正这些扭曲对于推进依赖几何场景理解的手术辅助系统至关重要。在这项工作中,我们的重点是角膜诱导畸变(cid)在手术显微镜图像的前眼。方法:采用具有立体融合层的卷积神经网络(CNN)来预测立体图像的畸变分布图(DDMs),以校正立体图像中的cid。为了实现监督学习,我们引入了CIDCAT,这是一个通过使用数字眼模型的渲染管道生成的合成手术显微镜数据集。我们通过使用语义分割的辅助任务对特征编码器进行正则化,解决了合成训练数据与真实手术图像的未标记目标域之间的域差距。结果:校正模型将合成CIDCAT图像的角膜瞳孔半径误差从8.56%降低到0.72%,结构相似性提高9%以上。我们的语义分割驱动域正则化技术使翻译到真实的手术图像。结论:CIDCAT数据集能够对CID进行调查并实现CID整改模型。我们提出的CID校正模型成功地最小化了CID,同时保持了图像的完整性。意义:这项工作代表了眼科手术计算机辅助和机器人解决方案研究的重大进展,这些解决方案依赖于对患者眼睛的无扭曲、三维理解。
{"title":"Rectification of Cornea Induced Distortions in Microscopic Images for Assisted Ophthalmic Surgery.","authors":"Rebekka Peter, Erik Oberschulte, Atharva Vaidya, Thomas Lindemeier, Franziska Mathis-Ullrich, Eleonora Tagliabue","doi":"10.1109/TBME.2026.3656209","DOIUrl":"https://doi.org/10.1109/TBME.2026.3656209","url":null,"abstract":"<p><strong>Objective: </strong>Image distortions induced by the high refractive power of the eye's optical components challenge the accuracy of geometric information derived from intra-operative sensor data in ophthalmic surgery. Correcting these distortions is vital for advancing surgical assistance systems that rely on geometric scene comprehension. In this work, we focus on cornea induced distortions (CIDs) in surgical microscope images of the anterior eye.</p><p><strong>Methods: </strong>We employ a convolutional neural network (CNN) with stereo fusion layers to predict distortion distribution maps (DDMs) to correct CIDs in stereo images. To enable supervised learning, we introduce CIDCAT, a synthetic surgical microscope dataset generated through a rendering pipeline using a digital eye model. We address the domain gap between the synthetic training data and the unlabeled target domain of real surgical images by employing an auxiliary task of semantic segmentation to regularizes the feature encoder.</p><p><strong>Results: </strong>Our rectification model reduces the cornea induced pupil radius error from 8.56% to 0.72% and improves the structural similarity by over 9% for synthetic CIDCAT images. Our semantic segmentation driven domain regularization technique enables the translation to real surgical images.</p><p><strong>Conclusion: </strong>The CIDCAT dataset enables the investigation of CIDs and the implementation of a CID rectification model. Our proposed CID rectification model demonstrate successful minimization of CIDs while preserving image integrity.</p><p><strong>Significance: </strong>The work represents a significant advancement in research on computer-assistance and robotic solutions for ophthalmic surgery that rely on a distortion-free, three-dimensional understanding of the patient's eye.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146010292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised Speech Representations for Sleep Apnea Severity Prediction. 睡眠呼吸暂停严重程度预测的自监督语音表征。
IF 4.5 2区 医学 Q2 ENGINEERING, BIOMEDICAL Pub Date : 2026-01-20 DOI: 10.1109/TBME.2026.3656326
Behrad TaghiBeyglou, Jiahao Geng, Dominick McDaulid, Papina Gnaneswaran, Oviga Yasokaran, Alexander Chow, Raymond Ng, Ronald D Chervin, Devin L Brown, Azadeh Yadollahi

Objective: Obstructive sleep apnea (OSA) is a prevalent yet underdiagnosed condition characterized by repetitive upper airway obstruction during sleep. Current gold standard diagnostic standards rely on polysomnography (PSG), which is resource-intensive. Since upper airway characteristics impact both OSA and speech production, speech processing has emerged as a promising alternative for OSA screening. However, prior work has focused primarily on acoustic features. This study aims to develop a speech-based screening and severity estimation pipeline for OSA using self-supervised learning (SSL) and multimodal acoustic features.

Methods: We proposed a novel fusion framework combining SSL-derived speech representations from pre-trained neural networks with traditional acoustic features and time-frequency representations of speech phase and magnitude. Elongated vowels recorded during wakefulness were used to screen for OSA at two apnea-hypopnea index (AHI) thresholds (10 and 30 events/hour) and to estimate AHI. Data were collected across three research sites, comprising participants of varied sex, race, and OSA severity.

Results: For OSA screening, the models achieved balanced accuracies of 0.79 (AHI $geq$10) and 0.74 (AHI $geq$30) in females, and 0.80 and 0.78 in males, respectively. AHI estimation yielded mean absolute errors of 12.0 events/hour (r = 0.63) in females and 14.7 events/hour (r = 0.52) in males.

Conclusion: Our results demonstrate the feasibility of using speech, especially vowel phonation during wakefulness, as a biomarker for OSA risk and severity estimation. The approach generalizes well across diverse demographic groups.

Significance: This study presents a significant step toward accessible, low-burden, and cost-effective OSA screening, with broad implications for scalable sleep health assessments.

目的:阻塞性睡眠呼吸暂停(OSA)是一种常见但未被诊断的疾病,其特征是睡眠期间反复出现上呼吸道阻塞。目前的金标准诊断标准依赖于多导睡眠图(PSG),这是资源密集型的。由于上呼吸道特征影响OSA和语音产生,语音处理已成为OSA筛查的一种有希望的替代方法。然而,先前的工作主要集中在声学特征上。本研究旨在利用自监督学习(SSL)和多模态声学特征开发基于语音的OSA筛查和严重程度估计管道。方法:我们提出了一种新的融合框架,将来自预训练神经网络的ssl衍生语音表示与传统声学特征以及语音相位和幅度的时频表示相结合。在清醒时记录的拉长元音用于在两个呼吸暂停低通气指数(AHI)阈值(10和30事件/小时)下筛查OSA并估计AHI。数据从三个研究地点收集,包括不同性别、种族和OSA严重程度的参与者。结果:对于OSA筛查,模型在女性中分别达到了0.79 (AHI $geq$ 10)和0.74 (AHI $geq$ 30)的平衡精度,在男性中分别达到了0.80和0.78。在AHI估计中,女性的平均绝对误差为12.0事件/小时(r = 0.63),男性为14.7事件/小时(r = 0.52)。结论:我们的研究结果证明了使用语音,特别是清醒时的元音发音作为OSA风险和严重程度评估的生物标志物的可行性。这种方法适用于不同的人口群体。意义:本研究向可获得、低负担、低成本的OSA筛查迈出了重要一步,对可扩展的睡眠健康评估具有广泛意义。
{"title":"Self-Supervised Speech Representations for Sleep Apnea Severity Prediction.","authors":"Behrad TaghiBeyglou, Jiahao Geng, Dominick McDaulid, Papina Gnaneswaran, Oviga Yasokaran, Alexander Chow, Raymond Ng, Ronald D Chervin, Devin L Brown, Azadeh Yadollahi","doi":"10.1109/TBME.2026.3656326","DOIUrl":"https://doi.org/10.1109/TBME.2026.3656326","url":null,"abstract":"<p><strong>Objective: </strong>Obstructive sleep apnea (OSA) is a prevalent yet underdiagnosed condition characterized by repetitive upper airway obstruction during sleep. Current gold standard diagnostic standards rely on polysomnography (PSG), which is resource-intensive. Since upper airway characteristics impact both OSA and speech production, speech processing has emerged as a promising alternative for OSA screening. However, prior work has focused primarily on acoustic features. This study aims to develop a speech-based screening and severity estimation pipeline for OSA using self-supervised learning (SSL) and multimodal acoustic features.</p><p><strong>Methods: </strong>We proposed a novel fusion framework combining SSL-derived speech representations from pre-trained neural networks with traditional acoustic features and time-frequency representations of speech phase and magnitude. Elongated vowels recorded during wakefulness were used to screen for OSA at two apnea-hypopnea index (AHI) thresholds (10 and 30 events/hour) and to estimate AHI. Data were collected across three research sites, comprising participants of varied sex, race, and OSA severity.</p><p><strong>Results: </strong>For OSA screening, the models achieved balanced accuracies of 0.79 (AHI $geq$10) and 0.74 (AHI $geq$30) in females, and 0.80 and 0.78 in males, respectively. AHI estimation yielded mean absolute errors of 12.0 events/hour (r = 0.63) in females and 14.7 events/hour (r = 0.52) in males.</p><p><strong>Conclusion: </strong>Our results demonstrate the feasibility of using speech, especially vowel phonation during wakefulness, as a biomarker for OSA risk and severity estimation. The approach generalizes well across diverse demographic groups.</p><p><strong>Significance: </strong>This study presents a significant step toward accessible, low-burden, and cost-effective OSA screening, with broad implications for scalable sleep health assessments.</p>","PeriodicalId":13245,"journal":{"name":"IEEE Transactions on Biomedical Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.5,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146010280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Biomedical Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1