首页 > 最新文献

IEEE Journal of Biomedical and Health Informatics最新文献

英文 中文
Continuous Cuffless Blood Pressure Estimation via Effective and Efficient Broad Learning Model. 基于有效和高效的广义学习模型的连续无袖带血压测量。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3604464
Chunlin Zhang, Pingyu Hu, Zhan Shen, Xiaorong Ding

Hypertension is a critical cardiovascular risk factor, underscoring the necessity of accessible blood pressure (BP) monitoring for its prevention, detection, and management. While cuffless BP estimation using wearable cardiovascular signals via deep learning models (DLMs) offers a promising solution, their implementation often entails high computational costs. This study addresses these challenges by proposing an end-to-end broad learning model (BLM) for efficient cuffless BP estimation. Unlike DLMs that prioritize network depth, the BLM increases network width, thereby reducing computational complexity and enhancing training efficiency for continuous BP estimation. An incremental learning mode is also explored to provide high memory efficiency and flexibility. Validation on the University of California Irvine (UCI) database (403.67 hours) demonstrated that the standard BLM (SBLM) achieved a mean absolute error (MAE) of 11.72 mmHg for arterial BP (ABP) waveform estimation, performance comparable to DLMs such as long short-term memory (LSTM) and the one-dimensional convolutional neural network (1D-CNN), while improving training efficiency by 25.20 times. The incremental BLM (IBLM) offered horizontal scalability by expanding through node addition in a single layer, maintaining predictive performance while reducing storage demands through support for incremental learning with streaming or partial datasets. For systolic and diastolic BP prediction, the SBLM achieved MAEs (mean error $pm$ standard deviation) of 3.04 mmHg (2.85 $pm$ 4.15 mmHg) and 2.57 mmHg (-2.47 $pm$ 3.03 mmHg), respectively. This study highlights the potential of BLM for personalized, real-time, continuous cuffless BP monitoring, presenting a practical solution for healthcare applications.

高血压是心血管疾病和全因死亡的一个主要危险因素,因此方便易行的血压(BP)测量,如无袖带方法,对其预防、检测和管理至关重要。通过深度学习模型(DLMs)使用可穿戴心血管信号进行无袖扣BP估计提供了一个很有前途的解决方案。然而,dlm的实现通常需要很高的计算成本和时间。本研究通过提供端到端广泛学习模型(BLM)来解决这些挑战,以实现有效和高效的无套BP估计。与dlm相比,BLM增加了网络宽度而不是深度,降低了计算复杂度,提高了连续BP估计的训练效率。我们还探索了一种提供高记忆效率和灵活性的增量学习模式。在加州大学欧文分校(UCI)数据库上进行的长达403.67小时的验证表明,标准BLM (SBLM)估计动脉血压(ABP)波形的平均绝对误差(MAE)为11.72 mmHg,与长短期记忆(LSTM)和一维卷积神经网络(1D-CNN)等dlm的性能相当,同时显著提高了25.20倍的训练效率。此外,增量BLM (IBLM)提供了一种水平可伸缩性方法,它涉及通过在单个层中添加节点而不是增加层数来扩展模型,用于增量学习,有效地更新模型,同时保持可比较的预测性能。这种方法通过支持流式或部分数据集的增量学习来减少存储需求。此外,SBLM预测收缩压(SBP)和舒张压(DBP)的平均绝对误差(MAE)(平均误差(ME)±标准差(SD))值分别为3.04 mmHg(2.85±4.15 mmHg)和2.57 mmHg(-2.47±3.03 mmHg)。这项研究强调了BLM在个性化、实时和连续无袖带血压监测方面的潜力,为医疗保健应用提供了一个实用的解决方案。
{"title":"Continuous Cuffless Blood Pressure Estimation via Effective and Efficient Broad Learning Model.","authors":"Chunlin Zhang, Pingyu Hu, Zhan Shen, Xiaorong Ding","doi":"10.1109/JBHI.2025.3604464","DOIUrl":"10.1109/JBHI.2025.3604464","url":null,"abstract":"<p><p>Hypertension is a critical cardiovascular risk factor, underscoring the necessity of accessible blood pressure (BP) monitoring for its prevention, detection, and management. While cuffless BP estimation using wearable cardiovascular signals via deep learning models (DLMs) offers a promising solution, their implementation often entails high computational costs. This study addresses these challenges by proposing an end-to-end broad learning model (BLM) for efficient cuffless BP estimation. Unlike DLMs that prioritize network depth, the BLM increases network width, thereby reducing computational complexity and enhancing training efficiency for continuous BP estimation. An incremental learning mode is also explored to provide high memory efficiency and flexibility. Validation on the University of California Irvine (UCI) database (403.67 hours) demonstrated that the standard BLM (SBLM) achieved a mean absolute error (MAE) of 11.72 mmHg for arterial BP (ABP) waveform estimation, performance comparable to DLMs such as long short-term memory (LSTM) and the one-dimensional convolutional neural network (1D-CNN), while improving training efficiency by 25.20 times. The incremental BLM (IBLM) offered horizontal scalability by expanding through node addition in a single layer, maintaining predictive performance while reducing storage demands through support for incremental learning with streaming or partial datasets. For systolic and diastolic BP prediction, the SBLM achieved MAEs (mean error $pm$ standard deviation) of 3.04 mmHg (2.85 $pm$ 4.15 mmHg) and 2.57 mmHg (-2.47 $pm$ 3.03 mmHg), respectively. This study highlights the potential of BLM for personalized, real-time, continuous cuffless BP monitoring, presenting a practical solution for healthcare applications.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1101-1114"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144952165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Deep Learning Approach for Epileptic Seizure Detection in EEG signals. 用于脑电图信号中癫痫发作检测的混合深度学习方法。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2023.3265983
Ijaz Ahmad, Xin Wang, Danish Javeed, Prabhat Kumar, Oluwarotimi Williams Samuel, Shixiong Chen

Early detection and proper treatment of epilepsy is essential and meaningful to those who suffer from this disease. The adoption of deep learning (DL) techniques for automated epileptic seizure detection using electroencephalography (EEG) signals has shown great potential in making the most appropriate and fast medical decisions. However, DL algorithms have high computational complexity and suffer low accuracy with imbalanced medical data in multi seizure-classification task. Motivated from the aforementioned challenges, we present a simple and effective hybrid DL approach for epileptic seizure detection in EEG signals. Specifically, first we use a K-means Synthetic minority oversampling technique (SMOTE) to balance the sampling data. Second, we integrate a 1D convolutional neural network (CNN) with a Bidirectional Long Short-Term Memory (BiLSTM) network based on Truncated Backpropagation Through Time (TBPTT) to efficiently extract spatial and temporal sequence information while reducing computational complexity. Finally, the proposed DL architecture uses softmax and sigmoid classifiers at the classification layer to perform multi and binary seizure-classification tasks. In addition, the 10-fold cross-validation technique is performed to show the significance of the proposed DL approach. Experimental results using the publicly available UCI epileptic seizure recognition data set shows better performance in terms of precision, sensitivity, specificity, and F1-score over some baseline DL algorithms and recent state-of-the-art techniques.

对于癫痫患者来说,及早发现和正确治疗癫痫是至关重要的,也是非常有意义的。采用深度学习(DL)技术利用脑电图(EEG)信号进行癫痫发作自动检测,在做出最合适、最快速的医疗决策方面显示出巨大的潜力。然而,深度学习算法具有较高的计算复杂性,而且在多癫痫发作分类任务中,不平衡医疗数据的准确性较低。鉴于上述挑战,我们提出了一种简单有效的混合 DL 方法,用于检测脑电信号中的癫痫发作。具体来说,首先,我们使用 K-means 合成少数过采样技术(SMOTE)来平衡采样数据。其次,我们整合了一维卷积神经网络(CNN)和基于时间截断反向传播(TBPTT)的双向长短期记忆(BiLSTM)网络,以有效提取空间和时间序列信息,同时降低计算复杂度。最后,拟议的 DL 架构在分类层使用 softmax 和 sigmoid 分类器来执行多重和二元癫痫发作分类任务。此外,还采用了 10 倍交叉验证技术,以显示所提出的 DL 方法的重要性。使用公开可用的 UCI 癫痫发作识别数据集进行的实验结果表明,在精确度、灵敏度、特异性和 F1 分数方面,该方法都优于一些基准 DL 算法和最新的先进技术。
{"title":"A Hybrid Deep Learning Approach for Epileptic Seizure Detection in EEG signals.","authors":"Ijaz Ahmad, Xin Wang, Danish Javeed, Prabhat Kumar, Oluwarotimi Williams Samuel, Shixiong Chen","doi":"10.1109/JBHI.2023.3265983","DOIUrl":"10.1109/JBHI.2023.3265983","url":null,"abstract":"<p><p>Early detection and proper treatment of epilepsy is essential and meaningful to those who suffer from this disease. The adoption of deep learning (DL) techniques for automated epileptic seizure detection using electroencephalography (EEG) signals has shown great potential in making the most appropriate and fast medical decisions. However, DL algorithms have high computational complexity and suffer low accuracy with imbalanced medical data in multi seizure-classification task. Motivated from the aforementioned challenges, we present a simple and effective hybrid DL approach for epileptic seizure detection in EEG signals. Specifically, first we use a K-means Synthetic minority oversampling technique (SMOTE) to balance the sampling data. Second, we integrate a 1D convolutional neural network (CNN) with a Bidirectional Long Short-Term Memory (BiLSTM) network based on Truncated Backpropagation Through Time (TBPTT) to efficiently extract spatial and temporal sequence information while reducing computational complexity. Finally, the proposed DL architecture uses softmax and sigmoid classifiers at the classification layer to perform multi and binary seizure-classification tasks. In addition, the 10-fold cross-validation technique is performed to show the significance of the proposed DL approach. Experimental results using the publicly available UCI epileptic seizure recognition data set shows better performance in terms of precision, sensitivity, specificity, and F1-score over some baseline DL algorithms and recent state-of-the-art techniques.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1019-1029"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9274448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CNNViT-MILF-a: A Novel Architecture Leveraging the Synergy of CNN and ViT for Motor Imagery Classification. cnnviti - milf - A:一种利用CNN和ViT协同作用的运动图像分类新架构。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3587026
Zhenxi Zhao, Yingyu Cao, Hongbin Yu, Huixian Yu, Junfen Huang

Accurate motor imagery (MI) classification in EEG-based brain-computer interfaces (BCIs) is essential for applications in engineering, medicine, and artificial intelligence. Due to the limitations of single-model approaches, hybrid model architectures have emerged as a promising direction. In particular, convolutional neural networks (CNNs) and vision transformers (ViTs) demonstrate strong complementary capabilities, leading to enhanced performance. This study proposes a series of novel models, termed as CNNViT-MI, to explore the synergy of CNNs and ViTs for MI classification. Specifically, five fusion strategies were defined: parallel integration, sequential integration, hierarchical integration, early fusion, and late fusion. Based on these strategies, eight candidate models were developed. Experiments were conducted on four datasets: BCI competition IV dataset 2a, BCI competition IV dataset 2b, high gamma dataset, and a self-collected MI-GS dataset. The results demonstrate that CNNViT-MILF-a achieves the best performance among all candidates by leveraging ViT as the backbone for global feature extraction and incorporating CNN-based local representations through a late fusion strategy. Compared to the best-performing state-of-the-art (SOTA) methods, mean accuracy was improved by 2.27%, 2.31%, 0.74%, and 2.50% on the respective datasets, confirming the model's effectiveness and broad applicability, other metrics showed similar improvements. In addition, significance analysis, ablation studies, and visualization analysis were conducted, and corresponding clinical integration and rehabilitation protocols were developed to support practical use in healthcare.

在基于脑电图的脑机接口(bci)中,准确的运动图像(MI)分类对于工程、医学和人工智能的应用至关重要。由于单一模型方法的局限性,混合模型体系结构已经成为一个有前途的方向。特别是卷积神经网络(cnn)和视觉变压器(ViTs)表现出很强的互补能力,从而提高了性能。本研究提出了一系列被称为CNNViT-MI的新模型,以探索cnn和vit在MI分类中的协同作用。具体而言,定义了五种融合策略:并行融合、顺序融合、分层融合、早期融合和晚期融合。基于这些策略,开发了8个候选模型。实验在4个数据集上进行:BCI competition IV dataset 2a、BCI competition IV dataset 2b、高伽马数据集和自收集的MI-GS数据集。结果表明,cnnit - milf -a利用ViT作为全局特征提取的骨干,并通过后期融合策略融合基于cnn的局部表征,在所有候选图像中取得了最佳性能。与表现最好的最先进的(SOTA)方法相比,在各自的数据集上,平均准确率提高了2.27%,2.31%,0.74%和2.50%,证实了模型的有效性和广泛适用性,其他指标也显示出类似的改进。此外,还进行了显著性分析、消融研究和可视化分析,并制定了相应的临床整合和康复方案,以支持在医疗保健中的实际应用。
{"title":"CNNViT-MILF-a: A Novel Architecture Leveraging the Synergy of CNN and ViT for Motor Imagery Classification.","authors":"Zhenxi Zhao, Yingyu Cao, Hongbin Yu, Huixian Yu, Junfen Huang","doi":"10.1109/JBHI.2025.3587026","DOIUrl":"10.1109/JBHI.2025.3587026","url":null,"abstract":"<p><p>Accurate motor imagery (MI) classification in EEG-based brain-computer interfaces (BCIs) is essential for applications in engineering, medicine, and artificial intelligence. Due to the limitations of single-model approaches, hybrid model architectures have emerged as a promising direction. In particular, convolutional neural networks (CNNs) and vision transformers (ViTs) demonstrate strong complementary capabilities, leading to enhanced performance. This study proposes a series of novel models, termed as CNNViT-MI, to explore the synergy of CNNs and ViTs for MI classification. Specifically, five fusion strategies were defined: parallel integration, sequential integration, hierarchical integration, early fusion, and late fusion. Based on these strategies, eight candidate models were developed. Experiments were conducted on four datasets: BCI competition IV dataset 2a, BCI competition IV dataset 2b, high gamma dataset, and a self-collected MI-GS dataset. The results demonstrate that CNNViT-MILF-a achieves the best performance among all candidates by leveraging ViT as the backbone for global feature extraction and incorporating CNN-based local representations through a late fusion strategy. Compared to the best-performing state-of-the-art (SOTA) methods, mean accuracy was improved by 2.27%, 2.31%, 0.74%, and 2.50% on the respective datasets, confirming the model's effectiveness and broad applicability, other metrics showed similar improvements. In addition, significance analysis, ablation studies, and visualization analysis were conducted, and corresponding clinical integration and rehabilitation protocols were developed to support practical use in healthcare.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1153-1165"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144583776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Psychological Assessments With Open-Ended Questionnaires and Large Language Models: An ASD Case Study. 运用开放式问卷和大型语言模型加强心理评估:ASD个案研究。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3599643
Alberto Altozano, Maria Eleonora Minissi, Lucia Gomez-Zaragoza, Luna Maddalon, Mariano Alcaniz, Javier Marin-Morales

Open-ended questionnaires allow respondents to express freely, capturing richer information than close-ended formats, but they are harder to analyze. Recent natural language processing advancements enable automatic assessment of open-ended responses, yet its use in psychological classification is underexplored. This study proposes a methodology using pre-trained large language models (LLMs) for automatic classification of open-ended questionnaires, applied to autism spectrum disorder (ASD) classification via parental reports. We compare multiple training strategies using transcribed responses from 51 parents (26 with typically developing children, 25 with ASD), exploring variations in model fine-tuning, input representation, and specificity. Subject-level predictions are derived by aggregating 12 individual question responses. Our best approach achieved 84% subject-wise accuracy and 1.0 ROC-AUC using an OpenAI embedding model, per-question training, including questions in the input, and combining the predictions with a voting system. In addition, a zero-shot evaluation using GPT-4o was conducted, yielding comparable results, underscoring the potential of both compact, local models and large out-of-the-box LLMs. To enhance transparency, we explored interpretability methods. Proprietary LLMs like GPT-4o offered no direct explanation, and OpenAI embedding models showed limited interpretability. However, locally deployable LLMs provided the highest interpretability. This highlights a trade-off between proprietary models' performance and local models' explainability. Our findings validate LLMs for automatically classifying open-ended questionnaires, offering a scalable, cost-effective complement for ASD assessment. These results suggest broader applicability for psychological analysis of other conditions, advancing LLM use in mental health research.

开放式问卷允许受访者自由表达,比封闭式问卷获取更丰富的信息,但更难分析。最近自然语言处理的进步使开放式反应的自动评估成为可能,但其在心理分类中的应用尚未得到充分探索。本研究提出了一种使用预训练大语言模型(LLMs)对开放式问卷进行自动分类的方法,并将其应用于通过父母报告对自闭症谱系障碍(ASD)进行分类。我们使用51位家长(26位为正常发育儿童,25位为ASD儿童)的转录反应来比较多种训练策略,探索模型微调、输入表征和特异性的变化。学科层面的预测是通过汇总12个单独的问题回答得出的。我们最好的方法是使用OpenAI嵌入模型,每个问题的训练,包括输入中的问题,并将预测与投票系统相结合,达到84%的主题精度和1.0的ROC-AUC。此外,使用gpt - 40进行了零射击评估,得出了类似的结果,强调了紧凑的本地模型和大型开箱即用的llm的潜力。为了提高透明度,我们探索了可解释性方法。像gpt - 40这样的专有法学硕士没有提供直接的解释,OpenAI嵌入模型的可解释性有限。然而,本地可部署的llm提供了最高的可解释性。这突出了专有模型的性能和本地模型的可解释性之间的权衡。我们的研究结果验证了法学硕士对开放式问卷的自动分类,为ASD评估提供了一种可扩展的、具有成本效益的补充。这些结果表明对其他疾病的心理分析具有更广泛的适用性,促进了法学硕士在心理健康研究中的应用。
{"title":"Enhancing Psychological Assessments With Open-Ended Questionnaires and Large Language Models: An ASD Case Study.","authors":"Alberto Altozano, Maria Eleonora Minissi, Lucia Gomez-Zaragoza, Luna Maddalon, Mariano Alcaniz, Javier Marin-Morales","doi":"10.1109/JBHI.2025.3599643","DOIUrl":"10.1109/JBHI.2025.3599643","url":null,"abstract":"<p><p>Open-ended questionnaires allow respondents to express freely, capturing richer information than close-ended formats, but they are harder to analyze. Recent natural language processing advancements enable automatic assessment of open-ended responses, yet its use in psychological classification is underexplored. This study proposes a methodology using pre-trained large language models (LLMs) for automatic classification of open-ended questionnaires, applied to autism spectrum disorder (ASD) classification via parental reports. We compare multiple training strategies using transcribed responses from 51 parents (26 with typically developing children, 25 with ASD), exploring variations in model fine-tuning, input representation, and specificity. Subject-level predictions are derived by aggregating 12 individual question responses. Our best approach achieved 84% subject-wise accuracy and 1.0 ROC-AUC using an OpenAI embedding model, per-question training, including questions in the input, and combining the predictions with a voting system. In addition, a zero-shot evaluation using GPT-4o was conducted, yielding comparable results, underscoring the potential of both compact, local models and large out-of-the-box LLMs. To enhance transparency, we explored interpretability methods. Proprietary LLMs like GPT-4o offered no direct explanation, and OpenAI embedding models showed limited interpretability. However, locally deployable LLMs provided the highest interpretability. This highlights a trade-off between proprietary models' performance and local models' explainability. Our findings validate LLMs for automatically classifying open-ended questionnaires, offering a scalable, cost-effective complement for ASD assessment. These results suggest broader applicability for psychological analysis of other conditions, advancing LLM use in mental health research.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1707-1720"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144859134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Graph Learning With Multi-Hypergraph Reasoning Networks for Focal Liver Lesion Classification in Multimodal Magnetic Resonance Imaging. 基于多超图推理网络的多模态图学习在多模态磁共振成像中的肝病灶分类。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3639185
Shaocong Mo, Ming Cai, Lanfen Lin, Ruofeng Tong, Fang Wang, Qingqing Chen, Wenbin Ji, Yinhao Li, Hongjie Hu, Yen-Wei Chen

Multimodal magnetic resonance imaging (MRI) is instrumental in differentiating liver lesions. The major challenge involves modeling reliable connections and simultaneously learning complementary information across various MRI sequences. While previous studies have primarily focused on multimodal integration in a pair-wise manner using few modalities, our research seeks to advance a more comprehensive understanding of interaction modeling by establishing complex high-order correlations among the diverse modalities in multimodal MRI. In this paper, we introduce a multimodal graph learning with multi-hypergraph reasoning network to capture the full spectrum of both pair-wise and group-wise relationships among different modalities. Specifically, a weight-shared encoder extracts features from regions of interest (ROI) images across all modalities. Subsequently, a collection of uniform hypergraphs are constructed with varying vertex configurations, allowing for the modeling of not only pair-wise correlations but also the high-order collaborations for relational reasoning. Following information propagation through the hypergraph message passing, adaptive intra-modality fusion module is proposed to effectively fuse feature representations from different hypergraphs of the same modality. Finally, all refined features are concatenated to prepare for the classification task. Our experimental evaluations, including focal liver lesions classification using the LLD-MMRI2023 dataset and early recurrence prediction of hepatocellular carcinoma using our internal datasets, demonstrate that our method significantly surpasses the performance of existing approaches, indicating the effectiveness of our model in handling both pair-wise and group-wise interactions across multiple modalities.

多模态磁共振成像(MRI)有助于鉴别肝脏病变。主要的挑战包括建立可靠的连接,同时学习不同MRI序列的互补信息。虽然以前的研究主要集中在使用少量模态以成对的方式进行多模态整合,但我们的研究旨在通过在多模态MRI中建立不同模态之间复杂的高阶相关性来推进对相互作用建模的更全面理解。在本文中,我们引入了一个多模态图学习和多超图推理网络,以捕获不同模态之间的成对和群明智关系的全谱。具体来说,权重共享编码器从所有模态的感兴趣区域(ROI)图像中提取特征。随后,用不同的顶点配置构造了一组一致的超图,不仅可以建模成对关联,还可以建模用于关系推理的高阶协作。在超图消息传递的基础上,提出自适应模态内融合模块,有效融合相同模态的不同超图的特征表示。最后,将所有细化的特征连接起来,为分类任务做准备。我们的实验评估,包括使用LLD-MMRI2023数据集进行局灶性肝病变分类,以及使用我们的内部数据集进行肝细胞癌早期复发预测,表明我们的方法显著优于现有方法的性能,表明我们的模型在处理跨多种方式的配对和组交互方面都是有效的。
{"title":"Multimodal Graph Learning With Multi-Hypergraph Reasoning Networks for Focal Liver Lesion Classification in Multimodal Magnetic Resonance Imaging.","authors":"Shaocong Mo, Ming Cai, Lanfen Lin, Ruofeng Tong, Fang Wang, Qingqing Chen, Wenbin Ji, Yinhao Li, Hongjie Hu, Yen-Wei Chen","doi":"10.1109/JBHI.2025.3639185","DOIUrl":"10.1109/JBHI.2025.3639185","url":null,"abstract":"<p><p>Multimodal magnetic resonance imaging (MRI) is instrumental in differentiating liver lesions. The major challenge involves modeling reliable connections and simultaneously learning complementary information across various MRI sequences. While previous studies have primarily focused on multimodal integration in a pair-wise manner using few modalities, our research seeks to advance a more comprehensive understanding of interaction modeling by establishing complex high-order correlations among the diverse modalities in multimodal MRI. In this paper, we introduce a multimodal graph learning with multi-hypergraph reasoning network to capture the full spectrum of both pair-wise and group-wise relationships among different modalities. Specifically, a weight-shared encoder extracts features from regions of interest (ROI) images across all modalities. Subsequently, a collection of uniform hypergraphs are constructed with varying vertex configurations, allowing for the modeling of not only pair-wise correlations but also the high-order collaborations for relational reasoning. Following information propagation through the hypergraph message passing, adaptive intra-modality fusion module is proposed to effectively fuse feature representations from different hypergraphs of the same modality. Finally, all refined features are concatenated to prepare for the classification task. Our experimental evaluations, including focal liver lesions classification using the LLD-MMRI2023 dataset and early recurrence prediction of hepatocellular carcinoma using our internal datasets, demonstrate that our method significantly surpasses the performance of existing approaches, indicating the effectiveness of our model in handling both pair-wise and group-wise interactions across multiple modalities.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1404-1417"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145654299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Remote Heart Rate Estimation Network Based on Spatial-Temporal-Channel Learning From Facial Videos. 基于人脸视频时空通道学习的鲁棒远程心率估计网络。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3597997
Jun Yang, Chen Zhu, Renbiao Wu

Non-contact heart rate detection technology leverages changes in skin color to estimate heart rate, enhancing the convenience of health monitoring, particularly in situations requiring real-time, contact-free observation. However, current video-based methods face various limitations, including restricted feature extraction capabilities, redundant spatial information, and ineffective motion artifact processing. To address these problems, a novel end-to-end heart rate estimation network, Spatial-Temporal-Channel Network (STCNet), is proposed. Firstly, in order to solve the problem of redundant spatial information in current video-based heart rate estimation methods, a spatial attention learning (SAL) unit is designed to highlight the effective information of the facial region. Next, an improved temporal shift module (TSMP) with long-range temporal information perception is proposed. On this basis, A temporal-channel learning (TCL) unit is designed to achieve the interaction of information across different frames' channels, aiming to address the insufficient capability of existing models in extracting periodic features of heartbeat. Finally, combining the SAL and TCL units, a feature extraction block (FEB) is designed. A feature extraction network is constructed by stacking multiple layers of FEBs to achieve accurate heart rate estimation. Numerous experiments are conducted on the UBFC-rPPG dataset and the PURE dataset to verify the effectiveness and generalization ability of our model. Notably, compared to the state-of-the-art CIN-rPPG, our model achieves a 0.27 bpm reduction in mean absolute error (MAE) and a 0.19 bpm reduction in root mean square error (RMSE), in intra-dataset testing on the PURE dataset. Experimental results demonstrate that our proposed model outperforms other mainstream models.

非接触式心率检测技术利用皮肤颜色的变化来估计心率,增强了健康监测的便利性,特别是在需要实时、无接触观察的情况下。然而,目前基于视频的方法面临各种限制,包括限制的特征提取能力,冗余的空间信息和无效的运动伪影处理。为了解决这些问题,提出了一种新的端到端心率估计网络——时空信道网络(STCNet)。首先,为了解决当前基于视频的心率估计方法中空间信息冗余的问题,设计了空间注意学习(SAL)单元,突出显示面部区域的有效信息;在此基础上,提出了一种具有长时距信息感知的改进时移模块(TSMP)。在此基础上,设计了时序信道学习(TCL)单元,实现了不同帧信道间的信息交互,解决了现有模型在提取心跳周期特征方面能力不足的问题。最后,结合SAL和TCL单元,设计了特征提取块(FEB)。通过多层feb叠加构建特征提取网络,实现准确的心率估计。在UBFC-rPPG数据集和PURE数据集上进行了大量实验,验证了我们模型的有效性和泛化能力。值得注意的是,与最先进的CIN-rPPG相比,我们的模型在PURE数据集的数据集内测试中实现了平均绝对误差(MAE)减少0.27 bpm,均方根误差(RMSE)减少0.19 bpm。实验结果表明,该模型优于其他主流模型。
{"title":"Robust Remote Heart Rate Estimation Network Based on Spatial-Temporal-Channel Learning From Facial Videos.","authors":"Jun Yang, Chen Zhu, Renbiao Wu","doi":"10.1109/JBHI.2025.3597997","DOIUrl":"10.1109/JBHI.2025.3597997","url":null,"abstract":"<p><p>Non-contact heart rate detection technology leverages changes in skin color to estimate heart rate, enhancing the convenience of health monitoring, particularly in situations requiring real-time, contact-free observation. However, current video-based methods face various limitations, including restricted feature extraction capabilities, redundant spatial information, and ineffective motion artifact processing. To address these problems, a novel end-to-end heart rate estimation network, Spatial-Temporal-Channel Network (STCNet), is proposed. Firstly, in order to solve the problem of redundant spatial information in current video-based heart rate estimation methods, a spatial attention learning (SAL) unit is designed to highlight the effective information of the facial region. Next, an improved temporal shift module (TSMP) with long-range temporal information perception is proposed. On this basis, A temporal-channel learning (TCL) unit is designed to achieve the interaction of information across different frames' channels, aiming to address the insufficient capability of existing models in extracting periodic features of heartbeat. Finally, combining the SAL and TCL units, a feature extraction block (FEB) is designed. A feature extraction network is constructed by stacking multiple layers of FEBs to achieve accurate heart rate estimation. Numerous experiments are conducted on the UBFC-rPPG dataset and the PURE dataset to verify the effectiveness and generalization ability of our model. Notably, compared to the state-of-the-art CIN-rPPG, our model achieves a 0.27 bpm reduction in mean absolute error (MAE) and a 0.19 bpm reduction in root mean square error (RMSE), in intra-dataset testing on the PURE dataset. Experimental results demonstrate that our proposed model outperforms other mainstream models.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1258-1271"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144834977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Atlas Brain Network Classification Through Consistency Distillation and Complementary Information Fusion. 基于一致性蒸馏和互补信息融合的多图谱脑网络分类。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3610111
Jiaxing Xu, Mengcheng Lan, Xia Dong, Kai He, Wei Zhang, Qingtian Bian, Yiping Ke

Brain network analysis plays a crucial role in identifying distinctive patterns associated with neurological disorders. Functional magnetic resonance imaging (fMRI) enables the construction of brain networks by analyzing correlations in blood-oxygen-level-dependent (BOLD) signals across different brain regions, known as regions of interest (ROIs). These networks are typically constructed using atlases that parcellate the brain based on various hypotheses of functional and anatomical divisions. However, there is no standard atlas for brain network classification, leading to limitations in detecting abnormalities in disorders. Recent methods leveraging multiple atlases fail to ensure consistency across atlases and lack effective ROI-level information exchange, limiting their efficacy. To address these challenges, we propose the Atlas-Integrated Distillation and Fusion network (AIDFusion), a novel framework designed to enhance brain network classification using fMRI data. AIDFusion introduces a disentangle Transformer to filter out inconsistent atlas-specific information and distill meaningful cross-atlas connections. Additionally, it enforces subject- and population-level consistency constraints to improve cross-atlas coherence. To further enhance feature integration, AIDFusion incorporates an inter-atlas message-passing mechanism that facilitates the fusion of complementary information across brain regions. We evaluate AIDFusion on four resting-state fMRI datasets encompassing different neurological disorders. Experimental results demonstrate its superior classification performance and computational efficiency compared to state-of-the-art methods. Furthermore, a case study highlights AIDFusion's ability to extract interpretable patterns that align with established neuroscience findings, reinforcing its potential as a robust tool for multi-atlas brain network analysis.

脑网络分析在识别与神经系统疾病相关的独特模式方面起着至关重要的作用。功能性磁共振成像(fMRI)通过分析不同大脑区域(即感兴趣区域(roi))中血氧水平依赖(BOLD)信号的相关性来构建大脑网络。这些网络通常是用地图集构建的,地图集根据各种功能和解剖划分的假设将大脑包裹起来。然而,没有标准的脑网络分类图谱,导致在检测异常的障碍限制。最近利用多个地图集的方法无法确保地图集之间的一致性,并且缺乏有效的roi级信息交换,从而限制了它们的有效性。为了应对这些挑战,我们提出了Atlas-Integrated Distillation and Fusion network (AIDFusion),这是一种利用fMRI数据增强脑网络分类的新框架。AIDFusion引入了一个解缠绕变压器,过滤掉不一致的图谱特定信息,提取出有意义的跨图谱连接。此外,它加强了学科和人口水平的一致性约束,以提高跨图谱的一致性。为了进一步增强特征整合,AIDFusion结合了一个图谱间信息传递机制,促进了脑区域间互补信息的融合。我们在包含不同神经系统疾病的四个静息状态fMRI数据集上评估AIDFusion。实验结果表明,该方法具有较好的分类性能和计算效率。此外,一个案例研究强调了AIDFusion提取与已建立的神经科学发现相一致的可解释模式的能力,加强了其作为多图谱脑网络分析的强大工具的潜力。该代码可在https://github.com/AngusMonroe/AIDFusion上公开获得。
{"title":"Multi-Atlas Brain Network Classification Through Consistency Distillation and Complementary Information Fusion.","authors":"Jiaxing Xu, Mengcheng Lan, Xia Dong, Kai He, Wei Zhang, Qingtian Bian, Yiping Ke","doi":"10.1109/JBHI.2025.3610111","DOIUrl":"10.1109/JBHI.2025.3610111","url":null,"abstract":"<p><p>Brain network analysis plays a crucial role in identifying distinctive patterns associated with neurological disorders. Functional magnetic resonance imaging (fMRI) enables the construction of brain networks by analyzing correlations in blood-oxygen-level-dependent (BOLD) signals across different brain regions, known as regions of interest (ROIs). These networks are typically constructed using atlases that parcellate the brain based on various hypotheses of functional and anatomical divisions. However, there is no standard atlas for brain network classification, leading to limitations in detecting abnormalities in disorders. Recent methods leveraging multiple atlases fail to ensure consistency across atlases and lack effective ROI-level information exchange, limiting their efficacy. To address these challenges, we propose the Atlas-Integrated Distillation and Fusion network (AIDFusion), a novel framework designed to enhance brain network classification using fMRI data. AIDFusion introduces a disentangle Transformer to filter out inconsistent atlas-specific information and distill meaningful cross-atlas connections. Additionally, it enforces subject- and population-level consistency constraints to improve cross-atlas coherence. To further enhance feature integration, AIDFusion incorporates an inter-atlas message-passing mechanism that facilitates the fusion of complementary information across brain regions. We evaluate AIDFusion on four resting-state fMRI datasets encompassing different neurological disorders. Experimental results demonstrate its superior classification performance and computational efficiency compared to state-of-the-art methods. Furthermore, a case study highlights AIDFusion's ability to extract interpretable patterns that align with established neuroscience findings, reinforcing its potential as a robust tool for multi-atlas brain network analysis.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1568-1579"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145075230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Position Paper: Artificial Intelligence in Medical Image Analysis: Advances, Clinical Translation, and Emerging Frontiers. 立场文件:医学图像分析中的人工智能:进展、临床翻译和新兴前沿。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3649496
A S Panayides, H Chen, N D Filipovic, T Geroski, J Hou, K Lekadir, K Marias, G K Matsopoulos, G Papanastasiou, P Sarder, G Tourassi, S A Tsaftaris, H Fu, E Kyriacou, C P Loizou, M Zervakis, J H Saltz, F E Shamout, K C L Wong, J Yao, A Amini, D I Fotiadis, C S Pattichis, M S Pattichis

Over the past five years, artificial intelligence (AI) has introduced new models and methods for addressing the challenges associated with the broader adoption of AI models and systems in medicine. This paper reviews recent advances in AI for medical image and video analysis, outlines emerging paradigms, highlights pathways for successful clinical translation, and provides recommendations for future work. Hybrid Convolutional Neural Network (CNN) Transformer architectures now deliver state-of-the-art results in segmentation, classification, reconstruction, synthesis, and registration. Foundation and generative AI models enable the use of transfer learning to smaller datasets with limited ground truth. Federated learning supports privacy-preserving collaboration across institutions. Explainable and trustworthy AI approaches have become essential to foster clinician trust, ensure regulatory compliance, and facilitate ethical deployment. Together, these developments pave the way for integrating AI into radiology, pathology, and wider healthcare workflows.

在过去五年中,人工智能(AI)引入了新的模型和方法,以应对与在医学中广泛采用AI模型和系统相关的挑战。本文回顾了人工智能在医学图像和视频分析方面的最新进展,概述了新兴范例,重点介绍了成功临床翻译的途径,并为未来的工作提供了建议。混合卷积神经网络(CNN)变压器架构现在在分割、分类、重建、合成和注册方面提供了最先进的结果。基础和生成人工智能模型可以将迁移学习用于具有有限基础真理的较小数据集。联邦学习支持跨机构的隐私保护协作。可解释和可信赖的人工智能方法对于培养临床医生的信任、确保法规遵守和促进道德部署至关重要。总之,这些发展为将人工智能集成到放射学、病理学和更广泛的医疗保健工作流程中铺平了道路。
{"title":"Position Paper: Artificial Intelligence in Medical Image Analysis: Advances, Clinical Translation, and Emerging Frontiers.","authors":"A S Panayides, H Chen, N D Filipovic, T Geroski, J Hou, K Lekadir, K Marias, G K Matsopoulos, G Papanastasiou, P Sarder, G Tourassi, S A Tsaftaris, H Fu, E Kyriacou, C P Loizou, M Zervakis, J H Saltz, F E Shamout, K C L Wong, J Yao, A Amini, D I Fotiadis, C S Pattichis, M S Pattichis","doi":"10.1109/JBHI.2025.3649496","DOIUrl":"10.1109/JBHI.2025.3649496","url":null,"abstract":"<p><p>Over the past five years, artificial intelligence (AI) has introduced new models and methods for addressing the challenges associated with the broader adoption of AI models and systems in medicine. This paper reviews recent advances in AI for medical image and video analysis, outlines emerging paradigms, highlights pathways for successful clinical translation, and provides recommendations for future work. Hybrid Convolutional Neural Network (CNN) Transformer architectures now deliver state-of-the-art results in segmentation, classification, reconstruction, synthesis, and registration. Foundation and generative AI models enable the use of transfer learning to smaller datasets with limited ground truth. Federated learning supports privacy-preserving collaboration across institutions. Explainable and trustworthy AI approaches have become essential to foster clinician trust, ensure regulatory compliance, and facilitate ethical deployment. Together, these developments pave the way for integrating AI into radiology, pathology, and wider healthcare workflows.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1187-1202"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145878178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant Features. MTS-LOF:通过闭塞不变特征进行医学时间序列表征学习。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2024.3373439
Huayu Li, Ana S Carreon-Rascon, Xiwen Chen, Geng Yuan, Ao Li

Medical time series data are indispensable in healthcare, providing critical insights for disease diagnosis, treatment planning, and patient management. The exponential growth in data complexity, driven by advanced sensor technologies, has presented challenges related to data labeling. Self-supervised learning (SSL) has emerged as a transformative approach to address these challenges, eliminating the need for extensive human annotation. In this study, we introduce a novel framework for Medical Time Series Representation Learning, known as MTS-LOF. MTS-LOF leverages the strengths of Joint-Embedding SSL and Masked Autoencoder (MAE) methods, offering a unique approach to representation learning for medical time series data. By combining these techniques, MTS-LOF enhances the potential of healthcare applications by providing more sophisticated, context-rich representations. Additionally, MTS-LOF employs a multi-masking strategy to facilitate occlusion-invariant feature learning. This approach allows the model to create multiple views of the data by masking portions of it. By minimizing the discrepancy between the representations of these masked patches and the fully visible patches, MTS-LOF learns to capture rich contextual information within medical time series datasets. The results of experiments conducted on diverse medical time series datasets demonstrate the superiority of MTS-LOF over other methods. These findings hold promise for significantly enhancing healthcare applications by improving representation learning. Furthermore, our work delves into the integration of Joint-Embedding SSL and MAE techniques, shedding light on the intricate interplay between temporal and structural dependencies in healthcare data. This understanding is crucial, as it allows us to grasp the complexities of healthcare data analysis.

医疗时间序列数据在医疗保健领域不可或缺,为疾病诊断、治疗计划和患者管理提供了重要的洞察力。在先进传感器技术的推动下,数据复杂性呈指数级增长,这给数据标注带来了挑战。自我监督学习(SSL)已成为应对这些挑战的变革性方法,无需大量人工标注。在本研究中,我们介绍了一种用于医学时间序列表示学习的新型框架,即 MTS-LOF。MTS-LOF 充分利用了联合嵌入 SSL 和掩码自动编码器 (MAE) 方法的优势,为医学时间序列数据的表示学习提供了一种独特的方法。通过结合这些技术,MTS-LOF 可提供更复杂、上下文丰富的表示,从而增强医疗保健应用的潜力。此外,MTS-LOF 还采用了多重掩码策略,以促进与闭塞无关的特征学习。这种方法允许模型通过屏蔽部分数据来创建多个数据视图。MTS-LOF 通过最大限度地减少这些遮挡斑块与完全可见斑块的表征之间的差异,学会捕捉医疗时间序列数据集中丰富的上下文信息。在各种医疗时间序列数据集上进行的实验结果表明,MTS-LOF 优于其他方法。这些发现为通过改进表征学习来显著提高医疗保健应用带来了希望。此外,我们的工作还深入研究了联合嵌入 SSL 和 MAE 技术的整合,揭示了医疗数据中时间和结构依赖性之间错综复杂的相互作用。这种理解至关重要,因为它能让我们掌握医疗数据分析的复杂性。
{"title":"MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant Features.","authors":"Huayu Li, Ana S Carreon-Rascon, Xiwen Chen, Geng Yuan, Ao Li","doi":"10.1109/JBHI.2024.3373439","DOIUrl":"10.1109/JBHI.2024.3373439","url":null,"abstract":"<p><p>Medical time series data are indispensable in healthcare, providing critical insights for disease diagnosis, treatment planning, and patient management. The exponential growth in data complexity, driven by advanced sensor technologies, has presented challenges related to data labeling. Self-supervised learning (SSL) has emerged as a transformative approach to address these challenges, eliminating the need for extensive human annotation. In this study, we introduce a novel framework for Medical Time Series Representation Learning, known as MTS-LOF. MTS-LOF leverages the strengths of Joint-Embedding SSL and Masked Autoencoder (MAE) methods, offering a unique approach to representation learning for medical time series data. By combining these techniques, MTS-LOF enhances the potential of healthcare applications by providing more sophisticated, context-rich representations. Additionally, MTS-LOF employs a multi-masking strategy to facilitate occlusion-invariant feature learning. This approach allows the model to create multiple views of the data by masking portions of it. By minimizing the discrepancy between the representations of these masked patches and the fully visible patches, MTS-LOF learns to capture rich contextual information within medical time series datasets. The results of experiments conducted on diverse medical time series datasets demonstrate the superiority of MTS-LOF over other methods. These findings hold promise for significantly enhancing healthcare applications by improving representation learning. Furthermore, our work delves into the integration of Joint-Embedding SSL and MAE techniques, shedding light on the intricate interplay between temporal and structural dependencies in healthcare data. This understanding is crucial, as it allows us to grasp the complexities of healthcare data analysis.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"958-969"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140039274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuro-BERT: Rethinking Masked Autoencoding for Self-Supervised Neurological Pretraining. Neuro-BERT:反思用于自我监督神经学预训练的屏蔽自动编码。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2024.3415959
Di Wu, Siyuan Li, Jie Yang, Mohamad Sawan

Deep learning associated with neurological signals is poised to drive major advancements in diverse fields such as medical diagnostics, neurorehabilitation, and brain-computer interfaces. The challenge in harnessing the full potential of these signals lies in the dependency on extensive, high-quality annotated data, which is often scarce and expensive to acquire, requiring specialized infrastructure and domain expertise. To address the appetite for data in deep learning, we present Neuro-BERT, a self-supervised pre-training framework of neurological signals based on masked autoencoding in the Fourier domain. The intuition behind our approach is simple: frequency and phase distribution of neurological signals can reveal intricate neurological activities. We propose a novel pre-training task dubbed Fourier Inversion Prediction (FIP), which randomly masks out a portion of the input signal and then predicts the missing information using the Fourier inversion theorem. Pre-trained models can be potentially used for various downstream tasks such as sleep stage classification and gesture recognition. Unlike contrastive-based methods, which strongly rely on carefully hand-crafted augmentations and siamese structure, our approach works reasonably well with a simple transformer encoder with no augmentation requirements. By evaluating our method on several benchmark datasets, we show that Neuro-BERT improves downstream neurological-related tasks by a large margin.

与神经信号相关的深度学习有望推动医疗诊断、神经康复和脑机接口等不同领域的重大进展。利用这些信号的全部潜力所面临的挑战在于对大量高质量注释数据的依赖,而这些数据往往稀缺且获取成本高昂,需要专门的基础设施和领域专业知识。为了解决深度学习对数据的需求,我们提出了 Neuro-BERT,这是一种基于傅立叶域屏蔽自动编码的神经信号自监督预训练框架。我们的方法背后的直觉很简单:神经信号的频率和相位分布可以揭示错综复杂的神经活动。我们提出了一种被称为傅立叶反转预测(FIP)的新颖预训练任务,即随机屏蔽掉部分输入信号,然后利用傅立叶反转定理预测缺失的信息。预训练模型可用于睡眠阶段分类和手势识别等各种下游任务。基于对比的方法主要依赖于精心手工制作的增强和连体结构,而我们的方法与之不同,只需使用简单的变压器编码器即可,无需增强。通过在多个基准数据集上对我们的方法进行评估,我们发现 Neuro-BERT 能显著改善下游神经相关任务。
{"title":"Neuro-BERT: Rethinking Masked Autoencoding for Self-Supervised Neurological Pretraining.","authors":"Di Wu, Siyuan Li, Jie Yang, Mohamad Sawan","doi":"10.1109/JBHI.2024.3415959","DOIUrl":"10.1109/JBHI.2024.3415959","url":null,"abstract":"<p><p>Deep learning associated with neurological signals is poised to drive major advancements in diverse fields such as medical diagnostics, neurorehabilitation, and brain-computer interfaces. The challenge in harnessing the full potential of these signals lies in the dependency on extensive, high-quality annotated data, which is often scarce and expensive to acquire, requiring specialized infrastructure and domain expertise. To address the appetite for data in deep learning, we present Neuro-BERT, a self-supervised pre-training framework of neurological signals based on masked autoencoding in the Fourier domain. The intuition behind our approach is simple: frequency and phase distribution of neurological signals can reveal intricate neurological activities. We propose a novel pre-training task dubbed Fourier Inversion Prediction (FIP), which randomly masks out a portion of the input signal and then predicts the missing information using the Fourier inversion theorem. Pre-trained models can be potentially used for various downstream tasks such as sleep stage classification and gesture recognition. Unlike contrastive-based methods, which strongly rely on carefully hand-crafted augmentations and siamese structure, our approach works reasonably well with a simple transformer encoder with no augmentation requirements. By evaluating our method on several benchmark datasets, we show that Neuro-BERT improves downstream neurological-related tasks by a large margin.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"970-980"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141418695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Biomedical and Health Informatics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1