首页 > 最新文献

IEEE Journal of Biomedical and Health Informatics最新文献

英文 中文
IIMCNet: Intra- and Inter-Modality Correlation Network for Hybrid EEG-fNIRS Brain-Computer Interface. 混合脑机接口的模态内和模态间关联网络。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-01 DOI: 10.1109/JBHI.2025.3594203
Xiaoyang Yuan, Yan Zhang, Peter Rolfe

Hybrid Brain-Computer Interface (BCI) enhances accuracy and reliability by leveraging the complementary information provided by multi-modality signal fusion. EEG-fNIRS, a fusion of electroencephalogram (EEG) and functional near-infrared spectroscopy (fNIRS), have emerged as the suitable techniques for real-world BCI applications due to their portability and economic viability. Existing methods typically focus on the high-level feature representation with late-fusion or early-fusion strategies during the recognition tasks. However, they usually overlook the joint feature extraction of both intra-modality and inter-modality, which is crucial for optimizing BCI performance. In this study, we introduce an Intra- and Inter-modality Correlation Network (IIMCNet) to integrate both the inherent features derived from individual modalities: EEG, deoxygenated hemoglobin (HbR), and oxygenated hemoglobin (HbO), as well as the cross-modality features between EEG-HbR, EEG-HbO, and HbR-HbO data. The intra-modality correlation features are generated using a late fusion method (Intra-net), which combines the uni-modality features extracted by E-Net and f-Net. Concurrently, the inter-modality correlation features are extracted using an early fusion method (Inter-net). Inter-net is consist of three dilated convolution-based C-Nets that focus on neurovascular coupling across modalities. Finally, three intra-modality features, three inter-modality features, and the concatenate hybrid feature are fed into deep supervision module to enhance robustness and accuracy. Experiment results demonstrate the IIMCNet exhibits superior performance compared to methods that rely solely on either intra-modality or inter-modality correlation networks. Furthermore, IIMCNet outperforms other state-of-the-art methods in motor imagery and mental arithmetic tasks, respectively.

混合脑机接口(BCI)利用多模态信号融合提供的互补信息来提高准确性和可靠性。EEG-fNIRS是脑电图(EEG)和功能近红外光谱(fNIRS)的融合,由于其便携性和经济可行性,已成为现实世界BCI应用的合适技术。现有的方法主要集中在识别过程中采用后期或早期融合策略对高级特征进行表征。然而,他们往往忽视了对BCI性能优化至关重要的模态内和模态间的联合特征提取。在这项研究中,我们引入了一个模态内和模态间相关网络(IIMCNet)来整合来自单个模态的固有特征:EEG、脱氧血红蛋白(HbR)和氧合血红蛋白(HbO),以及EEG-HbR、EEG-HbO和HbR-HbO数据之间的跨模态特征。结合E-Net和f-Net提取的单模态特征,采用后期融合方法(Intra-net)生成模态内相关特征。同时,采用早期融合方法(Inter-net)提取模态间相关特征。internet由三个基于扩展卷积的c -net组成,这些c -net专注于跨模式的神经血管耦合。最后,将3个模态内特征、3个模态间特征和连接混合特征输入到深度监督模块中,以提高鲁棒性和准确性。实验结果表明,与仅依赖于模态内或模态间相关网络的方法相比,IIMCNet具有优越的性能。此外,IIMCNet在运动意象和心算任务方面分别优于其他最先进的方法。(代码可从github.com/Y-xiaoyang/IIMCNet获得)。
{"title":"IIMCNet: Intra- and Inter-Modality Correlation Network for Hybrid EEG-fNIRS Brain-Computer Interface.","authors":"Xiaoyang Yuan, Yan Zhang, Peter Rolfe","doi":"10.1109/JBHI.2025.3594203","DOIUrl":"10.1109/JBHI.2025.3594203","url":null,"abstract":"<p><p>Hybrid Brain-Computer Interface (BCI) enhances accuracy and reliability by leveraging the complementary information provided by multi-modality signal fusion. EEG-fNIRS, a fusion of electroencephalogram (EEG) and functional near-infrared spectroscopy (fNIRS), have emerged as the suitable techniques for real-world BCI applications due to their portability and economic viability. Existing methods typically focus on the high-level feature representation with late-fusion or early-fusion strategies during the recognition tasks. However, they usually overlook the joint feature extraction of both intra-modality and inter-modality, which is crucial for optimizing BCI performance. In this study, we introduce an Intra- and Inter-modality Correlation Network (IIMCNet) to integrate both the inherent features derived from individual modalities: EEG, deoxygenated hemoglobin (HbR), and oxygenated hemoglobin (HbO), as well as the cross-modality features between EEG-HbR, EEG-HbO, and HbR-HbO data. The intra-modality correlation features are generated using a late fusion method (Intra-net), which combines the uni-modality features extracted by E-Net and f-Net. Concurrently, the inter-modality correlation features are extracted using an early fusion method (Inter-net). Inter-net is consist of three dilated convolution-based C-Nets that focus on neurovascular coupling across modalities. Finally, three intra-modality features, three inter-modality features, and the concatenate hybrid feature are fed into deep supervision module to enhance robustness and accuracy. Experiment results demonstrate the IIMCNet exhibits superior performance compared to methods that rely solely on either intra-modality or inter-modality correlation networks. Furthermore, IIMCNet outperforms other state-of-the-art methods in motor imagery and mental arithmetic tasks, respectively.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"113-124"},"PeriodicalIF":6.8,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144764852","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Replacing Attention With Modality-Wise Convolution for Energy-Efficient PPG-Based Heart Rate Estimation Using Knowledge Distillation. 用模态卷积代替注意力实现基于ppg的高效心率估计。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-01-01 DOI: 10.1109/JBHI.2025.3580474
Panagiotis Kasnesis, Lazaros Toumanidis, Daniele Jahier Pagliari, Alessio Burrello

Continuous monitoring of Hearth Rate (HR) based on photoplethysmography (PPG) sensors is an essential capability of nearly all wrist-worn devices. However, arm movements lead to the creation of Motion Artifacts (MA), affecting the accuracy of HR tracking using PPG sensors. This problem is commonly tackled by exploiting the recorded accelerometer data to correlate them with the PPG signal and eventually clean it. Thus, automatic fusion techniques based on Deep Learning (DL) algorithms have been proposed, but they are considered too large and complex to be deployed on wearable devices. The current work presents a novel and lightweight DL architecture, PULSE, improving sensor fusion by applying a multi-head cross-attention layer to the extracted temporal features. Moreover, we propose a relation-based knowledge distillation mechanism to pass PULSE's knowledge to a student network that uses modality-wise convolutions to replace the attention module and mimic the teacher's performance with 5× fewer parameters. The teacher and student are evaluated on two datasets: a) PPG-DaLiA the most extensive available dataset, with PULSE achieving close performance to the best state-of-the-art model, and b) WESAD with PULSE reducing the mean absolute error by 22.6%. The student model is further compressed using post-training quantization and deployed on two commercial-off-the-shelf microcontrollers, demonstrating its suitability for real-time execution, having a close-to-state-of-the-art MAE of 4.81 BPM (+0.40 BPM) on the PPG-DaLiA, but a 10.9× lower memory footprint of 37.9 kB, and consuming 45.9× lower energy (0.577 mJ).

基于光电容积脉搏波(PPG)传感器的连续监测炉底率(HR)是几乎所有腕带设备的基本功能。然而,手臂运动导致运动伪影(MA)的产生,影响使用PPG传感器进行HR跟踪的准确性。这个问题通常是通过利用记录的加速度计数据将它们与PPG信号相关联并最终清除它来解决的。因此,基于深度学习(DL)算法的自动融合技术已经被提出,但它们被认为过于庞大和复杂,无法部署在可穿戴设备上。目前的工作提出了一种新的轻量级DL架构PULSE,通过对提取的时间特征应用多头交叉注意层来改善传感器融合。此外,我们提出了一种基于关系的知识蒸馏机制,将PULSE的知识传递给学生网络,该网络使用模态智能卷积取代注意力模块,并以少5倍的参数模拟教师的表现。教师和学生在两个数据集上进行评估:a) PPG-DaLiA是最广泛的可用数据集,其中PULSE的性能接近最先进的模型,b) WESAD的PULSE将平均绝对误差降低了22.6%。学生模型使用训练后量化进一步压缩,并部署在两个商用现成的微控制器上,证明了其实时执行的适用性,在PPG-DaLiA上具有接近最先进的4.81 BPM (+0.40 BPM)的MAE,但内存占用降低了10.9倍,为37.9 kB,能耗降低了45.9倍(0.577 mJ)。
{"title":"Replacing Attention With Modality-Wise Convolution for Energy-Efficient PPG-Based Heart Rate Estimation Using Knowledge Distillation.","authors":"Panagiotis Kasnesis, Lazaros Toumanidis, Daniele Jahier Pagliari, Alessio Burrello","doi":"10.1109/JBHI.2025.3580474","DOIUrl":"10.1109/JBHI.2025.3580474","url":null,"abstract":"<p><p>Continuous monitoring of Hearth Rate (HR) based on photoplethysmography (PPG) sensors is an essential capability of nearly all wrist-worn devices. However, arm movements lead to the creation of Motion Artifacts (MA), affecting the accuracy of HR tracking using PPG sensors. This problem is commonly tackled by exploiting the recorded accelerometer data to correlate them with the PPG signal and eventually clean it. Thus, automatic fusion techniques based on Deep Learning (DL) algorithms have been proposed, but they are considered too large and complex to be deployed on wearable devices. The current work presents a novel and lightweight DL architecture, PULSE, improving sensor fusion by applying a multi-head cross-attention layer to the extracted temporal features. Moreover, we propose a relation-based knowledge distillation mechanism to pass PULSE's knowledge to a student network that uses modality-wise convolutions to replace the attention module and mimic the teacher's performance with 5× fewer parameters. The teacher and student are evaluated on two datasets: a) PPG-DaLiA the most extensive available dataset, with PULSE achieving close performance to the best state-of-the-art model, and b) WESAD with PULSE reducing the mean absolute error by 22.6%. The student model is further compressed using post-training quantization and deployed on two commercial-off-the-shelf microcontrollers, demonstrating its suitability for real-time execution, having a close-to-state-of-the-art MAE of 4.81 BPM (+0.40 BPM) on the PPG-DaLiA, but a 10.9× lower memory footprint of 37.9 kB, and consuming 45.9× lower energy (0.577 mJ).</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"339-352"},"PeriodicalIF":6.8,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144316793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Variance Extrapolated Class-Imbalance-Aware Domain Adaptive Myocardial Segmentation in Multi-Sequence Cardiac MRI. 基于方差外推类不平衡感知域的多序列心脏MRI自适应心肌分割。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-31 DOI: 10.1109/JBHI.2025.3649765
Fangxu Xing, Xiaofeng Liu, Iman Aganj, Georges El Fakhri, Panki Kim, Byoung Wook Choi, Jonghye Woo

Fully automated myocardial segmentation from cardiac magnetic resonance imaging (MRI) is vital for efficient diagnosis and treatment planning. Although numerous automated methods have been proposed, they typically focus on single MRI sequences and therefore have difficulties in generalizing across vendors and across cardiac MRI protocols. Simultaneous analysis of complementary cardiac MRI sequences, such as cine, T1 mapping, and late gadolinium enhancement (LGE) MRI, remains challenging due to their distinct image characteristics and scanner-specific variations. To address these issues, we propose an unsupervised domain adaptation approach that allows robust myocardial segmentation across multi-vendor cine, T1, and LGE MRI data. In particular, we introduce a class- imbalance self-training framework to transfer information learned from a source domain with labels to any unlabeled target domain, while maintaining consistent performance across different MRI sequences. Our framework iteratively refines segmentation accuracy by generating pseudo-labels for target data using a hardness-aware strategy, thus effectively addressing the problem of class imbalance in cardiac MRI segmentation. To mitigate data scarcity following pseudo-label selection, we employ a variance-guided vicinal feature extrapolation, which expands data points in the feature space into a probabilistic distribution. This, in turn, facilitates joint source-target training by generating a larger intersection in the feature space. Experimental results demonstrate that our framework outperforms existing methods when assessed using the Dice coefficient and Hausdorff distance. Our framework enables cardiac evaluation across MRI protocols without sequence-specific manual annotations.

全自动心肌分割从心脏磁共振成像(MRI)是至关重要的有效诊断和治疗计划。尽管已经提出了许多自动化方法,但它们通常集中在单个MRI序列上,因此在跨供应商和跨心脏MRI协议的推广方面存在困难。同时分析互补的心脏MRI序列,如电影、T1成像和晚期钆增强(LGE) MRI,由于其不同的图像特征和扫描仪特异性变化,仍然具有挑战性。为了解决这些问题,我们提出了一种无监督域自适应方法,允许跨多供应商电影、T1和LGE MRI数据进行稳健的心肌分割。特别是,我们引入了一个类不平衡自训练框架,将从带标签的源域学习到的信息转移到任何未标记的目标域,同时在不同的MRI序列中保持一致的性能。我们的框架通过使用硬度感知策略为目标数据生成伪标签来迭代地改进分割精度,从而有效地解决了心脏MRI分割中的类不平衡问题。为了减轻伪标签选择后的数据稀缺性,我们采用方差引导的邻近特征外推,将特征空间中的数据点扩展为概率分布。这反过来又通过在特征空间中生成更大的交集来促进源-目标联合训练。实验结果表明,在使用Dice系数和Hausdorff距离进行评估时,我们的框架优于现有的方法。我们的框架可以跨MRI协议进行心脏评估,而无需特定序列的手动注释。
{"title":"Variance Extrapolated Class-Imbalance-Aware Domain Adaptive Myocardial Segmentation in Multi-Sequence Cardiac MRI.","authors":"Fangxu Xing, Xiaofeng Liu, Iman Aganj, Georges El Fakhri, Panki Kim, Byoung Wook Choi, Jonghye Woo","doi":"10.1109/JBHI.2025.3649765","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649765","url":null,"abstract":"<p><p>Fully automated myocardial segmentation from cardiac magnetic resonance imaging (MRI) is vital for efficient diagnosis and treatment planning. Although numerous automated methods have been proposed, they typically focus on single MRI sequences and therefore have difficulties in generalizing across vendors and across cardiac MRI protocols. Simultaneous analysis of complementary cardiac MRI sequences, such as cine, T1 mapping, and late gadolinium enhancement (LGE) MRI, remains challenging due to their distinct image characteristics and scanner-specific variations. To address these issues, we propose an unsupervised domain adaptation approach that allows robust myocardial segmentation across multi-vendor cine, T1, and LGE MRI data. In particular, we introduce a class- imbalance self-training framework to transfer information learned from a source domain with labels to any unlabeled target domain, while maintaining consistent performance across different MRI sequences. Our framework iteratively refines segmentation accuracy by generating pseudo-labels for target data using a hardness-aware strategy, thus effectively addressing the problem of class imbalance in cardiac MRI segmentation. To mitigate data scarcity following pseudo-label selection, we employ a variance-guided vicinal feature extrapolation, which expands data points in the feature space into a probabilistic distribution. This, in turn, facilitates joint source-target training by generating a larger intersection in the feature space. Experimental results demonstrate that our framework outperforms existing methods when assessed using the Dice coefficient and Hausdorff distance. Our framework enables cardiac evaluation across MRI protocols without sequence-specific manual annotations.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145878290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SMFusion: Semantic-Preserving Fusion of Multimodal Medical Images for Enhanced Clinical Diagnosis. SMFusion:保留语义的多模态医学图像融合以增强临床诊断。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-31 DOI: 10.1109/JBHI.2025.3649749
Haozhe Xiang, Han Zhang, Yu Cheng, Xiongwen Quan, Wanwan Huang

Multimodal medical image fusion plays a crucial role in medical diagnosis by integrating complementary information from different modalities to enhance image readability and clinical applicability. However, existing methods mainly follow computer vision standards for feature extraction and fusion strategy formulation, overlooking the rich semantic information inherent in medical images. To address this limitation, we propose a novel semantic-guided medical image fusion approach that, for the first time, incorporates medical prior knowledge into the fusion process. Specifically, we construct a publicly available multimodal medical image-text dataset, upon which text descriptions generated by BiomedGPT are encoded and semantically aligned with image features in a high-dimensional space via a semantic interaction alignment module. During this process, a cross attention based linear transformation automatically maps the relationship between textual and visual features to facilitate comprehensive learning. The aligned features are then embedded into a text-injection module for further feature-level fusion. Unlike traditional methods, we further generate diagnostic reports from the fused images to assess the preservation of medical information. Additionally, we design a medical semantic loss function to enhance the retention of textual cues from the source images. Experimental results on test datasets demonstrate that the proposed method achieves superior performance in both qualitative and quantitative evaluations while preserving more critical medical information.

多模态医学图像融合通过整合不同模态的互补信息,提高图像的可读性和临床适用性,在医学诊断中起着至关重要的作用。然而,现有的方法主要遵循计算机视觉标准进行特征提取和融合策略制定,忽略了医学图像所固有的丰富的语义信息。为了解决这一限制,我们提出了一种新的语义引导医学图像融合方法,该方法首次将医学先验知识纳入融合过程。具体来说,我们构建了一个公开可用的多模态医学图像-文本数据集,在此数据集上,通过语义交互对齐模块对生物gpt生成的文本描述进行编码,并在高维空间中与图像特征进行语义对齐。在此过程中,基于交叉注意的线性转换自动映射文本和视觉特征之间的关系,以促进全面学习。然后将对齐的特征嵌入到文本注入模块中,以进一步进行特征级融合。与传统方法不同,我们进一步从融合的图像中生成诊断报告,以评估医疗信息的保存情况。此外,我们设计了一个医学语义损失函数来增强源图像文本线索的保留。在测试数据集上的实验结果表明,该方法在保留更多关键医学信息的同时,在定性和定量评估方面都取得了优异的性能。
{"title":"SMFusion: Semantic-Preserving Fusion of Multimodal Medical Images for Enhanced Clinical Diagnosis.","authors":"Haozhe Xiang, Han Zhang, Yu Cheng, Xiongwen Quan, Wanwan Huang","doi":"10.1109/JBHI.2025.3649749","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649749","url":null,"abstract":"<p><p>Multimodal medical image fusion plays a crucial role in medical diagnosis by integrating complementary information from different modalities to enhance image readability and clinical applicability. However, existing methods mainly follow computer vision standards for feature extraction and fusion strategy formulation, overlooking the rich semantic information inherent in medical images. To address this limitation, we propose a novel semantic-guided medical image fusion approach that, for the first time, incorporates medical prior knowledge into the fusion process. Specifically, we construct a publicly available multimodal medical image-text dataset, upon which text descriptions generated by BiomedGPT are encoded and semantically aligned with image features in a high-dimensional space via a semantic interaction alignment module. During this process, a cross attention based linear transformation automatically maps the relationship between textual and visual features to facilitate comprehensive learning. The aligned features are then embedded into a text-injection module for further feature-level fusion. Unlike traditional methods, we further generate diagnostic reports from the fused images to assess the preservation of medical information. Additionally, we design a medical semantic loss function to enhance the retention of textual cues from the source images. Experimental results on test datasets demonstrate that the proposed method achieves superior performance in both qualitative and quantitative evaluations while preserving more critical medical information.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145878117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Position Paper: Artificial Intelligence in Medical Image Analysis: Advances, Clinical Translation, and Emerging Frontiers. 立场文件:医学图像分析中的人工智能:进展、临床翻译和新兴前沿。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-31 DOI: 10.1109/JBHI.2025.3649496
A S Panayides, H Chen, N D Filipovic, T Geroski, J Hou, K Lekadir, K Marias, G K Matsopoulos, G Papanastasiou, P Sarder, G Tourassi, S A Tsaftaris, H Fu, E Kyriacou, C P Loizou, M Zervakis, J H Saltz, F E Shamout, K C L Wong, J Yao, A Amini, D I Fotiadis, C S Pattichis, M S Pattichis

Over the past five years, artificial intelligence (AI) has introduced new models and methods for addressing the challenges associated with the broader adoption of AI models and systems in medicine. This paper reviews recent advances in AI for medical image and video analysis, outlines emerging paradigms, highlights pathways for successful clinical translation, and provides recommendations for future work. Hybrid Convolutional Neural Network (CNN) Transformer architectures now deliver state-of-the-art results in segmentation, classification, reconstruction, synthesis, and registration. Foundation and generative AI models enable the use of transfer learning to smaller datasets with limited ground truth. Federated learning supports privacy-preserving collaboration across institutions. Explainable and trustworthy AI approaches have become essential to foster clinician trust, ensure regulatory compliance, and facilitate ethical deployment. Together, these developments pave the way for integrating AI into radiology, pathology, and wider healthcare workflows.

在过去五年中,人工智能(AI)引入了新的模型和方法,以应对与在医学中广泛采用AI模型和系统相关的挑战。本文回顾了人工智能在医学图像和视频分析方面的最新进展,概述了新兴范例,重点介绍了成功临床翻译的途径,并为未来的工作提供了建议。混合卷积神经网络(CNN)变压器架构现在在分割、分类、重建、合成和注册方面提供了最先进的结果。基础和生成人工智能模型可以将迁移学习用于具有有限基础真理的较小数据集。联邦学习支持跨机构的隐私保护协作。可解释和可信赖的人工智能方法对于培养临床医生的信任、确保法规遵守和促进道德部署至关重要。总之,这些发展为将人工智能集成到放射学、病理学和更广泛的医疗保健工作流程中铺平了道路。
{"title":"Position Paper: Artificial Intelligence in Medical Image Analysis: Advances, Clinical Translation, and Emerging Frontiers.","authors":"A S Panayides, H Chen, N D Filipovic, T Geroski, J Hou, K Lekadir, K Marias, G K Matsopoulos, G Papanastasiou, P Sarder, G Tourassi, S A Tsaftaris, H Fu, E Kyriacou, C P Loizou, M Zervakis, J H Saltz, F E Shamout, K C L Wong, J Yao, A Amini, D I Fotiadis, C S Pattichis, M S Pattichis","doi":"10.1109/JBHI.2025.3649496","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649496","url":null,"abstract":"<p><p>Over the past five years, artificial intelligence (AI) has introduced new models and methods for addressing the challenges associated with the broader adoption of AI models and systems in medicine. This paper reviews recent advances in AI for medical image and video analysis, outlines emerging paradigms, highlights pathways for successful clinical translation, and provides recommendations for future work. Hybrid Convolutional Neural Network (CNN) Transformer architectures now deliver state-of-the-art results in segmentation, classification, reconstruction, synthesis, and registration. Foundation and generative AI models enable the use of transfer learning to smaller datasets with limited ground truth. Federated learning supports privacy-preserving collaboration across institutions. Explainable and trustworthy AI approaches have become essential to foster clinician trust, ensure regulatory compliance, and facilitate ethical deployment. Together, these developments pave the way for integrating AI into radiology, pathology, and wider healthcare workflows.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145878178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Contrastive Prototype Learning for Resilient Brain Tumor Segmentation With Missing Modalities. 基于多模态对比原型学习的缺失模态弹性脑肿瘤分割。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-31 DOI: 10.1109/JBHI.2025.3649819
Heran Xi, Yu Ye, Jinghua Zhu, Jinbao Li

Multimodal fusion is an effective solution for holistic brain tumor diagnosis; however, it faces challenges under missing modalities. Traditional multi-encoder architectures can easily capture modality-specific features, while single-encoder architectures readily obtain modality-shared features. The reverse, however, is challenging. In this paper, we propose a two-stage dual-view prototype learning framework to extract the modality-specific feature and the class-specific feature simultaneously. In the first stage, we utilize the Transformer decoder to learn the modality-prototypes that are used to optimize the modality reconstruction task. A masked autoencoder is introduced to generate shared features of incomplete modalities. The learned modality-prototypes that contain modality-specific features are blended with the modality-share features for the reconstruction process. In the second stage, we learn the class-prototypes through the Transformer decoder to generate a segmentation mask through voxel-to-prototype comparison. A masked modality strategy is introduced to handle random modality absence during training. Furthermore, modality-view and class-view contrastive learning strategies are developed to enhance prototype learning. We conduct experiments on BraTS2020 and BraTS2018; the experimental results demonstrate the superior performance of our model under various missing modality scenarios. On BraTS2020, our model achieves DSC improvements of 5.9% for ET, 0.5% for TC, and 0.2% for WT compared to state-of-the-art methods. Notably, in the challenging T1C modality missing scenario, our model achieved clinically significant gains of 9.5% for ET and 1.8% for TC. The code is available at https://github.com/Xiheran/MCPL.

多模态融合是脑肿瘤整体诊断的有效解决方案;然而,它面临着缺失模式的挑战。传统的多编码器体系结构可以很容易地捕获特定于模态的特征,而单编码器体系结构很容易获得模态共享的特征。然而,反之则具有挑战性。在本文中,我们提出了一个两阶段双视图原型学习框架,以同时提取特定模态特征和特定类特征。在第一阶段,我们利用Transformer解码器来学习用于优化模态重建任务的模态原型。引入掩码自编码器生成不完全模态的共享特征。在重构过程中,学习到的包含模式特定特征的模式原型与模式共享特征混合在一起。在第二阶段,我们通过Transformer解码器学习类原型,通过体素与原型的比较生成分割掩码。在训练过程中引入了一种掩模态策略来处理随机模态缺失。此外,本文还提出了模态观和类观对比学习策略,以促进原型学习。我们在BraTS2020和BraTS2018上进行了实验;实验结果表明,该模型在各种缺失模态场景下都具有良好的性能。在BraTS2020上,与最先进的方法相比,我们的模型对ET的DSC提高了5.9%,对TC提高了0.5%,对WT提高了0.2%。值得注意的是,在具有挑战性的T1C模式缺失情况下,我们的模型在ET和TC方面取得了9.5%和1.8%的临床显著收益。代码可在https://github.com/Xiheran/MCPL上获得。
{"title":"Multimodal Contrastive Prototype Learning for Resilient Brain Tumor Segmentation With Missing Modalities.","authors":"Heran Xi, Yu Ye, Jinghua Zhu, Jinbao Li","doi":"10.1109/JBHI.2025.3649819","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649819","url":null,"abstract":"<p><p>Multimodal fusion is an effective solution for holistic brain tumor diagnosis; however, it faces challenges under missing modalities. Traditional multi-encoder architectures can easily capture modality-specific features, while single-encoder architectures readily obtain modality-shared features. The reverse, however, is challenging. In this paper, we propose a two-stage dual-view prototype learning framework to extract the modality-specific feature and the class-specific feature simultaneously. In the first stage, we utilize the Transformer decoder to learn the modality-prototypes that are used to optimize the modality reconstruction task. A masked autoencoder is introduced to generate shared features of incomplete modalities. The learned modality-prototypes that contain modality-specific features are blended with the modality-share features for the reconstruction process. In the second stage, we learn the class-prototypes through the Transformer decoder to generate a segmentation mask through voxel-to-prototype comparison. A masked modality strategy is introduced to handle random modality absence during training. Furthermore, modality-view and class-view contrastive learning strategies are developed to enhance prototype learning. We conduct experiments on BraTS2020 and BraTS2018; the experimental results demonstrate the superior performance of our model under various missing modality scenarios. On BraTS2020, our model achieves DSC improvements of 5.9% for ET, 0.5% for TC, and 0.2% for WT compared to state-of-the-art methods. Notably, in the challenging T1C modality missing scenario, our model achieved clinically significant gains of 9.5% for ET and 1.8% for TC. The code is available at https://github.com/Xiheran/MCPL.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145878171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure-Aware Consensus Representation Learning with Dual-Channel Attention for Multi-Omics Cancer Subtype Clustering. 基于双通道关注的结构感知共识表示学习用于多组癌亚型聚类。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-30 DOI: 10.1109/JBHI.2025.3649227
Yong Zhang, Kun Liu, Wenzhe Liu, Jiongcheng Zhu, Jianfeng Zhong

Cancer is characterized by complex subtypes and pronounced heterogeneity, which pose significant challenges for accurate identification and effective treatment. In response, multi-omics clustering has emerged as a powerful approach for integrating heterogeneous biological data to identify cancer subtypes, thereby playing a crucial role in early diagnosis and precision medicine. Despite promising progress, existing multi-omics clustering methods face two key limitations. First, most methods focus on mining the common information across omics but neglect the unique heterogeneity features of each omics. Second, representation learning and clustering are often decoupled, preventing joint optimization of feature representations and the clustering affinity matrix, ultimately leading to suboptimal performance. To tackle these difficulties, we propose a novel Structure-Aware Consensus Representation Learning with Dual-Channel Attention for Multi-Omics Cancer Subtype Clustering(SACR-DCA). SACR-DCA integrates two pivotal modules: (1) The multi-omics specific feature extraction and common representation fusion module, which uniquely captures both omics-specific characteristics and their shared information via a dual-channel attention fusion framework; (2) The clustering-oriented structure-aware representation learning and consensus enhancement module, which enhances consensus representations through structure-aware learning to boost clustering efficacy, leveraging a Cauchy-Schwarz (CS) divergence constraint for clustering adaptability. Performance experiments on ten real-world datasets fully demonstrate that our method outperforms existing methods. The source code is available at https://github.com/liukun2000/SACR-DCA.

癌症具有复杂的亚型和明显的异质性,这对准确识别和有效治疗构成了重大挑战。因此,多组学聚类已成为一种整合异质生物学数据以识别癌症亚型的强大方法,从而在早期诊断和精准医疗中发挥重要作用。尽管有很大的进展,现有的多组学聚类方法面临两个关键的局限性。首先,大多数方法侧重于挖掘组学之间的共同信息,而忽略了每个组学独特的异质性特征。其次,表示学习和聚类通常是解耦的,这阻碍了特征表示和聚类亲和矩阵的联合优化,最终导致次优性能。为了解决这些困难,我们提出了一种新的具有双通道关注的结构感知共识表示学习,用于多组学癌症亚型聚类(SACR-DCA)。SACR-DCA集成了两个关键模块:(1)多组学特异性特征提取和共同表征融合模块,该模块通过双通道注意力融合框架独特地捕获组学特异性特征及其共享信息;(2)面向聚类的结构感知表征学习和共识增强模块,通过结构感知学习增强共识表征,提高聚类效率,利用Cauchy-Schwarz (CS)发散约束提高聚类适应性。在10个真实数据集上的性能实验充分证明了我们的方法优于现有的方法。源代码可从https://github.com/liukun2000/SACR-DCA获得。
{"title":"Structure-Aware Consensus Representation Learning with Dual-Channel Attention for Multi-Omics Cancer Subtype Clustering.","authors":"Yong Zhang, Kun Liu, Wenzhe Liu, Jiongcheng Zhu, Jianfeng Zhong","doi":"10.1109/JBHI.2025.3649227","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649227","url":null,"abstract":"<p><p>Cancer is characterized by complex subtypes and pronounced heterogeneity, which pose significant challenges for accurate identification and effective treatment. In response, multi-omics clustering has emerged as a powerful approach for integrating heterogeneous biological data to identify cancer subtypes, thereby playing a crucial role in early diagnosis and precision medicine. Despite promising progress, existing multi-omics clustering methods face two key limitations. First, most methods focus on mining the common information across omics but neglect the unique heterogeneity features of each omics. Second, representation learning and clustering are often decoupled, preventing joint optimization of feature representations and the clustering affinity matrix, ultimately leading to suboptimal performance. To tackle these difficulties, we propose a novel Structure-Aware Consensus Representation Learning with Dual-Channel Attention for Multi-Omics Cancer Subtype Clustering(SACR-DCA). SACR-DCA integrates two pivotal modules: (1) The multi-omics specific feature extraction and common representation fusion module, which uniquely captures both omics-specific characteristics and their shared information via a dual-channel attention fusion framework; (2) The clustering-oriented structure-aware representation learning and consensus enhancement module, which enhances consensus representations through structure-aware learning to boost clustering efficacy, leveraging a Cauchy-Schwarz (CS) divergence constraint for clustering adaptability. Performance experiments on ten real-world datasets fully demonstrate that our method outperforms existing methods. The source code is available at https://github.com/liukun2000/SACR-DCA.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145862465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Prediction of circRNA-Drug Associations Based on Bipartite Graph Transformer. 基于二部图转换器的circrna -药物关联预测。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-30 DOI: 10.1109/JBHI.2025.3649178
Zihan Zhang, Yuchen Zhang, Xiujuan Lei

Circular RNAs (circRNAs) represent a distinctive class of non-coding RNAs with covalently closed loop structures that play crucial regulatory roles in drug response. While existing computational methods have achieved certain progress in prediction tasks, they primarily relied on circRNA genotypes and traditional molecular fingerprints, with limited utilization of multi-omics data and inadequate consideration of heterogeneous network topology. To address these limitations, this study proposed the CircRNA-Drug Bipartite Graph Transformer (CDBGT) framework to predict associations. Rather than limiting to associations between circRNA genotypes and drugs, this study integrated circRNA-drug response and target association information from multiple databases. CDBGT employed pre-trained models RNA-FM and ChemBERTa to extract features of sequence and molecular fingerprint and utilized multi-omics data to construct similarity matrices. The framework incorporated a bipartite graph transformer with topological positional encoding, comprehensively considering degree encoding, degree ranking encoding and spectral encoding to extract topological information from heterogeneous networks. Experimental results showed that CDBGT performed stably in 5-fold cross-validation. On the Response dataset, it achieved ROC-AUC of 0.9674 and PR-AUC of 0.9540, while on the Target dataset it reached ROC-AUC of 0.8621. Compared with existing methods, it showed an improvement of 3.20 to 26.87 percentage points in ROC-AUC. Ablation experiments demonstrated the necessity of each module. Through literature-supported case studies, this work suggested potential directions for circRNA-based therapeutic research.

环状rna (circRNAs)是一类独特的非编码rna,具有共价闭环结构,在药物反应中发挥重要的调节作用。虽然现有的计算方法在预测任务上取得了一定的进展,但它们主要依赖于circRNA基因型和传统的分子指纹,对多组学数据的利用有限,对异构网络拓扑的考虑不足。为了解决这些局限性,本研究提出了CircRNA-Drug Bipartite Graph Transformer (CDBGT)框架来预测关联。该研究没有局限于circRNA基因型与药物之间的关联,而是整合了来自多个数据库的circRNA-药物反应和靶标关联信息。CDBGT采用预训练模型RNA-FM和ChemBERTa提取序列特征和分子指纹,利用多组学数据构建相似矩阵。该框架将二部图变换与拓扑位置编码相结合,综合考虑度编码、度排序编码和谱编码,从异构网络中提取拓扑信息。实验结果表明,CDBGT在5次交叉验证中表现稳定。在Response数据集上,ROC-AUC为0.9674,PR-AUC为0.9540,在Target数据集上,ROC-AUC为0.8621。与现有方法相比,ROC-AUC提高3.20 ~ 26.87个百分点。烧蚀实验证明了各模块的必要性。通过文献支持的案例研究,这项工作为基于环状rna的治疗研究提出了潜在的方向。
{"title":"Prediction of circRNA-Drug Associations Based on Bipartite Graph Transformer.","authors":"Zihan Zhang, Yuchen Zhang, Xiujuan Lei","doi":"10.1109/JBHI.2025.3649178","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649178","url":null,"abstract":"<p><p>Circular RNAs (circRNAs) represent a distinctive class of non-coding RNAs with covalently closed loop structures that play crucial regulatory roles in drug response. While existing computational methods have achieved certain progress in prediction tasks, they primarily relied on circRNA genotypes and traditional molecular fingerprints, with limited utilization of multi-omics data and inadequate consideration of heterogeneous network topology. To address these limitations, this study proposed the CircRNA-Drug Bipartite Graph Transformer (CDBGT) framework to predict associations. Rather than limiting to associations between circRNA genotypes and drugs, this study integrated circRNA-drug response and target association information from multiple databases. CDBGT employed pre-trained models RNA-FM and ChemBERTa to extract features of sequence and molecular fingerprint and utilized multi-omics data to construct similarity matrices. The framework incorporated a bipartite graph transformer with topological positional encoding, comprehensively considering degree encoding, degree ranking encoding and spectral encoding to extract topological information from heterogeneous networks. Experimental results showed that CDBGT performed stably in 5-fold cross-validation. On the Response dataset, it achieved ROC-AUC of 0.9674 and PR-AUC of 0.9540, while on the Target dataset it reached ROC-AUC of 0.8621. Compared with existing methods, it showed an improvement of 3.20 to 26.87 percentage points in ROC-AUC. Ablation experiments demonstrated the necessity of each module. Through literature-supported case studies, this work suggested potential directions for circRNA-based therapeutic research.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145862281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dual Domain Collaborative Network for Polyp Segmentation. 一种用于息肉分割的双域协同网络。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-30 DOI: 10.1109/JBHI.2025.3649384
Yao Tong, Zuojian Zhou, Kongfa Hu, Tao Yang, Andre Kaup, Xin Li

Accurate polyp segmentation in colonoscopy images is essential for early colorectal cancer detection but remains a challenging problem due to the limitations in existing methods for optimizing boundary features and aligning cross-level representations. Specifically, the indistinct polyp boundaries and scale variations across different feature levels pose significant challenges for segmentation accuracy. To address these issues, we propose a dual domain collaborative network (DDCNet) that introduces two novel modules: a frequency context enhancement module (FCEM), which operates in the frequency domain to refine high- and low-frequency features, and a cross-level shift recalibrated fusion module (CSFM), which improves multi scale feature alignment in the spatial domain. The FCEM improves boundary precision by adaptively refining high frequency boundary features and enhancing low-frequency contextual information, while the CSFM mitigates cross level feature misalignment by dynamically recalibrating multi-scale features throughout the encoder-decoder architecture. Additionally, we design a hybrid loss function that integrates boundary, cross-entropy, and frequency consistency losses to further boost segmentation performance. Experimental results on three benchmark datasets (Kvasir SEG, CVC-ClinicDB, and CVC-ColonDB) demonstrate that DDCNet achieves state-of-the-art performance, with Dice coefficients of 0.9343, 0.9447, and 0.8155, respectively. These results represent improvements of 1.0%-1.5% over the best existing methods. Ablation studies further validate the individual contributions of FCEM, CSFM, and the hybrid loss function. Additionally, we compared the proposed loss function with three commonly used functions.

结肠镜图像中准确的息肉分割对于早期结直肠癌检测至关重要,但由于现有方法在优化边界特征和对齐跨水平表示方面的局限性,仍然是一个具有挑战性的问题。具体而言,息肉边界模糊和不同特征水平的尺度变化对分割精度提出了重大挑战。为了解决这些问题,我们提出了一个双域协作网络(DDCNet),该网络引入了两个新模块:频率上下文增强模块(FCEM),它在频域中工作以细化高低频特征,以及跨电平移位再校准融合模块(CSFM),它改善了空间域的多尺度特征对齐。FCEM通过自适应细化高频边界特征和增强低频上下文信息来提高边界精度,而CSFM通过在整个编码器-解码器架构中动态重新校准多尺度特征来减轻跨电平特征错位。此外,我们设计了一个混合损失函数,集成了边界、交叉熵和频率一致性损失,以进一步提高分割性能。在Kvasir SEG、CVC-ClinicDB和CVC-ColonDB三个基准数据集上的实验结果表明,DDCNet达到了最先进的性能,Dice系数分别为0.9343、0.9447和0.8155。这些结果比现有的最佳方法提高了1.0% ~ 1.5%。消融研究进一步验证了FCEM、CSFM和混合损失函数的单独贡献。此外,我们将所提出的损失函数与三种常用函数进行了比较。
{"title":"A Dual Domain Collaborative Network for Polyp Segmentation.","authors":"Yao Tong, Zuojian Zhou, Kongfa Hu, Tao Yang, Andre Kaup, Xin Li","doi":"10.1109/JBHI.2025.3649384","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3649384","url":null,"abstract":"<p><p>Accurate polyp segmentation in colonoscopy images is essential for early colorectal cancer detection but remains a challenging problem due to the limitations in existing methods for optimizing boundary features and aligning cross-level representations. Specifically, the indistinct polyp boundaries and scale variations across different feature levels pose significant challenges for segmentation accuracy. To address these issues, we propose a dual domain collaborative network (DDCNet) that introduces two novel modules: a frequency context enhancement module (FCEM), which operates in the frequency domain to refine high- and low-frequency features, and a cross-level shift recalibrated fusion module (CSFM), which improves multi scale feature alignment in the spatial domain. The FCEM improves boundary precision by adaptively refining high frequency boundary features and enhancing low-frequency contextual information, while the CSFM mitigates cross level feature misalignment by dynamically recalibrating multi-scale features throughout the encoder-decoder architecture. Additionally, we design a hybrid loss function that integrates boundary, cross-entropy, and frequency consistency losses to further boost segmentation performance. Experimental results on three benchmark datasets (Kvasir SEG, CVC-ClinicDB, and CVC-ColonDB) demonstrate that DDCNet achieves state-of-the-art performance, with Dice coefficients of 0.9343, 0.9447, and 0.8155, respectively. These results represent improvements of 1.0%-1.5% over the best existing methods. Ablation studies further validate the individual contributions of FCEM, CSFM, and the hybrid loss function. Additionally, we compared the proposed loss function with three commonly used functions.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145862192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MEDL-DDI: Example-Driven Learning With Multi-Source Features for Predicting Drug-Drug Interaction. MEDL-DDI:用多源特征预测药物-药物相互作用的例子驱动学习。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2025-12-26 DOI: 10.1109/JBHI.2025.3648723
Haixue Zhao, Yunjiong Liu, Peiliang Zhang, Xiaoping Min, Ming Wang, Chao Che

Accurate drug-drug interaction (DDI) prediction is crucial for optimizing the efficacy of combination therapies and minimizing adverse effects. Most existing methods rely on single features and struggle to integrate structural and sequential drug information. Additionally, prediction bias caused by class imbalance remains a significant challenge. To address these issues, this study proposes a multi-source example-driven learning framework for DDI (MEDL-DDI) that jointly models structural and sequential drug representations to achieve robust multimodal fusion and mitigate class imbalance. MEDL-DDI enriches SMILES with chemical knowledge, extracts global semantic features via a Transformer, and identifies key substructures through a graph information bottleneck. Moreover, an example-driven mechanism guided by example centers enhances the model's ability to recognize minority classes. Experimental results on three benchmark datasets validate that MEDL-DDI outperforms state-of-the-art methods. The case study on cardiovascular drug interactions further highlights MEDL-DDI's practical value and applicability.

准确预测药物-药物相互作用(DDI)对于优化联合治疗的疗效和减少不良反应至关重要。大多数现有方法依赖于单一特征,难以整合结构和顺序药物信息。此外,由阶级不平衡引起的预测偏差仍然是一个重大挑战。为了解决这些问题,本研究提出了一个多源示例驱动的DDI学习框架(MEDL-DDI),该框架联合建模结构和顺序药物表示,以实现鲁棒的多模态融合并减轻类不平衡。MEDL-DDI通过化学知识丰富SMILES,通过Transformer提取全局语义特征,并通过图信息瓶颈识别关键子结构。此外,由示例中心引导的示例驱动机制增强了模型识别少数类的能力。在三个基准数据集上的实验结果验证了MEDL-DDI优于最先进的方法。心血管药物相互作用的案例研究进一步凸显了MEDL-DDI的实用价值和适用性。
{"title":"MEDL-DDI: Example-Driven Learning With Multi-Source Features for Predicting Drug-Drug Interaction.","authors":"Haixue Zhao, Yunjiong Liu, Peiliang Zhang, Xiaoping Min, Ming Wang, Chao Che","doi":"10.1109/JBHI.2025.3648723","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3648723","url":null,"abstract":"<p><p>Accurate drug-drug interaction (DDI) prediction is crucial for optimizing the efficacy of combination therapies and minimizing adverse effects. Most existing methods rely on single features and struggle to integrate structural and sequential drug information. Additionally, prediction bias caused by class imbalance remains a significant challenge. To address these issues, this study proposes a multi-source example-driven learning framework for DDI (MEDL-DDI) that jointly models structural and sequential drug representations to achieve robust multimodal fusion and mitigate class imbalance. MEDL-DDI enriches SMILES with chemical knowledge, extracts global semantic features via a Transformer, and identifies key substructures through a graph information bottleneck. Moreover, an example-driven mechanism guided by example centers enhances the model's ability to recognize minority classes. Experimental results on three benchmark datasets validate that MEDL-DDI outperforms state-of-the-art methods. The case study on cardiovascular drug interactions further highlights MEDL-DDI's practical value and applicability.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.8,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145843742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Biomedical and Health Informatics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1