首页 > 最新文献

IEEE Journal of Biomedical and Health Informatics最新文献

英文 中文
Two-Steps Neural Networks for an Automated Cerebrovascular Landmark Detection Along the Circle of Willis. 两步神经网络自动检测沿威利斯圈的脑血管地标。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3606992
Rafic Nader, Vincent L'Allinec, Romain Bourcier, Florent Autrusseau

Intracranial aneurysms (ICA) commonly occur in specific segments of the Circle of Willis (CoW), primarily, onto thirteen major arterial bifurcations. An accurate detection of these critical landmarks is necessary for a prompt and efficient diagnosis. We introduce a fully automated landmark detection approach for CoW bifurcations using a two-step neural networks process. Initially, an object detection network identifies regions of interest (ROIs) proximal to the landmark locations. Subsequently, a modified U-Net with deep supervision is exploited to accurately locate the bifurcations. This two-step method reduces various problems, such as the missed detections caused by two landmarks being close to each other and having similar visual characteristics, especially when processing the complete MRA Time-of-Flight (TOF). Additionally, it accounts for the anatomical variability of the CoW, which affects the number of detectable landmarks per scan. We assessed the effectiveness of our approach using two cerebral MRA datasets: our In-House dataset which had varying numbers of landmarks, and a public dataset with standardized landmark configuration. Our experimental results demonstrate that our method achieves the highest level of performance on a bifurcation detection task.

颅内动脉瘤(ICA)通常发生在威利斯圈(CoW)的特定段,主要发生在13个主要动脉分叉上。准确检测这些关键标志对于及时有效的诊断是必要的。我们介绍了一种全自动地标检测方法,用于CoW分岔使用两步神经网络过程。首先,目标检测网络识别靠近地标位置的感兴趣区域(roi)。随后,利用改进的U-Net进行深度监督,精确定位分叉点。这种两步法在处理完整的MRA飞行时间(TOF)时,减少了各种问题,例如由于两个地标彼此靠近且具有相似的视觉特征而导致的漏检问题。此外,它解释了CoW的解剖学变异性,这影响了每次扫描可检测到的标志的数量。我们使用两个大脑MRA数据集评估了我们方法的有效性:我们的内部数据集具有不同数量的地标,以及具有标准化地标配置的公共数据集。实验结果表明,我们的方法在分支检测任务上达到了最高的性能水平。
{"title":"Two-Steps Neural Networks for an Automated Cerebrovascular Landmark Detection Along the Circle of Willis.","authors":"Rafic Nader, Vincent L'Allinec, Romain Bourcier, Florent Autrusseau","doi":"10.1109/JBHI.2025.3606992","DOIUrl":"10.1109/JBHI.2025.3606992","url":null,"abstract":"<p><p>Intracranial aneurysms (ICA) commonly occur in specific segments of the Circle of Willis (CoW), primarily, onto thirteen major arterial bifurcations. An accurate detection of these critical landmarks is necessary for a prompt and efficient diagnosis. We introduce a fully automated landmark detection approach for CoW bifurcations using a two-step neural networks process. Initially, an object detection network identifies regions of interest (ROIs) proximal to the landmark locations. Subsequently, a modified U-Net with deep supervision is exploited to accurately locate the bifurcations. This two-step method reduces various problems, such as the missed detections caused by two landmarks being close to each other and having similar visual characteristics, especially when processing the complete MRA Time-of-Flight (TOF). Additionally, it accounts for the anatomical variability of the CoW, which affects the number of detectable landmarks per scan. We assessed the effectiveness of our approach using two cerebral MRA datasets: our In-House dataset which had varying numbers of landmarks, and a public dataset with standardized landmark configuration. Our experimental results demonstrate that our method achieves the highest level of performance on a bifurcation detection task.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1353-1364"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145023207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Medical Image Privacy in Federated Learning: Segmentation-Reorganization and Sparsified Gradient Matching Attacks. 联邦学习中的医学图像隐私:分割重组和稀疏梯度匹配攻击。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3593631
Kaimin Wei, Jin Qian, Chengkun Jia, Jinpeng Chen, Jilian Zhang, Yongdong Wu, Jinyu Zhu, Yuhan Guo

In modern medicine, the widespread use of medical imaging has greatly improved diagnostic and treatment efficiency. However, these images contain sensitive personal information, and any leakage could seriously compromise patient privacy, leading to ethical and legal issues. Federated learning (FL), an emerging privacy-preserving technique, transmits gradients rather than raw data for model training. Yet, recent studies reveal that gradient inversion attacks can exploit this information to reconstruct private data, posing a significant threat to FL. Current attacks remain limited in image resolution, similarity, and batch processing, and thus do not yet pose a significant risk to FL. To address this, we propose a novel gradient inversion attack based on sparsified gradient matching and segmentation reorganization (SR) to reconstruct high-resolution, high-similarity medical images in batch mode. Specifically, an $L_{1}$ loss function optimises the gradient sparsification process, while the SR strategy enhances image resolution. An adaptive learning rate adjustment mechanism is also employed to improve optimisation stability and avoid local optima. Experimental results demonstrate that our method significantly outperforms state-of-the-art approaches in both visual quality and quantitative metrics, achieving up to a 146% improvement in similarity.

在现代医学中,医学影像学的广泛应用大大提高了诊断和治疗效率。然而,这些图像包含敏感的个人信息,任何泄露都可能严重损害患者的隐私,导致道德和法律问题。联邦学习(FL)是一种新兴的隐私保护技术,它传输梯度而不是原始数据用于模型训练。然而,最近的研究表明,梯度反演攻击可以利用这些信息来重建私人数据,对FL构成重大威胁。目前的攻击在图像分辨率、相似性和批处理方面仍然有限,因此尚未对FL构成重大风险。为了解决这个问题,我们提出了一种基于稀疏梯度匹配和分割重组(SR)的新型梯度反演攻击来重建高分辨率。批处理模式下的高相似度医学图像。具体来说,$L_{1}$损失函数优化了梯度稀疏化过程,而SR策略增强了图像分辨率。采用自适应学习率调整机制,提高优化稳定性,避免局部最优。实验结果表明,我们的方法在视觉质量和定量指标上都明显优于最先进的方法,相似度提高了146%。
{"title":"Medical Image Privacy in Federated Learning: Segmentation-Reorganization and Sparsified Gradient Matching Attacks.","authors":"Kaimin Wei, Jin Qian, Chengkun Jia, Jinpeng Chen, Jilian Zhang, Yongdong Wu, Jinyu Zhu, Yuhan Guo","doi":"10.1109/JBHI.2025.3593631","DOIUrl":"10.1109/JBHI.2025.3593631","url":null,"abstract":"<p><p>In modern medicine, the widespread use of medical imaging has greatly improved diagnostic and treatment efficiency. However, these images contain sensitive personal information, and any leakage could seriously compromise patient privacy, leading to ethical and legal issues. Federated learning (FL), an emerging privacy-preserving technique, transmits gradients rather than raw data for model training. Yet, recent studies reveal that gradient inversion attacks can exploit this information to reconstruct private data, posing a significant threat to FL. Current attacks remain limited in image resolution, similarity, and batch processing, and thus do not yet pose a significant risk to FL. To address this, we propose a novel gradient inversion attack based on sparsified gradient matching and segmentation reorganization (SR) to reconstruct high-resolution, high-similarity medical images in batch mode. Specifically, an $L_{1}$ loss function optimises the gradient sparsification process, while the SR strategy enhances image resolution. An adaptive learning rate adjustment mechanism is also employed to improve optimisation stability and avoid local optima. Experimental results demonstrate that our method significantly outperforms state-of-the-art approaches in both visual quality and quantitative metrics, achieving up to a 146% improvement in similarity.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1443-1451"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144952133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Cancer Research With Synthetic Data Generation in Low-Data Scenarios. 低数据场景下合成数据生成推进癌症研究。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3595371
Patricia A Apellaniz, Borja Arroyo Galende, Ana Jimenez, Juan Parras, Santiago Zazo

The scarcity of medical data, particularly in Survival Analysis (SA) for cancer-related diseases, challenges data-driven healthcare research. While Synthetic Tabular Data Generation (STDG) models have been proposed to address this issue, most rely on datasets with abundant samples, which do not reflect real-world limitations. We suggest using an STDG approach that leverages transfer learning and meta-learning techniques to create an artificial inductive bias, guiding generative models trained on limited samples. Experiments on classification datasets across varying sample sizes validated the method's robustness, with further clinical utility assessment on cancer-related SA data. While divergence-based similarity validation proved effective in capturing improvements in generation quality, clinical utility validation showed limited sensitivity to sample size, highlighting its shortcomings. In SA experiments, we observed that altering the task can reveal if relationships among variables are accurately generated, with most cases benefiting from the proposed methodology. Our findings confirm the method's ability to generate high-quality synthetic data under constrained conditions. We emphasize the need to complement utility-based validation with similarity metrics, particularly in low-data settings, to assess STDG performance reliably.

医疗数据的缺乏,特别是在癌症相关疾病的生存分析(SA)中,给数据驱动的医疗保健研究带来了挑战。虽然已经提出了合成表格数据生成(STDG)模型来解决这个问题,但大多数模型依赖于具有丰富样本的数据集,而不能反映现实世界的局限性。我们建议使用STDG方法,利用迁移学习和元学习技术来创建人工归纳偏差,指导在有限样本上训练的生成模型。最初的实验是在更大的分类数据集上进行的,这使我们能够在不同的样本量和丰富与稀缺的数据场景下评估方法。我们主要采用临床效用验证癌症相关SA数据,因为基于差异的相似性验证是不可行的。该方法在受限数据条件下改进了STDG,基于散度的相似性验证被证明是数据质量的稳健度量。相反,无论样本量大小,临床效用验证都得出了类似的结果,这表明其在统计确认有效STDG方面的局限性。在SA实验中,我们观察到,改变任务可以揭示变量之间的关系是否准确地生成,大多数情况下受益于所提出的方法。我们的研究强调了该方法通过在受限条件下有效生成高质量合成数据来解决医疗数据稀缺问题的有效性。当有足够的数据可用时,基于差异的相似性验证是必不可少的,但仅靠临床效用验证是不够的,应该辅以相似性验证。这些发现强调了STDG方法在解决医疗数据稀缺问题方面的潜力和局限性。
{"title":"Advancing Cancer Research With Synthetic Data Generation in Low-Data Scenarios.","authors":"Patricia A Apellaniz, Borja Arroyo Galende, Ana Jimenez, Juan Parras, Santiago Zazo","doi":"10.1109/JBHI.2025.3595371","DOIUrl":"10.1109/JBHI.2025.3595371","url":null,"abstract":"<p><p>The scarcity of medical data, particularly in Survival Analysis (SA) for cancer-related diseases, challenges data-driven healthcare research. While Synthetic Tabular Data Generation (STDG) models have been proposed to address this issue, most rely on datasets with abundant samples, which do not reflect real-world limitations. We suggest using an STDG approach that leverages transfer learning and meta-learning techniques to create an artificial inductive bias, guiding generative models trained on limited samples. Experiments on classification datasets across varying sample sizes validated the method's robustness, with further clinical utility assessment on cancer-related SA data. While divergence-based similarity validation proved effective in capturing improvements in generation quality, clinical utility validation showed limited sensitivity to sample size, highlighting its shortcomings. In SA experiments, we observed that altering the task can reveal if relationships among variables are accurately generated, with most cases benefiting from the proposed methodology. Our findings confirm the method's ability to generate high-quality synthetic data under constrained conditions. We emphasize the need to complement utility-based validation with similarity metrics, particularly in low-data settings, to assess STDG performance reliably.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1666-1679"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144784194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WOADNet: A Wavelet-Inspired Orientational Adaptive Dictionary Network for CT Metal Artifact Reduction. wadnet:一种基于小波的CT金属伪影还原定向自适应字典网络。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3592024
Tong Jin, Jin Liu, Diandian Wang, Kun Wang, Chenlong Miao, Yikun Zhang, Dianlin Hu, Zhan Wu, Yang Chen

In computed tomography (CT), metal artifacts pose a persistent challenge to achieving high-quality imaging. Despite advancements in metal artifact reduction (MAR) techniques, many existing approaches have not fully leveraged the intrinsic a priori knowledge related to metal artifacts, improved model interpretability, or addressed the complex texture of CT images effectively. To address these limitations, we propose a novel and interpretable framework, the wavelet-inspired oriented adaptive dictionary network (WOADNet). WOADNet builds on sparse coding with orientational information in the wavelet domain. By exploring the discriminative features of artifacts and anatomical tissues, we adopt a high-precision filter parameterization strategy that incorporates multiangle rotations. Furthermore, we integrate a reweighted sparse constraint framework into the convolutional dictionary learning process and employ a cross-space, multiscale attention mechanism to construct an adaptive convolutional dictionary unit for the artifact feature encoder. This innovative design allows for flexible adjustment of weights and convolutional representations, resulting in significant image quality improvements. The experimental results using synthetic and clinical datasets demonstrate that WOADNet outperforms both traditional and state-of-the-art MAR methods in terms of suppressing artifacts.

在计算机断层扫描(CT)中,金属伪影对实现高质量成像构成了持续的挑战。尽管金属伪影还原(MAR)技术取得了进步,但许多现有的方法并没有充分利用与金属伪影相关的固有先验知识,提高模型的可解释性,或有效地处理CT图像的复杂纹理。为了解决这些限制,我们提出了一种新颖的可解释框架,即小波启发的面向自适应字典网络(WOADNet)。WOADNet基于小波域的方向信息稀疏编码。通过探索伪影和解剖组织的区别特征,我们采用了一种包含多角度旋转的高精度滤波参数化策略。此外,我们将一个重加权的稀疏约束框架整合到卷积字典学习过程中,并采用跨空间、多尺度注意机制为伪特征编码器构建自适应卷积字典单元。这种创新的设计允许灵活地调整权重和卷积表示,从而显著提高图像质量。使用合成和临床数据集的实验结果表明,WOADNet在抑制伪像方面优于传统和最先进的MAR方法。
{"title":"WOADNet: A Wavelet-Inspired Orientational Adaptive Dictionary Network for CT Metal Artifact Reduction.","authors":"Tong Jin, Jin Liu, Diandian Wang, Kun Wang, Chenlong Miao, Yikun Zhang, Dianlin Hu, Zhan Wu, Yang Chen","doi":"10.1109/JBHI.2025.3592024","DOIUrl":"10.1109/JBHI.2025.3592024","url":null,"abstract":"<p><p>In computed tomography (CT), metal artifacts pose a persistent challenge to achieving high-quality imaging. Despite advancements in metal artifact reduction (MAR) techniques, many existing approaches have not fully leveraged the intrinsic a priori knowledge related to metal artifacts, improved model interpretability, or addressed the complex texture of CT images effectively. To address these limitations, we propose a novel and interpretable framework, the wavelet-inspired oriented adaptive dictionary network (WOADNet). WOADNet builds on sparse coding with orientational information in the wavelet domain. By exploring the discriminative features of artifacts and anatomical tissues, we adopt a high-precision filter parameterization strategy that incorporates multiangle rotations. Furthermore, we integrate a reweighted sparse constraint framework into the convolutional dictionary learning process and employ a cross-space, multiscale attention mechanism to construct an adaptive convolutional dictionary unit for the artifact feature encoder. This innovative design allows for flexible adjustment of weights and convolutional representations, resulting in significant image quality improvements. The experimental results using synthetic and clinical datasets demonstrate that WOADNet outperforms both traditional and state-of-the-art MAR methods in terms of suppressing artifacts.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1452-1465"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144698429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing the Reliability of Affective Brain-Computer Interfaces by Using Specifically Designed Confidence Estimator. 利用特殊设计的置信度估计提高情感脑机接口的可靠性。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3594219
Jiaheng Wang, Zhenyu Wang, Tianheng Xu, Ang Li, Yuan Si, Ting Zhou, Xi Zhao, Honglin Hu

In recent years, the diverse applications of electroencephalography (EEG) - based affective brain-computer interfaces (aBCIs) are being extensively explored. However, due to adverse factors like noise and physiological variability, the recognition capability of aBCIs can unforeseeably suffer abrupt declines. Since the timing of these aBCI failures is unknown, placing trust in aBCIs without scrutiny can lead to undesirable consequences. To alleviate this issue, we propose an algorithm for estimating the reliability of aBCI (primarily Graph Convolutional Network), synchronously delivering a probabilistic confidence score upon aBCI decision completion, thereby reflecting the aBCI's real-time recognition capabilities. Methodologically, we use the Maximum Softmax Probability (MSP) from EEG recognition networks as confidence scores and leverage the Scaling Operator to calibrate them. Then, the Projection Operator is employed to address confidence estimation biases caused by noise and subject variability. For the numerical concentration of MSP, we provide fresh insights into its causes and propose corresponding solutions. The derivation of the estimator from the Maximum Entropy Principle is also substantiated for robust theoretical underpinnings. Finally, we confirm theoretically that the estimator does not compromise BCI performance. In experiments conducted on public datasets SEED and SEED-IV, the proposed algorithm demonstrates superior performance in estimating aBCIs reliability compared to other benchmarks, and commendable adaptability to new subjects. This research has the potential to lead to more trustworthy aBCIs and advance their broader application in complex real-world scenarios.

近年来,基于脑电图(EEG)的情感脑机接口(abci)的各种应用正在被广泛探索。然而,由于噪声和生理变异等不利因素,abci的识别能力可能会不可预见地突然下降。由于这些aBCI故障的时间是未知的,在没有审查的情况下信任aBCI可能会导致不良后果。为了缓解这一问题,我们提出了一种估计aBCI(主要是图卷积网络)可靠性的算法,在aBCI决策完成时同步提供概率置信度评分,从而反映aBCI的实时识别能力。在方法上,我们使用来自脑电图识别网络的最大软最大概率(MSP)作为置信度分数,并利用缩放算子对它们进行校准。然后,利用投影算子解决噪声和主体变异性引起的置信度估计偏差。对于MSP的数值浓度,我们对其产生的原因有了新的认识,并提出了相应的解决方案。从最大熵原理推导的估计量也证实了稳健的理论基础。最后,我们从理论上证实了该估计器不会损害BCI性能。在公共数据集SEED和SEED- iv上进行的实验中,与其他基准测试相比,该算法在估计abci可靠性方面表现出优异的性能,并且对新主题具有良好的适应性。这项研究有可能导致更值得信赖的abci,并在复杂的现实世界场景中推进其更广泛的应用。
{"title":"Enhancing the Reliability of Affective Brain-Computer Interfaces by Using Specifically Designed Confidence Estimator.","authors":"Jiaheng Wang, Zhenyu Wang, Tianheng Xu, Ang Li, Yuan Si, Ting Zhou, Xi Zhao, Honglin Hu","doi":"10.1109/JBHI.2025.3594219","DOIUrl":"10.1109/JBHI.2025.3594219","url":null,"abstract":"<p><p>In recent years, the diverse applications of electroencephalography (EEG) - based affective brain-computer interfaces (aBCIs) are being extensively explored. However, due to adverse factors like noise and physiological variability, the recognition capability of aBCIs can unforeseeably suffer abrupt declines. Since the timing of these aBCI failures is unknown, placing trust in aBCIs without scrutiny can lead to undesirable consequences. To alleviate this issue, we propose an algorithm for estimating the reliability of aBCI (primarily Graph Convolutional Network), synchronously delivering a probabilistic confidence score upon aBCI decision completion, thereby reflecting the aBCI's real-time recognition capabilities. Methodologically, we use the Maximum Softmax Probability (MSP) from EEG recognition networks as confidence scores and leverage the Scaling Operator to calibrate them. Then, the Projection Operator is employed to address confidence estimation biases caused by noise and subject variability. For the numerical concentration of MSP, we provide fresh insights into its causes and propose corresponding solutions. The derivation of the estimator from the Maximum Entropy Principle is also substantiated for robust theoretical underpinnings. Finally, we confirm theoretically that the estimator does not compromise BCI performance. In experiments conducted on public datasets SEED and SEED-IV, the proposed algorithm demonstrates superior performance in estimating aBCIs reliability compared to other benchmarks, and commendable adaptability to new subjects. This research has the potential to lead to more trustworthy aBCIs and advance their broader application in complex real-world scenarios.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1073-1086"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144764849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SaccpaNet: A Separable Atrous Convolution- Based Cascade Pyramid Attention Network to Estimate Body Landmarks Using Cross-Modal Knowledge Transfer for Under-Blanket Sleep Posture Classification. SaccpaNet:基于可分离无齿卷积的级联金字塔注意网络,利用跨模态知识转移估算身体地标,用于毯下睡姿分类。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2024.3432195
Andy Yiu-Chau Tam, Ye-Jiao Mao, Derek Ka-Hei Lai, Andy Chi-Ho Chan, Daphne Sze Ki Cheung, William Kearns, Duo Wai-Chi Wong, James Chung-Wai Cheung

The accuracy of sleep posture assessment in standard polysomnography might be compromised by the unfamiliar sleep lab environment. In this work, we aim to develop a depth camera-based sleep posture monitoring and classification system for home or community usage and tailor a deep learning model that can account for blanket interference. Our model included a joint coordinate estimation network (JCE) and sleep posture classification network (SPC). SaccpaNet (Separable Atrous Convolution-based Cascade Pyramid Attention Network) was developed using a combination of pyramidal structure of residual separable atrous convolution unit to reduce computational cost and enlarge receptive field. The Saccpa attention unit served as the core of JCE and SPC, while different backbones for SPC were also evaluated. The model was cross-modally pretrained by RGB images from the COCO whole body dataset and then trained/tested using dept image data collected from 150 participants performing seven sleep postures across four blanket conditions. Besides, we applied a data augmentation technique that used intra-class mix-up to synthesize blanket conditions; and an overlaid flip-cut to synthesize partially covered blanket conditions for a robustness that we referred to as the Post-hoc Data Augmentation Robustness Test (PhD-ART). Our model achieved an average precision of estimated joint coordinate (in terms of PCK@0.1) of 0.652 and demonstrated adequate robustness. The overall classification accuracy of sleep postures (F1-score) was 0.885 and 0.940, for 7- and 6-class classification, respectively. Our system was resistant to the interference of blanket, with a spread difference of 2.5%.

标准多导睡眠监测仪对睡眠姿势评估的准确性可能会受到陌生的睡眠实验室环境的影响。在这项工作中,我们旨在开发一种基于深度摄像头的睡眠姿势监测和分类系统,供家庭或社区使用,并定制一种可考虑毯子干扰的深度学习模型。我们的模型包括联合坐标估计网络(JCE)和睡姿分类网络(SPC)。SaccpaNet(基于可分离无齿卷积的级联金字塔注意网络)是利用残余可分离无齿卷积单元的金字塔结构组合开发的,以降低计算成本并扩大感受野。Saccpa 注意单元是 JCE 和 SPC 的核心,同时还对 SPC 的不同骨架进行了评估。该模型通过 COCO 全身数据集的 RGB 图像进行跨模态预训练,然后使用从 150 名参与者在四种毯子条件下的七种睡眠姿势中收集的深度图像数据进行训练/测试。此外,我们还应用了一种数据增强技术,即使用类内混合来合成毯子条件;以及一种覆盖翻转切割来合成部分覆盖的毯子条件,以实现我们称之为 "事后数据增强鲁棒性测试"(PhD-ART)的鲁棒性。我们的模型估计关节坐标的平均精度(以 PCK@0.1 计)达到了 0.652,表现出了足够的鲁棒性。睡眠姿势的总体分类准确率(F1-分数)分别为 0.885 和 0.940(7 级分类和 6 级分类)。我们的系统对毯子的干扰具有很强的抵抗力,传播差为 2.5%。
{"title":"SaccpaNet: A Separable Atrous Convolution- Based Cascade Pyramid Attention Network to Estimate Body Landmarks Using Cross-Modal Knowledge Transfer for Under-Blanket Sleep Posture Classification.","authors":"Andy Yiu-Chau Tam, Ye-Jiao Mao, Derek Ka-Hei Lai, Andy Chi-Ho Chan, Daphne Sze Ki Cheung, William Kearns, Duo Wai-Chi Wong, James Chung-Wai Cheung","doi":"10.1109/JBHI.2024.3432195","DOIUrl":"10.1109/JBHI.2024.3432195","url":null,"abstract":"<p><p>The accuracy of sleep posture assessment in standard polysomnography might be compromised by the unfamiliar sleep lab environment. In this work, we aim to develop a depth camera-based sleep posture monitoring and classification system for home or community usage and tailor a deep learning model that can account for blanket interference. Our model included a joint coordinate estimation network (JCE) and sleep posture classification network (SPC). SaccpaNet (Separable Atrous Convolution-based Cascade Pyramid Attention Network) was developed using a combination of pyramidal structure of residual separable atrous convolution unit to reduce computational cost and enlarge receptive field. The Saccpa attention unit served as the core of JCE and SPC, while different backbones for SPC were also evaluated. The model was cross-modally pretrained by RGB images from the COCO whole body dataset and then trained/tested using dept image data collected from 150 participants performing seven sleep postures across four blanket conditions. Besides, we applied a data augmentation technique that used intra-class mix-up to synthesize blanket conditions; and an overlaid flip-cut to synthesize partially covered blanket conditions for a robustness that we referred to as the Post-hoc Data Augmentation Robustness Test (PhD-ART). Our model achieved an average precision of estimated joint coordinate (in terms of PCK@0.1) of 0.652 and demonstrated adequate robustness. The overall classification accuracy of sleep postures (F1-score) was 0.885 and 0.940, for 7- and 6-class classification, respectively. Our system was resistant to the interference of blanket, with a spread difference of 2.5%.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1593-1604"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141751582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Video Polyp Segmentation by Deformable Alignment and Local Attention. 基于可变形对齐和局部关注的高效视频息肉分割。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3592897
Yifei Zhao, Xiaoying Wang, Junping Yin

Accurate and efficient Video Polyp Segmentation (VPS) is vital for the early detection of colorectal cancer and the effectivetreatment of polyps. However, achieving this remains highly challenging due to the inherent difficulty in modeling the spatial-temporal relationships within colonoscopy videos. Existing methods that directly associate video frames frequently fail to account for variations in polyp or background motion, leading to excessive noise and reduced segmentation accuracy. Conversely, approaches that rely on optical flow models to estimate motion and align frames incur significant computational overhead. To address these limitations, we propose a novel VPS framework, termed Deformable Alignment and Local Attention (DALA). In this framework, we first construct a shared encoder to jointly encode the feature representations of paired video frames. Subsequently, we introduce a Multi-Scale Frame Alignment (MSFA) module based on deformable convolution to estimate the motion between reference and anchor frames. The multi-scale architecture is designed to accommodate the scale variations of polyps arising from differing viewing angles and speeds during colonoscopy. Furthermore, Local Attention (LA) is employed to selectively aggregate the aligned features, yielding more precise spatial-temporal feature representations. Extensive experiments conducted on the challenging SUN-SEG dataset and PolypGen dataset demonstrate that DALA achieves superior performance compared to state-of-the-art models.

准确、高效的视频息肉分割(VPS)对于早期发现结直肠癌和有效治疗息肉至关重要。然而,由于在结肠镜检查视频中建模时空关系的固有困难,实现这一目标仍然具有很高的挑战性。现有的直接关联视频帧的方法经常不能解释息肉或背景运动的变化,导致过多的噪声和降低分割精度。相反,依赖于光流模型来估计运动和对齐帧的方法会产生显著的计算开销。为了解决这些限制,我们提出了一个新的VPS框架,称为可变形对齐和局部注意(DALA)。在该框架中,我们首先构建一个共享编码器,对成对视频帧的特征表示进行联合编码。随后,我们引入了一种基于可变形卷积的多尺度帧对齐(MSFA)模块来估计参考帧和锚帧之间的运动。多尺度结构的设计是为了适应结肠镜检查过程中因不同视角和速度而产生的息肉的尺度变化。此外,采用局部注意(Local Attention, LA)对对齐的特征进行选择性聚合,得到更精确的时空特征表示。在具有挑战性的SUN-SEG数据集和PolypGen数据集上进行的大量实验表明,与最先进的模型相比,DALA实现了卓越的性能。代码将在https://github.com/xff12138/DALA上公开。
{"title":"Efficient Video Polyp Segmentation by Deformable Alignment and Local Attention.","authors":"Yifei Zhao, Xiaoying Wang, Junping Yin","doi":"10.1109/JBHI.2025.3592897","DOIUrl":"10.1109/JBHI.2025.3592897","url":null,"abstract":"<p><p>Accurate and efficient Video Polyp Segmentation (VPS) is vital for the early detection of colorectal cancer and the effectivetreatment of polyps. However, achieving this remains highly challenging due to the inherent difficulty in modeling the spatial-temporal relationships within colonoscopy videos. Existing methods that directly associate video frames frequently fail to account for variations in polyp or background motion, leading to excessive noise and reduced segmentation accuracy. Conversely, approaches that rely on optical flow models to estimate motion and align frames incur significant computational overhead. To address these limitations, we propose a novel VPS framework, termed Deformable Alignment and Local Attention (DALA). In this framework, we first construct a shared encoder to jointly encode the feature representations of paired video frames. Subsequently, we introduce a Multi-Scale Frame Alignment (MSFA) module based on deformable convolution to estimate the motion between reference and anchor frames. The multi-scale architecture is designed to accommodate the scale variations of polyps arising from differing viewing angles and speeds during colonoscopy. Furthermore, Local Attention (LA) is employed to selectively aggregate the aligned features, yielding more precise spatial-temporal feature representations. Extensive experiments conducted on the challenging SUN-SEG dataset and PolypGen dataset demonstrate that DALA achieves superior performance compared to state-of-the-art models.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1534-1543"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144715061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Edge-Aware Diffusion Segmentation Model With Hessian Priors for Automated Diaphragm Thickness Measurement in Ultrasound Imaging. 超声成像中自动测量隔膜厚度的Hessian先验边缘感知扩散分割模型。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3601567
Chen-Long Miao, Yikang He, Baike Shi, Zhongkai Bian, Wenxue Yu, Yang Chen, Guang-Quan Zhou

The thickness of the diaphragm serves as a crucial biometric indicator, particularly in assessing rehabilitation and respiratory dysfunction. However, measuring diaphragm thickness from ultrasound images mainly depends on manual delineation of the fascia, which is subjective, time-consuming, and sensitive to the inherent speckle noise. In this study, we introduce an edge-aware diffusion segmentation model (ESADiff), which incorporates prior structural knowledge of the fascia to improve the accuracy and reliability of diaphragm thickness measurements in ultrasound imaging. We first apply a diffusion model, guided by annotations, to learn the image features while preserving edge details through an iterative denoising process. Specifically, we design an anisotropic edge-sensitive annotation refinement module that corrects inaccurate labels by integrating Hessian geometric priors with a backtracking shortest-path connection algorithm, further enhancing model accuracy. Moreover, a curvature-aware deformable convolution and edge-prior ranking loss function are proposed to leverage the shape prior knowledge of the fascia, allowing the model to selectively focus on relevant linear structures while mitigating the influence of noise on feature extraction. We evaluated the proposed model on an in-house diaphragm ultrasound dataset, a public calf muscle dataset, and an internal tongue muscle dataset to demonstrate robust generalization. Extensive experimental results demonstrate that our method achieves finer fascia segmentation and significantly improves the accuracy of thickness measurements compared to other state-of-the-art techniques, highlighting its potential for clinical applications.

横膈膜的厚度是一个重要的生物特征指标,特别是在评估康复和呼吸功能障碍方面。然而,从超声图像中测量隔膜厚度主要依赖于手工描绘筋膜,这是主观的,耗时的,并且对固有的散斑噪声敏感。在这项研究中,我们引入了一种边缘感知扩散分割模型(ESADiff),该模型结合了先前的筋膜结构知识,以提高超声成像中膈膜厚度测量的准确性和可靠性。我们首先应用一个扩散模型,在注释的指导下,学习图像特征,同时通过迭代去噪过程保留边缘细节。具体而言,我们设计了一个各向异性边缘敏感标注细化模块,通过整合Hessian几何先验和回溯最短路径连接算法来纠正不准确的标签,进一步提高模型精度。此外,提出了曲率感知的可变形卷积和边缘先验排序损失函数,利用筋膜的形状先验知识,使模型能够选择性地关注相关线性结构,同时减轻噪声对特征提取的影响。我们在内部隔膜超声数据集、公共小腿肌肉数据集和内部舌肌肉数据集上评估了所提出的模型,以证明鲁棒泛化。大量的实验结果表明,与其他最先进的技术相比,我们的方法实现了更精细的筋膜分割,显著提高了厚度测量的准确性,突出了其临床应用的潜力。
{"title":"Edge-Aware Diffusion Segmentation Model With Hessian Priors for Automated Diaphragm Thickness Measurement in Ultrasound Imaging.","authors":"Chen-Long Miao, Yikang He, Baike Shi, Zhongkai Bian, Wenxue Yu, Yang Chen, Guang-Quan Zhou","doi":"10.1109/JBHI.2025.3601567","DOIUrl":"10.1109/JBHI.2025.3601567","url":null,"abstract":"<p><p>The thickness of the diaphragm serves as a crucial biometric indicator, particularly in assessing rehabilitation and respiratory dysfunction. However, measuring diaphragm thickness from ultrasound images mainly depends on manual delineation of the fascia, which is subjective, time-consuming, and sensitive to the inherent speckle noise. In this study, we introduce an edge-aware diffusion segmentation model (ESADiff), which incorporates prior structural knowledge of the fascia to improve the accuracy and reliability of diaphragm thickness measurements in ultrasound imaging. We first apply a diffusion model, guided by annotations, to learn the image features while preserving edge details through an iterative denoising process. Specifically, we design an anisotropic edge-sensitive annotation refinement module that corrects inaccurate labels by integrating Hessian geometric priors with a backtracking shortest-path connection algorithm, further enhancing model accuracy. Moreover, a curvature-aware deformable convolution and edge-prior ranking loss function are proposed to leverage the shape prior knowledge of the fascia, allowing the model to selectively focus on relevant linear structures while mitigating the influence of noise on feature extraction. We evaluated the proposed model on an in-house diaphragm ultrasound dataset, a public calf muscle dataset, and an internal tongue muscle dataset to demonstrate robust generalization. Extensive experimental results demonstrate that our method achieves finer fascia segmentation and significantly improves the accuracy of thickness measurements compared to other state-of-the-art techniques, highlighting its potential for clinical applications.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1544-1554"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144952132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Cognitive Load Estimation With Radio Frequency Sensing and Pupillometry in Complex Auditory Environments. 复杂听觉环境下基于射频传感和瞳孔测量的多模态认知负荷估计。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3634072
Usman Anwar, Adeel Hussain, Mandar Gogate, Kia Dashtipour, Tughrul Arslan, Amir Hussain, Peter Lomax

The detection of listening effort or cognitive load (CL) has been a major research challenge in recent years. Most conventional techniques utilise physiological or audio-visual sensors and are privacy-invasive and computationally complex. The challenges of synchronization, data alignment and accessibility limitations potentially increase the noise and error probability, compromising the accuracy of CL estimates. This innovative work presents a multi-modal, non-invasive and privacy-preserving approach that combines Radio Frequency (RF) and pupillometry sensing to address these challenges. Custom RF sensors are first designed and developed to capture blood flow changes in specific brain regions with high spatial resolution. Next, multi-modal fusion with pupillometry sensing is proposed and shown to offer a robust assessment of cognitive and listening effort through pupil size and pupil dilation. Our novel approach evaluates RF sensing to estimate CL from cerebral blood flow variations utilizing pupillometry as a baseline. A first-of-its-kind, multi-modal dataset is collected as a new benchmark resource in a controlled environment with participants to comprehend target speech with varying background noise levels. The framework is statistically evaluated using intraclass correlation for pupillometry data (average ICC> 0.95). The correlation between pupillometry and RF data is established through Pearson's correlation (average PCC> 0.79). Further, CL is classified into high and low categories based on RF data using K-means clustering. Future work involves integrating RF sensors with glasses to estimate listening effort for hearing-aid users and utilising RF measurements to optimize speech enhancement based on individual's listening effort and complexity of acoustic environment.

听力努力或认知负荷(CL)的检测是近年来研究的一个重大挑战。大多数传统技术使用生理或视听传感器,并且侵犯隐私且计算复杂。同步、数据对齐和可访问性限制的挑战可能会增加噪声和错误概率,从而影响CL估计的准确性。这项创新工作提出了一种多模式、非侵入性和保护隐私的方法,该方法结合了射频(RF)和瞳孔测量传感来应对这些挑战。定制的射频传感器首先被设计和开发,以高空间分辨率捕获特定大脑区域的血流变化。接下来,提出了瞳孔测量传感的多模态融合,并通过瞳孔大小和瞳孔扩张提供了对认知和听力努力的可靠评估。我们的新方法评估射频传感,以估计CL从脑血流变化利用瞳孔测量作为基线。首先,在受控环境中收集多模态数据集作为新的基准资源,与参与者一起理解具有不同背景噪声水平的目标语音。该框架使用瞳孔测量数据的类内相关性进行统计评估(平均ICC> 0.95)。瞳孔测量与RF数据通过Pearson相关建立相关性(平均PCC> 0.79)。此外,基于RF数据,使用K-means聚类将CL分为高类和低类。未来的工作包括将射频传感器与眼镜集成,以估计助听器用户的听力努力,并根据个人的听力努力和声环境的复杂性利用射频测量来优化语音增强。
{"title":"Multimodal Cognitive Load Estimation With Radio Frequency Sensing and Pupillometry in Complex Auditory Environments.","authors":"Usman Anwar, Adeel Hussain, Mandar Gogate, Kia Dashtipour, Tughrul Arslan, Amir Hussain, Peter Lomax","doi":"10.1109/JBHI.2025.3634072","DOIUrl":"10.1109/JBHI.2025.3634072","url":null,"abstract":"<p><p>The detection of listening effort or cognitive load (CL) has been a major research challenge in recent years. Most conventional techniques utilise physiological or audio-visual sensors and are privacy-invasive and computationally complex. The challenges of synchronization, data alignment and accessibility limitations potentially increase the noise and error probability, compromising the accuracy of CL estimates. This innovative work presents a multi-modal, non-invasive and privacy-preserving approach that combines Radio Frequency (RF) and pupillometry sensing to address these challenges. Custom RF sensors are first designed and developed to capture blood flow changes in specific brain regions with high spatial resolution. Next, multi-modal fusion with pupillometry sensing is proposed and shown to offer a robust assessment of cognitive and listening effort through pupil size and pupil dilation. Our novel approach evaluates RF sensing to estimate CL from cerebral blood flow variations utilizing pupillometry as a baseline. A first-of-its-kind, multi-modal dataset is collected as a new benchmark resource in a controlled environment with participants to comprehend target speech with varying background noise levels. The framework is statistically evaluated using intraclass correlation for pupillometry data (average ICC> 0.95). The correlation between pupillometry and RF data is established through Pearson's correlation (average PCC> 0.79). Further, CL is classified into high and low categories based on RF data using K-means clustering. Future work involves integrating RF sensors with glasses to estimate listening effort for hearing-aid users and utilising RF measurements to optimize speech enhancement based on individual's listening effort and complexity of acoustic environment.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1605-1617"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145556854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
XAI Driven Intelligent IoMT Secure Data Management Framework. XAI 驱动的智能 IoMT 安全数据管理框架。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2024.3408215
Wei Liu, Feng Zhao, Lewis Nkenyereye, Shalli Rani, Keqin Li, Jianhui Lv

The Internet of Medical Things (IoMT) has transformed traditional healthcare systems by enabling real-time monitoring, remote diagnostics, and data-driven treatment. However, security and privacy remain significant concerns for IoMT adoption due to the sensitive nature of medical data. Therefore, we propose an integrated framework leveraging blockchain and explainable artificial intelligence (XAI) to enable secure, intelligent, and transparent management of IoMT data. First, the traceability and tamper-proof of blockchain are used to realize the secure transaction of IoMT data, transforming the secure transaction of IoMT data into a two-stage Stackelberg game. The dual-chain architecture is used to ensure the security and privacy protection of the transaction. The main-chain manages regular IoMT data transactions, while the side-chain deals with data trading activities aimed at resale. Simultaneously, the perceptual hash technology is used to realize data rights confirmation, which maximally protects the rights and interests of each participant in the transaction. Subsequently, medical time-series data is modeled using bidirectional simple recurrent units to detect anomalies and cyberthreats accurately while overcoming vanishing gradients. Lastly, an adversarial sample generation method based on local interpretable model-agnostic explanations is provided to evaluate, secure, and improve the anomaly detection model, as well as to make it more explainable and resilient to possible adversarial attacks. Simulation results are provided to illustrate the high performance of the integrated secure data management framework leveraging blockchain and XAI, compared with the benchmarks.

医疗物联网(IoMT)实现了实时监控、远程诊断和数据驱动的治疗,从而改变了传统的医疗保健系统。然而,由于医疗数据的敏感性,安全和隐私仍然是采用 IoMT 的重大问题。因此,我们提出了一个利用区块链和可解释人工智能(XAI)的集成框架,以实现 IoMT 数据的安全、智能和透明管理。首先,利用区块链的可追溯性和防篡改性实现 IoMT 数据的安全交易,将 IoMT 数据的安全交易转化为两阶段 Stackelberg 博弈。采用双链架构确保交易的安全性和隐私保护。主链管理常规的 IoMT 数据交易,侧链处理以转售为目的的数据交易活动。同时,利用感知哈希技术实现数据确权,最大限度地保护交易各参与方的权益。随后,利用双向简单递归单元对医疗时间序列数据进行建模,在克服梯度消失的同时准确检测异常和网络威胁。最后,提供了一种基于本地可解释模型的对抗样本生成方法,以评估、保护和改进异常检测模型,并使其更易于解释和抵御可能的对抗攻击。仿真结果表明,与基准相比,利用区块链和 XAI 的集成安全数据管理框架具有很高的性能。
{"title":"XAI Driven Intelligent IoMT Secure Data Management Framework.","authors":"Wei Liu, Feng Zhao, Lewis Nkenyereye, Shalli Rani, Keqin Li, Jianhui Lv","doi":"10.1109/JBHI.2024.3408215","DOIUrl":"10.1109/JBHI.2024.3408215","url":null,"abstract":"<p><p>The Internet of Medical Things (IoMT) has transformed traditional healthcare systems by enabling real-time monitoring, remote diagnostics, and data-driven treatment. However, security and privacy remain significant concerns for IoMT adoption due to the sensitive nature of medical data. Therefore, we propose an integrated framework leveraging blockchain and explainable artificial intelligence (XAI) to enable secure, intelligent, and transparent management of IoMT data. First, the traceability and tamper-proof of blockchain are used to realize the secure transaction of IoMT data, transforming the secure transaction of IoMT data into a two-stage Stackelberg game. The dual-chain architecture is used to ensure the security and privacy protection of the transaction. The main-chain manages regular IoMT data transactions, while the side-chain deals with data trading activities aimed at resale. Simultaneously, the perceptual hash technology is used to realize data rights confirmation, which maximally protects the rights and interests of each participant in the transaction. Subsequently, medical time-series data is modeled using bidirectional simple recurrent units to detect anomalies and cyberthreats accurately while overcoming vanishing gradients. Lastly, an adversarial sample generation method based on local interpretable model-agnostic explanations is provided to evaluate, secure, and improve the anomaly detection model, as well as to make it more explainable and resilient to possible adversarial attacks. Simulation results are provided to illustrate the high performance of the integrated secure data management framework leveraging blockchain and XAI, compared with the benchmarks.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"935-946"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141237831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Biomedical and Health Informatics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1