首页 > 最新文献

IEEE Journal of Biomedical and Health Informatics最新文献

英文 中文
Edge-Aware Diffusion Segmentation Model With Hessian Priors for Automated Diaphragm Thickness Measurement in Ultrasound Imaging. 超声成像中自动测量隔膜厚度的Hessian先验边缘感知扩散分割模型。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3601567
Chen-Long Miao, Yikang He, Baike Shi, Zhongkai Bian, Wenxue Yu, Yang Chen, Guang-Quan Zhou

The thickness of the diaphragm serves as a crucial biometric indicator, particularly in assessing rehabilitation and respiratory dysfunction. However, measuring diaphragm thickness from ultrasound images mainly depends on manual delineation of the fascia, which is subjective, time-consuming, and sensitive to the inherent speckle noise. In this study, we introduce an edge-aware diffusion segmentation model (ESADiff), which incorporates prior structural knowledge of the fascia to improve the accuracy and reliability of diaphragm thickness measurements in ultrasound imaging. We first apply a diffusion model, guided by annotations, to learn the image features while preserving edge details through an iterative denoising process. Specifically, we design an anisotropic edge-sensitive annotation refinement module that corrects inaccurate labels by integrating Hessian geometric priors with a backtracking shortest-path connection algorithm, further enhancing model accuracy. Moreover, a curvature-aware deformable convolution and edge-prior ranking loss function are proposed to leverage the shape prior knowledge of the fascia, allowing the model to selectively focus on relevant linear structures while mitigating the influence of noise on feature extraction. We evaluated the proposed model on an in-house diaphragm ultrasound dataset, a public calf muscle dataset, and an internal tongue muscle dataset to demonstrate robust generalization. Extensive experimental results demonstrate that our method achieves finer fascia segmentation and significantly improves the accuracy of thickness measurements compared to other state-of-the-art techniques, highlighting its potential for clinical applications.

横膈膜的厚度是一个重要的生物特征指标,特别是在评估康复和呼吸功能障碍方面。然而,从超声图像中测量隔膜厚度主要依赖于手工描绘筋膜,这是主观的,耗时的,并且对固有的散斑噪声敏感。在这项研究中,我们引入了一种边缘感知扩散分割模型(ESADiff),该模型结合了先前的筋膜结构知识,以提高超声成像中膈膜厚度测量的准确性和可靠性。我们首先应用一个扩散模型,在注释的指导下,学习图像特征,同时通过迭代去噪过程保留边缘细节。具体而言,我们设计了一个各向异性边缘敏感标注细化模块,通过整合Hessian几何先验和回溯最短路径连接算法来纠正不准确的标签,进一步提高模型精度。此外,提出了曲率感知的可变形卷积和边缘先验排序损失函数,利用筋膜的形状先验知识,使模型能够选择性地关注相关线性结构,同时减轻噪声对特征提取的影响。我们在内部隔膜超声数据集、公共小腿肌肉数据集和内部舌肌肉数据集上评估了所提出的模型,以证明鲁棒泛化。大量的实验结果表明,与其他最先进的技术相比,我们的方法实现了更精细的筋膜分割,显著提高了厚度测量的准确性,突出了其临床应用的潜力。
{"title":"Edge-Aware Diffusion Segmentation Model With Hessian Priors for Automated Diaphragm Thickness Measurement in Ultrasound Imaging.","authors":"Chen-Long Miao, Yikang He, Baike Shi, Zhongkai Bian, Wenxue Yu, Yang Chen, Guang-Quan Zhou","doi":"10.1109/JBHI.2025.3601567","DOIUrl":"10.1109/JBHI.2025.3601567","url":null,"abstract":"<p><p>The thickness of the diaphragm serves as a crucial biometric indicator, particularly in assessing rehabilitation and respiratory dysfunction. However, measuring diaphragm thickness from ultrasound images mainly depends on manual delineation of the fascia, which is subjective, time-consuming, and sensitive to the inherent speckle noise. In this study, we introduce an edge-aware diffusion segmentation model (ESADiff), which incorporates prior structural knowledge of the fascia to improve the accuracy and reliability of diaphragm thickness measurements in ultrasound imaging. We first apply a diffusion model, guided by annotations, to learn the image features while preserving edge details through an iterative denoising process. Specifically, we design an anisotropic edge-sensitive annotation refinement module that corrects inaccurate labels by integrating Hessian geometric priors with a backtracking shortest-path connection algorithm, further enhancing model accuracy. Moreover, a curvature-aware deformable convolution and edge-prior ranking loss function are proposed to leverage the shape prior knowledge of the fascia, allowing the model to selectively focus on relevant linear structures while mitigating the influence of noise on feature extraction. We evaluated the proposed model on an in-house diaphragm ultrasound dataset, a public calf muscle dataset, and an internal tongue muscle dataset to demonstrate robust generalization. Extensive experimental results demonstrate that our method achieves finer fascia segmentation and significantly improves the accuracy of thickness measurements compared to other state-of-the-art techniques, highlighting its potential for clinical applications.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1544-1554"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144952132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Cognitive Load Estimation With Radio Frequency Sensing and Pupillometry in Complex Auditory Environments. 复杂听觉环境下基于射频传感和瞳孔测量的多模态认知负荷估计。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3634072
Usman Anwar, Adeel Hussain, Mandar Gogate, Kia Dashtipour, Tughrul Arslan, Amir Hussain, Peter Lomax

The detection of listening effort or cognitive load (CL) has been a major research challenge in recent years. Most conventional techniques utilise physiological or audio-visual sensors and are privacy-invasive and computationally complex. The challenges of synchronization, data alignment and accessibility limitations potentially increase the noise and error probability, compromising the accuracy of CL estimates. This innovative work presents a multi-modal, non-invasive and privacy-preserving approach that combines Radio Frequency (RF) and pupillometry sensing to address these challenges. Custom RF sensors are first designed and developed to capture blood flow changes in specific brain regions with high spatial resolution. Next, multi-modal fusion with pupillometry sensing is proposed and shown to offer a robust assessment of cognitive and listening effort through pupil size and pupil dilation. Our novel approach evaluates RF sensing to estimate CL from cerebral blood flow variations utilizing pupillometry as a baseline. A first-of-its-kind, multi-modal dataset is collected as a new benchmark resource in a controlled environment with participants to comprehend target speech with varying background noise levels. The framework is statistically evaluated using intraclass correlation for pupillometry data (average ICC> 0.95). The correlation between pupillometry and RF data is established through Pearson's correlation (average PCC> 0.79). Further, CL is classified into high and low categories based on RF data using K-means clustering. Future work involves integrating RF sensors with glasses to estimate listening effort for hearing-aid users and utilising RF measurements to optimize speech enhancement based on individual's listening effort and complexity of acoustic environment.

听力努力或认知负荷(CL)的检测是近年来研究的一个重大挑战。大多数传统技术使用生理或视听传感器,并且侵犯隐私且计算复杂。同步、数据对齐和可访问性限制的挑战可能会增加噪声和错误概率,从而影响CL估计的准确性。这项创新工作提出了一种多模式、非侵入性和保护隐私的方法,该方法结合了射频(RF)和瞳孔测量传感来应对这些挑战。定制的射频传感器首先被设计和开发,以高空间分辨率捕获特定大脑区域的血流变化。接下来,提出了瞳孔测量传感的多模态融合,并通过瞳孔大小和瞳孔扩张提供了对认知和听力努力的可靠评估。我们的新方法评估射频传感,以估计CL从脑血流变化利用瞳孔测量作为基线。首先,在受控环境中收集多模态数据集作为新的基准资源,与参与者一起理解具有不同背景噪声水平的目标语音。该框架使用瞳孔测量数据的类内相关性进行统计评估(平均ICC> 0.95)。瞳孔测量与RF数据通过Pearson相关建立相关性(平均PCC> 0.79)。此外,基于RF数据,使用K-means聚类将CL分为高类和低类。未来的工作包括将射频传感器与眼镜集成,以估计助听器用户的听力努力,并根据个人的听力努力和声环境的复杂性利用射频测量来优化语音增强。
{"title":"Multimodal Cognitive Load Estimation With Radio Frequency Sensing and Pupillometry in Complex Auditory Environments.","authors":"Usman Anwar, Adeel Hussain, Mandar Gogate, Kia Dashtipour, Tughrul Arslan, Amir Hussain, Peter Lomax","doi":"10.1109/JBHI.2025.3634072","DOIUrl":"10.1109/JBHI.2025.3634072","url":null,"abstract":"<p><p>The detection of listening effort or cognitive load (CL) has been a major research challenge in recent years. Most conventional techniques utilise physiological or audio-visual sensors and are privacy-invasive and computationally complex. The challenges of synchronization, data alignment and accessibility limitations potentially increase the noise and error probability, compromising the accuracy of CL estimates. This innovative work presents a multi-modal, non-invasive and privacy-preserving approach that combines Radio Frequency (RF) and pupillometry sensing to address these challenges. Custom RF sensors are first designed and developed to capture blood flow changes in specific brain regions with high spatial resolution. Next, multi-modal fusion with pupillometry sensing is proposed and shown to offer a robust assessment of cognitive and listening effort through pupil size and pupil dilation. Our novel approach evaluates RF sensing to estimate CL from cerebral blood flow variations utilizing pupillometry as a baseline. A first-of-its-kind, multi-modal dataset is collected as a new benchmark resource in a controlled environment with participants to comprehend target speech with varying background noise levels. The framework is statistically evaluated using intraclass correlation for pupillometry data (average ICC> 0.95). The correlation between pupillometry and RF data is established through Pearson's correlation (average PCC> 0.79). Further, CL is classified into high and low categories based on RF data using K-means clustering. Future work involves integrating RF sensors with glasses to estimate listening effort for hearing-aid users and utilising RF measurements to optimize speech enhancement based on individual's listening effort and complexity of acoustic environment.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1605-1617"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145556854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
XAI Driven Intelligent IoMT Secure Data Management Framework. XAI 驱动的智能 IoMT 安全数据管理框架。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2024.3408215
Wei Liu, Feng Zhao, Lewis Nkenyereye, Shalli Rani, Keqin Li, Jianhui Lv

The Internet of Medical Things (IoMT) has transformed traditional healthcare systems by enabling real-time monitoring, remote diagnostics, and data-driven treatment. However, security and privacy remain significant concerns for IoMT adoption due to the sensitive nature of medical data. Therefore, we propose an integrated framework leveraging blockchain and explainable artificial intelligence (XAI) to enable secure, intelligent, and transparent management of IoMT data. First, the traceability and tamper-proof of blockchain are used to realize the secure transaction of IoMT data, transforming the secure transaction of IoMT data into a two-stage Stackelberg game. The dual-chain architecture is used to ensure the security and privacy protection of the transaction. The main-chain manages regular IoMT data transactions, while the side-chain deals with data trading activities aimed at resale. Simultaneously, the perceptual hash technology is used to realize data rights confirmation, which maximally protects the rights and interests of each participant in the transaction. Subsequently, medical time-series data is modeled using bidirectional simple recurrent units to detect anomalies and cyberthreats accurately while overcoming vanishing gradients. Lastly, an adversarial sample generation method based on local interpretable model-agnostic explanations is provided to evaluate, secure, and improve the anomaly detection model, as well as to make it more explainable and resilient to possible adversarial attacks. Simulation results are provided to illustrate the high performance of the integrated secure data management framework leveraging blockchain and XAI, compared with the benchmarks.

医疗物联网(IoMT)实现了实时监控、远程诊断和数据驱动的治疗,从而改变了传统的医疗保健系统。然而,由于医疗数据的敏感性,安全和隐私仍然是采用 IoMT 的重大问题。因此,我们提出了一个利用区块链和可解释人工智能(XAI)的集成框架,以实现 IoMT 数据的安全、智能和透明管理。首先,利用区块链的可追溯性和防篡改性实现 IoMT 数据的安全交易,将 IoMT 数据的安全交易转化为两阶段 Stackelberg 博弈。采用双链架构确保交易的安全性和隐私保护。主链管理常规的 IoMT 数据交易,侧链处理以转售为目的的数据交易活动。同时,利用感知哈希技术实现数据确权,最大限度地保护交易各参与方的权益。随后,利用双向简单递归单元对医疗时间序列数据进行建模,在克服梯度消失的同时准确检测异常和网络威胁。最后,提供了一种基于本地可解释模型的对抗样本生成方法,以评估、保护和改进异常检测模型,并使其更易于解释和抵御可能的对抗攻击。仿真结果表明,与基准相比,利用区块链和 XAI 的集成安全数据管理框架具有很高的性能。
{"title":"XAI Driven Intelligent IoMT Secure Data Management Framework.","authors":"Wei Liu, Feng Zhao, Lewis Nkenyereye, Shalli Rani, Keqin Li, Jianhui Lv","doi":"10.1109/JBHI.2024.3408215","DOIUrl":"10.1109/JBHI.2024.3408215","url":null,"abstract":"<p><p>The Internet of Medical Things (IoMT) has transformed traditional healthcare systems by enabling real-time monitoring, remote diagnostics, and data-driven treatment. However, security and privacy remain significant concerns for IoMT adoption due to the sensitive nature of medical data. Therefore, we propose an integrated framework leveraging blockchain and explainable artificial intelligence (XAI) to enable secure, intelligent, and transparent management of IoMT data. First, the traceability and tamper-proof of blockchain are used to realize the secure transaction of IoMT data, transforming the secure transaction of IoMT data into a two-stage Stackelberg game. The dual-chain architecture is used to ensure the security and privacy protection of the transaction. The main-chain manages regular IoMT data transactions, while the side-chain deals with data trading activities aimed at resale. Simultaneously, the perceptual hash technology is used to realize data rights confirmation, which maximally protects the rights and interests of each participant in the transaction. Subsequently, medical time-series data is modeled using bidirectional simple recurrent units to detect anomalies and cyberthreats accurately while overcoming vanishing gradients. Lastly, an adversarial sample generation method based on local interpretable model-agnostic explanations is provided to evaluate, secure, and improve the anomaly detection model, as well as to make it more explainable and resilient to possible adversarial attacks. Simulation results are provided to illustrate the high performance of the integrated secure data management framework leveraging blockchain and XAI, compared with the benchmarks.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"935-946"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141237831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Semantic Conditional Diffusion Model for Enhanced Personal Privacy Preservation in Medical Images. 增强医学影像个人隐私保护的语义条件扩散模型
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2024.3511583
Shudong Wang, Zhiyuan Zhao, Yawu Zhao, Luqi Wang, Yuanyuan Zhang, Jiehuan Wang, Sibo Qiao, Zhihan Lyu

Deep learning has significantly advanced medical image processing, yet the inherent inclusion of personally identifiable information (PII) within medical images-such as facial features, distinctive anatomical structures, rare lesions, or specific textural patterns-poses a critical risk to patient privacy during data transmission. To mitigate this risk, we introduce the Medical Semantic Diffusion Model (MSDM), a novel framework designed to synthesize medical images guided by semantic information, synthesis images with the same distribution as the original data, which effectively removes the PPI of the original data to ensure robust privacy protection. Unlike conventional techniques that combine semantic and noisy images for denoising, MSDM integrates Adaptive Batch Normalization (AdaBN) to encode semantic information into high-dimensional latent space, embedding it directly within the denoising neural network. This approach enhances image quality and semantic accuracy while ensuring that the synthetic and original images belong to the same distribution. In addition, to further accelerate synthesis and reduce dependency on manually crafted semantic masks, we propose the Spread Algorithm, which automatically generates these masks. Extensive experiments conducted on the BraTS 2021, MSD Lung, DSB18, and FIVES datasets confirm the efficacy of MSDM, yielding state-of-the-art results across several performance metrics. Augmenting datasets with MSDM-generated images in nnUNet segmentation experiments led to Dice scores of 0.6243, 0.9531, 0.9406, and 0.9562 underscoring its potential for enhancing both image quality and privacy-preserving data augmentation.

深度学习在医学图像处理方面有着显著的进步,但医学图像中固有的个人身份信息(PII),如面部特征、独特的解剖结构、罕见的病变或特定的纹理模式,在数据传输过程中对患者隐私构成了严重的风险。为了降低这种风险,我们引入了医学语义扩散模型(Medical Semantic Diffusion Model, MSDM),这是一种基于语义信息的医学图像合成框架,合成的图像与原始数据具有相同的分布,有效地去除了原始数据的PPI,从而保证了对隐私的鲁棒性保护。与传统的结合语义和噪声图像进行去噪的技术不同,MSDM集成了自适应批归一化(AdaBN)将语义信息编码到高维潜在空间中,并将其直接嵌入到去噪神经网络中。该方法在保证合成图像和原始图像属于同一分布的同时,提高了图像质量和语义精度。此外,为了进一步加快合成速度并减少对手工制作的语义掩码的依赖,我们提出了自动生成这些掩码的扩展算法。在BraTS 2021、MSD Lung、DSB18和fifs数据集上进行的大量实验证实了MSDM的有效性,在几个性能指标上产生了最先进的结果。在nnUNet分割实验中,使用msdm生成的图像增强数据集的Dice得分分别为0.6243、0.9531、0.9406和0.9562,这表明msdm在增强图像质量和保护隐私的数据增强方面具有潜力。
{"title":"A Semantic Conditional Diffusion Model for Enhanced Personal Privacy Preservation in Medical Images.","authors":"Shudong Wang, Zhiyuan Zhao, Yawu Zhao, Luqi Wang, Yuanyuan Zhang, Jiehuan Wang, Sibo Qiao, Zhihan Lyu","doi":"10.1109/JBHI.2024.3511583","DOIUrl":"10.1109/JBHI.2024.3511583","url":null,"abstract":"<p><p>Deep learning has significantly advanced medical image processing, yet the inherent inclusion of personally identifiable information (PII) within medical images-such as facial features, distinctive anatomical structures, rare lesions, or specific textural patterns-poses a critical risk to patient privacy during data transmission. To mitigate this risk, we introduce the Medical Semantic Diffusion Model (MSDM), a novel framework designed to synthesize medical images guided by semantic information, synthesis images with the same distribution as the original data, which effectively removes the PPI of the original data to ensure robust privacy protection. Unlike conventional techniques that combine semantic and noisy images for denoising, MSDM integrates Adaptive Batch Normalization (AdaBN) to encode semantic information into high-dimensional latent space, embedding it directly within the denoising neural network. This approach enhances image quality and semantic accuracy while ensuring that the synthetic and original images belong to the same distribution. In addition, to further accelerate synthesis and reduce dependency on manually crafted semantic masks, we propose the Spread Algorithm, which automatically generates these masks. Extensive experiments conducted on the BraTS 2021, MSD Lung, DSB18, and FIVES datasets confirm the efficacy of MSDM, yielding state-of-the-art results across several performance metrics. Augmenting datasets with MSDM-generated images in nnUNet segmentation experiments led to Dice scores of 0.6243, 0.9531, 0.9406, and 0.9562 underscoring its potential for enhancing both image quality and privacy-preserving data augmentation.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"853-864"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143541494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Plausible Deniable Medical Image Encryption by Large Language Models and Reversible Content-Aware Strategy. 基于大语言模型和可逆内容感知策略的可信可否认医学图像加密。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3565271
Yirui Wu, Xinfu Liu, Lucia Cascone, Michele Nappi, Shaohua Wan

There is a rising concern about healthcare system security, where data loss could bring lots of damages to patients and hospitals. As a promising encryption method for medical images, DNA encoding own characteristics of high speed, parallelism computation, minimal storage, and unbreakable cryptosystems. Inspired by the idea of involving Large Language Models(LLMs) to improve DNA encoding, we propose a medical image encryption method with LLM-enhanced DNA encoding, which consists of LLM enhancing module and content-aware permutation&diffusion module. Regarding medical images generally have plain backgrounds with low-entropy pixels, the first module compresses pixels into highly compact signals with features of probabilistic varying and plausibly deniability, serving as another LLM-based layer of defense against privacy breaches before DNA encoding. The second module not only adds permutation by randomly sampling from a redundant correlation between adjacent pixels to break the internal links between pixels but also performs a DNA-based diffusion process to greatly increase the complexity of cracking. Experiments on ChestXray-14, COVID-CT and fcon-1000 datasets show that the proposed method outperforms all comparative methods in sensitivity, correlation and entropy.

人们越来越关注医疗系统的安全性,数据丢失可能会给患者和医院带来很大的损失。DNA编码具有速度快、并行计算、存储空间小、密码系统不可破解等特点,是一种很有前途的医学图像加密方法。受利用大语言模型(LLM)改进DNA编码的思想启发,我们提出了一种基于LLM增强DNA编码的医学图像加密方法,该方法由LLM增强模块和内容感知置换扩散模块组成。由于医学图像通常具有低熵像素的普通背景,第一个模块将像素压缩成具有概率变化和合理否认特征的高度紧凑的信号,作为DNA编码之前的另一个基于llm的隐私泄露防御层。第二个模块不仅通过随机采样相邻像素之间的冗余相关性来增加排列,打破像素之间的内部联系,而且还进行了基于dna的扩散过程,大大增加了破解的复杂性。在ChestXray-14、COVID-CT和fcon-1000数据集上的实验表明,该方法在灵敏度、相关性和熵方面优于所有比较方法。
{"title":"Plausible Deniable Medical Image Encryption by Large Language Models and Reversible Content-Aware Strategy.","authors":"Yirui Wu, Xinfu Liu, Lucia Cascone, Michele Nappi, Shaohua Wan","doi":"10.1109/JBHI.2025.3565271","DOIUrl":"10.1109/JBHI.2025.3565271","url":null,"abstract":"<p><p>There is a rising concern about healthcare system security, where data loss could bring lots of damages to patients and hospitals. As a promising encryption method for medical images, DNA encoding own characteristics of high speed, parallelism computation, minimal storage, and unbreakable cryptosystems. Inspired by the idea of involving Large Language Models(LLMs) to improve DNA encoding, we propose a medical image encryption method with LLM-enhanced DNA encoding, which consists of LLM enhancing module and content-aware permutation&diffusion module. Regarding medical images generally have plain backgrounds with low-entropy pixels, the first module compresses pixels into highly compact signals with features of probabilistic varying and plausibly deniability, serving as another LLM-based layer of defense against privacy breaches before DNA encoding. The second module not only adds permutation by randomly sampling from a redundant correlation between adjacent pixels to break the internal links between pixels but also performs a DNA-based diffusion process to greatly increase the complexity of cracking. Experiments on ChestXray-14, COVID-CT and fcon-1000 datasets show that the proposed method outperforms all comparative methods in sensitivity, correlation and entropy.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"947-957"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143964998","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FusSADGCNN: Decoding the Impact of Transcranial Electrical Stimulation on Neuromodulation in Emotion Recognition and Emotion Elicitation. 解读经颅电刺激对情绪识别和情绪激发神经调节的影响。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3586906
Congming Tan, Jiayang Xu, Liangliang Hu, Yin Tian

Emotional neuromodulation refers to the direct manipulation of the nervous system using techniques such as electrical or magnetic stimulation to manage and adjust an individual's emotional experiences. Transcranial electrical stimulation (tES) targeting the right ventrolateral prefrontal cortex (rVLPFC) has been widely used to modulate emotions. However, the impact of emotions on brain network changes and modulation during tES remains unclear. In this study, we developed a subject-adaptive dynamic graph convolution network with fused features (FusSADGCNN) to decode the impact of tES on neuromodulation for emotion recognition and emotion elicitation. Specifically, we developed a fused feature, CPE, which integrates the average sub-frequency phase-locking value representing global functional connectivity with differential entropy characterizing local activation to explore network differences across emotional states, while incorporating an improved dynamic graph convolution to adaptively integrate multi-receptive neighborhood information for precise decoding of individual tES effects. On the SEED dataset and our laboratory data, the FusSADGCNN model outperforms the state-of-the-art methods. Furthermore, we utilized these tools to assess the emotional modulation states induced by tES. Results indicated that in the experiment involving music-elicited emotional modulation, the tools effectively identified improvements in negative emotions under true stimulation, with predictive accuracy significantly related to the average connectivity strength of the brain network. In the active facial emotion recognition modulation experiment, jointed stimulation of rVLPFC and temporo-parietal junction achieved better modulation effects. These findings highlight that the FusSADGCNN effectively evaluate the neuromodulation states during tES-induced emotional regulation, providing a reliable foundation for integrating emotion recognition and neuromodulation.

情绪神经调节是指使用电或磁刺激等技术直接操纵神经系统来管理和调整个人的情绪体验。经颅电刺激(tES)靶向右腹外侧前额叶皮层(rVLPFC)已被广泛用于调节情绪。然而,情绪对te期间大脑网络变化和调节的影响尚不清楚。在这项研究中,我们开发了一个具有融合特征的主体自适应动态图卷积网络(FusSADGCNN)来解码tES对情绪识别和情绪激发的神经调节的影响。具体来说,我们开发了一种融合特征CPE,它将代表全局功能连通性的平均次频锁相值与表征局部激活的微分熵相结合,以探索不同情绪状态下的网络差异,同时结合改进的动态图卷积自适应整合多接受邻域信息,以精确解码个体te效应。在SEED数据集和我们的实验室数据上,FusSADGCNN模型优于最先进的方法。此外,我们利用这些工具来评估tES诱导的情绪调节状态。结果表明,在涉及音乐诱发情绪调节的实验中,这些工具有效地识别了真实刺激下负面情绪的改善,预测准确性与大脑网络的平均连接强度显著相关。在主动面部情绪识别调制实验中,rVLPFC和颞顶叶连接的联合刺激获得了更好的调制效果。这些结果表明,FusSADGCNN能够有效评估tes诱导的情绪调节过程中的神经调节状态,为情绪识别和神经调节的整合提供了可靠的基础。
{"title":"FusSADGCNN: Decoding the Impact of Transcranial Electrical Stimulation on Neuromodulation in Emotion Recognition and Emotion Elicitation.","authors":"Congming Tan, Jiayang Xu, Liangliang Hu, Yin Tian","doi":"10.1109/JBHI.2025.3586906","DOIUrl":"10.1109/JBHI.2025.3586906","url":null,"abstract":"<p><p>Emotional neuromodulation refers to the direct manipulation of the nervous system using techniques such as electrical or magnetic stimulation to manage and adjust an individual's emotional experiences. Transcranial electrical stimulation (tES) targeting the right ventrolateral prefrontal cortex (rVLPFC) has been widely used to modulate emotions. However, the impact of emotions on brain network changes and modulation during tES remains unclear. In this study, we developed a subject-adaptive dynamic graph convolution network with fused features (FusSADGCNN) to decode the impact of tES on neuromodulation for emotion recognition and emotion elicitation. Specifically, we developed a fused feature, CPE, which integrates the average sub-frequency phase-locking value representing global functional connectivity with differential entropy characterizing local activation to explore network differences across emotional states, while incorporating an improved dynamic graph convolution to adaptively integrate multi-receptive neighborhood information for precise decoding of individual tES effects. On the SEED dataset and our laboratory data, the FusSADGCNN model outperforms the state-of-the-art methods. Furthermore, we utilized these tools to assess the emotional modulation states induced by tES. Results indicated that in the experiment involving music-elicited emotional modulation, the tools effectively identified improvements in negative emotions under true stimulation, with predictive accuracy significantly related to the average connectivity strength of the brain network. In the active facial emotion recognition modulation experiment, jointed stimulation of rVLPFC and temporo-parietal junction achieved better modulation effects. These findings highlight that the FusSADGCNN effectively evaluate the neuromodulation states during tES-induced emotional regulation, providing a reliable foundation for integrating emotion recognition and neuromodulation.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1087-1100"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144600248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating ECG and PCG Signals through a Dual-Modal ViT for Coronary Artery Disease Detection. 利用双模态ViT集成ECG和PCG信号用于冠状动脉疾病检测。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3589257
Xu Liu, Ling You, Chengcong Lv, Mingyuan Chen, Lianhuan Wei, Yineng Zheng, Xingming Guo

Cardiovascular disease (CVD) remains the leading cause of mortality worldwide, with coronary artery disease (CAD) being the most prevalent form. To improve screening efficiency, there is a critical need for accurate, non-invasive, and cost-effective CAD detection methods. This study presents Co-Attention Dual-Modal ViT (CAD-ViT), a novel classification framework based on the Vision Transformer that integrates both electrocardiogram (ECG) and phonocardiogram (PCG) signals. Unlike prior approaches that process ECG and PCG features independently or fuse them through simple concatenation, the proposed model introduces two key modules: a Co-Attention mechanism that enables bidirectional cross-modal interaction to effectively capture complementary features between ECG and PCG signals, and a Dynamic Weighted Fusion (DWF) module that adaptively adjusts the contribution of each modality for robust feature fusion. CAD-ViT is evaluated on a private clinical dataset comprising 132 CAD and 101 non-CAD subjects, achieving an accuracy of 97.08%, precision of 97.18%, specificity of 98.52%, F1-score of 97.04, and recall of 96.94%. Additional validation on two public datasets confirms the model's robustness and generalization capability. These results demonstrate the effectiveness of the proposed approach and its potential for practical deployment in CAD screening using multimodal biosignals.

心血管疾病(CVD)仍然是世界范围内死亡的主要原因,冠状动脉疾病(CAD)是最普遍的形式。为了提高筛查效率,迫切需要准确、无创、经济高效的CAD检测方法。本研究提出了一种基于视觉变换器的新分类框架——共同注意双模态ViT (CAD-ViT),该框架集成了心电图(ECG)和心音图(PCG)信号。与之前独立处理ECG和PCG特征或通过简单连接融合它们的方法不同,该模型引入了两个关键模块:一个共同注意机制,使双向跨模态交互能够有效捕获ECG和PCG信号之间的互补特征,以及一个动态加权融合(DWF)模块,该模块自适应调整每个模态的贡献,以实现稳健的特征融合。CAD- vit在包含132例CAD和101例非CAD受试者的私人临床数据集上进行评估,准确率为97.08%,精密度为97.18%,特异性为98.52%,f1评分为97.04,召回率为96.94%。在两个公共数据集上的额外验证证实了该模型的鲁棒性和泛化能力。这些结果证明了所提出的方法的有效性及其在使用多模态生物信号的CAD筛选中实际部署的潜力。
{"title":"Integrating ECG and PCG Signals through a Dual-Modal ViT for Coronary Artery Disease Detection.","authors":"Xu Liu, Ling You, Chengcong Lv, Mingyuan Chen, Lianhuan Wei, Yineng Zheng, Xingming Guo","doi":"10.1109/JBHI.2025.3589257","DOIUrl":"10.1109/JBHI.2025.3589257","url":null,"abstract":"<p><p>Cardiovascular disease (CVD) remains the leading cause of mortality worldwide, with coronary artery disease (CAD) being the most prevalent form. To improve screening efficiency, there is a critical need for accurate, non-invasive, and cost-effective CAD detection methods. This study presents Co-Attention Dual-Modal ViT (CAD-ViT), a novel classification framework based on the Vision Transformer that integrates both electrocardiogram (ECG) and phonocardiogram (PCG) signals. Unlike prior approaches that process ECG and PCG features independently or fuse them through simple concatenation, the proposed model introduces two key modules: a Co-Attention mechanism that enables bidirectional cross-modal interaction to effectively capture complementary features between ECG and PCG signals, and a Dynamic Weighted Fusion (DWF) module that adaptively adjusts the contribution of each modality for robust feature fusion. CAD-ViT is evaluated on a private clinical dataset comprising 132 CAD and 101 non-CAD subjects, achieving an accuracy of 97.08%, precision of 97.18%, specificity of 98.52%, F1-score of 97.04, and recall of 96.94%. Additional validation on two public datasets confirms the model's robustness and generalization capability. These results demonstrate the effectiveness of the proposed approach and its potential for practical deployment in CAD screening using multimodal biosignals.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1128-1139"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144642431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy Preserved Blood Glucose Level Cross-Prediction: An Asynchronous Decentralized Federated Learning Approach. 隐私保护血糖水平交叉预测:一种异步分散的联邦学习方法。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3573954
Chengzhe Piao, Taiyu Zhu, Yu Wang, Stephanie E Baldeweg, Paul Taylor, Pantelis Georgiou, Jiahao Sun, Jun Wang, Kezhi Li

Newly diagnosed Type 1 Diabetes (T1D) patients often struggle to obtain effective Blood Glucose (BG) prediction models due to the lack of sufficient BG data from Continuous Glucose Monitoring (CGM), presenting a significant "cold start" problem in patient care. Utilizing population models to address this challenge is a potential solution, but collecting patient data for training population models in a privacy-conscious manner is challenging, especially given that such data is often stored on personal devices. Considering the privacy protection and addressing the "cold start" problem in diabetes care, we propose "GluADFL", blood Glucose prediction by Asynchronous Decentralized Federated Learning. We compared GluADFL with eight baseline methods using four distinct T1D datasets, comprising 298 participants, which demonstrated its superior performance in accurately predicting BG levels for cross-patient analysis. Furthermore, patients' data might be stored and shared across various communication networks in GluADFL, ranging from highly interconnected (e.g., random, performs the best among others) to more structured topologies (e.g., cluster and ring), suitable for various social networks. The asynchronous training framework supports flexible participation. By adjusting the ratios of inactive participants, we found it remains stable if less than 70% are inactive. Our results confirm that GluADFL offers a practical, privacy-preserved solution for BG prediction in T1D, significantly enhancing the quality of diabetes management.

由于连续血糖监测(CGM)缺乏足够的血糖数据,新诊断的1型糖尿病(T1D)患者往往难以获得有效的血糖(BG)预测模型,这在患者护理中提出了一个重要的“冷启动”问题。利用人口模型来应对这一挑战是一个潜在的解决方案,但是以一种具有隐私意识的方式收集患者数据以训练人口模型是具有挑战性的,特别是考虑到这些数据通常存储在个人设备上。考虑到隐私保护和解决糖尿病护理中的“冷启动”问题,我们提出了“GluADFL”,即基于异步分散联邦学习的血糖预测。我们使用四种不同的T1D数据集(包括298名参与者)将GluADFL与八种基线方法进行了比较,结果表明其在跨患者分析中准确预测BG水平方面具有优越的性能。此外,在GluADFL中,患者的数据可以在各种通信网络中存储和共享,从高度互连(例如,随机的,在其他网络中表现最好)到更结构化的拓扑(例如,集群和环),适用于各种社交网络。异步培训框架支持灵活的参与。通过调整不运动参与者的比例,我们发现如果不运动的比例低于70%,则保持稳定。我们的研究结果证实,GluADFL为T1D患者的血糖预测提供了一种实用的、保密的解决方案,显著提高了糖尿病管理的质量。
{"title":"Privacy Preserved Blood Glucose Level Cross-Prediction: An Asynchronous Decentralized Federated Learning Approach.","authors":"Chengzhe Piao, Taiyu Zhu, Yu Wang, Stephanie E Baldeweg, Paul Taylor, Pantelis Georgiou, Jiahao Sun, Jun Wang, Kezhi Li","doi":"10.1109/JBHI.2025.3573954","DOIUrl":"10.1109/JBHI.2025.3573954","url":null,"abstract":"<p><p>Newly diagnosed Type 1 Diabetes (T1D) patients often struggle to obtain effective Blood Glucose (BG) prediction models due to the lack of sufficient BG data from Continuous Glucose Monitoring (CGM), presenting a significant \"cold start\" problem in patient care. Utilizing population models to address this challenge is a potential solution, but collecting patient data for training population models in a privacy-conscious manner is challenging, especially given that such data is often stored on personal devices. Considering the privacy protection and addressing the \"cold start\" problem in diabetes care, we propose \"GluADFL\", blood Glucose prediction by Asynchronous Decentralized Federated Learning. We compared GluADFL with eight baseline methods using four distinct T1D datasets, comprising 298 participants, which demonstrated its superior performance in accurately predicting BG levels for cross-patient analysis. Furthermore, patients' data might be stored and shared across various communication networks in GluADFL, ranging from highly interconnected (e.g., random, performs the best among others) to more structured topologies (e.g., cluster and ring), suitable for various social networks. The asynchronous training framework supports flexible participation. By adjusting the ratios of inactive participants, we found it remains stable if less than 70% are inactive. Our results confirm that GluADFL offers a practical, privacy-preserved solution for BG prediction in T1D, significantly enhancing the quality of diabetes management.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"839-852"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144855062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MFDP: Multi-View Feature Integration and Enhanced Disease Prompting for Radiology Report Generation. MFDP:多视图特征集成和增强的疾病提示放射学报告生成。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3632356
Yongxu Zhao, Kequan Yang, Yuanchen Wu, Xiaoqiang Li

Radiology report generation aims to automatically produce diagnostic reports from medical images, reducing radiologists' workload. Most existing models commonly use an encoder-decoder architecture, where the text decoder generates reports based on encoded image tokens. However, these approaches have two major limitations: 1) they always use a single-view feature or simple static fusion multi-view feature, which fails to capture complementary information from multi-view images, and 2) they lack explicit diagnostic information related to the disease during the text decoding process, resulting in reduced clinical accuracy and relevance of the generated report. To deal with the above limitations, this paper proposes a novel framework employing Multi-view Feature Integration and Enhanced Disease Prompting for Radiology Report Generation, called MFDP. Specifically, MFDP introduces two key innovations:1) the Multi-view Feature Fusion (MFF) module is designed to dynamically integrate multi-view images (e.g., frontal and lateral views) through a multi-view attention mechanism that adaptively captures inter-view dependencies, enriching the decoder's input features to generate more comprehensive reports. 2) the Enhanced Disease Prompting (EDP) module is designed to provide explicit diagnostic information by constructing enhanced disease prompts to guide the text decoding process. Experiments on two benchmark datasets, MIMIC-CXR and IU X-Ray, demonstrate that the proposed MFDP is competitive in both Clinical Efficacy (CE) and Natural Language Generation (NLG) metrics. Notably, MFDP achieves a 10% average improvement in CE Recall compared to SOTA models, enabling more precise localization of critical abnormalities while maintaining diagnostic completeness.

放射学报告生成旨在从医学图像中自动生成诊断报告,减少放射科医生的工作量。大多数现有模型通常使用编码器-解码器体系结构,其中文本解码器根据编码的图像令牌生成报告。然而,这些方法有两个主要的局限性:1)它们总是使用单视图特征或简单的静态融合多视图特征,无法从多视图图像中捕获互补信息;2)它们在文本解码过程中缺乏明确的与疾病相关的诊断信息,导致生成报告的临床准确性和相关性降低。为了解决上述局限性,本文提出了一种采用多视图特征集成和增强疾病提示的放射学报告生成框架,称为MFDP。具体来说,MFDP引入了两个关键创新:1)多视图特征融合(MFF)模块旨在通过自适应捕获视图间依赖关系的多视图注意机制动态集成多视图图像(例如,正面和侧面视图),丰富解码器的输入特征以生成更全面的报告。2)增强疾病提示(Enhanced Disease prompts, EDP)模块通过构建增强疾病提示来指导文本解码过程,提供明确的诊断信息。在MIMIC-CXR和IU X-Ray两个基准数据集上的实验表明,所提出的MFDP在临床疗效(CE)和自然语言生成(NLG)指标上都具有竞争力。值得注意的是,与SOTA模型相比,MFDP在CE召回方面平均提高了10%,能够更精确地定位关键异常,同时保持诊断的完整性。
{"title":"MFDP: Multi-View Feature Integration and Enhanced Disease Prompting for Radiology Report Generation.","authors":"Yongxu Zhao, Kequan Yang, Yuanchen Wu, Xiaoqiang Li","doi":"10.1109/JBHI.2025.3632356","DOIUrl":"10.1109/JBHI.2025.3632356","url":null,"abstract":"<p><p>Radiology report generation aims to automatically produce diagnostic reports from medical images, reducing radiologists' workload. Most existing models commonly use an encoder-decoder architecture, where the text decoder generates reports based on encoded image tokens. However, these approaches have two major limitations: 1) they always use a single-view feature or simple static fusion multi-view feature, which fails to capture complementary information from multi-view images, and 2) they lack explicit diagnostic information related to the disease during the text decoding process, resulting in reduced clinical accuracy and relevance of the generated report. To deal with the above limitations, this paper proposes a novel framework employing Multi-view Feature Integration and Enhanced Disease Prompting for Radiology Report Generation, called MFDP. Specifically, MFDP introduces two key innovations:1) the Multi-view Feature Fusion (MFF) module is designed to dynamically integrate multi-view images (e.g., frontal and lateral views) through a multi-view attention mechanism that adaptively captures inter-view dependencies, enriching the decoder's input features to generate more comprehensive reports. 2) the Enhanced Disease Prompting (EDP) module is designed to provide explicit diagnostic information by constructing enhanced disease prompts to guide the text decoding process. Experiments on two benchmark datasets, MIMIC-CXR and IU X-Ray, demonstrate that the proposed MFDP is competitive in both Clinical Efficacy (CE) and Natural Language Generation (NLG) metrics. Notably, MFDP achieves a 10% average improvement in CE Recall compared to SOTA models, enabling more precise localization of critical abnormalities while maintaining diagnostic completeness.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1378-1391"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145512608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Plantar Perfusion Imaging for Peripheral Arterial Disease Screening: A Proof-of-Concept Study. 足底灌注成像外周动脉疾病筛查:一项概念验证研究。
IF 6.8 2区 医学 Q1 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 DOI: 10.1109/JBHI.2025.3591844
Yukai Huang, Ningbo Zhao, Dongmin Huang, Yonglong Ye, Zi Luo, Hongzhou Lu, Min He, Wenjin Wang

The diagnosis of peripheral artery disease (PAD) typically relies on specialized equipment such as ultrasound. The delayed PAD detection of these approaches may lead to amputation and even death. To achieve rapid and ubiquitous PAD screening, we propose a novel concept of camera-based plantar perfusion imaging (CPPI) for PAD diagnosis and severity classification. Specifically, we performed a simulation trial that used an RGB camera to record the plantar video of 20 subjects and a cuff with different pressures applied to the left leg to simulate different degrees of lower limb blockage. We generated the plantar perfusion maps using remote photoplethysmography imaging and proposed a multi-view perfusion (MVP) feature set to represent the perfusion maps for PAD classification. The experimental results show that the Pearson correlation coefficients between MVP and Doppler ultrasound (clinical reference) features were larger than 0.9. MVP feature combined with Support Vector Machine obtains 91.47% accuracy in distinguishing the normal and obstructed states, and 76.48% accuracy in differentiating four different degrees of vascular obstruction. The clinical benchmark demonstrated the potential of CPPI as a rapid, sensitive, and easy-to-use diagnostic tool for PAD, suitable for large-scale screening in home or community settings.

外周动脉疾病(PAD)的诊断通常依赖于专门的设备,如超声。这些入路的PAD检测延迟可能导致截肢甚至死亡。为了实现快速和普遍的PAD筛查,我们提出了一种基于相机的足底灌注成像(CPPI)用于PAD诊断和严重程度分类的新概念。具体来说,我们进行了一项模拟试验,使用RGB摄像机记录20名受试者的足底视频,并在左腿上施加不同压力的袖带来模拟不同程度的下肢阻塞。我们使用远程光容积脉搏波成像生成足底灌注图,并提出了一个多视图灌注(MVP)特征集来表示灌注图,用于PAD分类。实验结果表明,MVP与多普勒超声(临床参考)特征的Pearson相关系数均大于0.9。MVP特征结合支持向量机对正常和阻塞状态的区分准确率为91.47%,对四种不同程度血管阻塞的区分准确率为76.48%。临床基准证明了CPPI作为一种快速、敏感和易于使用的PAD诊断工具的潜力,适合在家庭或社区环境中进行大规模筛查。
{"title":"Plantar Perfusion Imaging for Peripheral Arterial Disease Screening: A Proof-of-Concept Study.","authors":"Yukai Huang, Ningbo Zhao, Dongmin Huang, Yonglong Ye, Zi Luo, Hongzhou Lu, Min He, Wenjin Wang","doi":"10.1109/JBHI.2025.3591844","DOIUrl":"10.1109/JBHI.2025.3591844","url":null,"abstract":"<p><p>The diagnosis of peripheral artery disease (PAD) typically relies on specialized equipment such as ultrasound. The delayed PAD detection of these approaches may lead to amputation and even death. To achieve rapid and ubiquitous PAD screening, we propose a novel concept of camera-based plantar perfusion imaging (CPPI) for PAD diagnosis and severity classification. Specifically, we performed a simulation trial that used an RGB camera to record the plantar video of 20 subjects and a cuff with different pressures applied to the left leg to simulate different degrees of lower limb blockage. We generated the plantar perfusion maps using remote photoplethysmography imaging and proposed a multi-view perfusion (MVP) feature set to represent the perfusion maps for PAD classification. The experimental results show that the Pearson correlation coefficients between MVP and Doppler ultrasound (clinical reference) features were larger than 0.9. MVP feature combined with Support Vector Machine obtains 91.47% accuracy in distinguishing the normal and obstructed states, and 76.48% accuracy in differentiating four different degrees of vascular obstruction. The clinical benchmark demonstrated the potential of CPPI as a rapid, sensitive, and easy-to-use diagnostic tool for PAD, suitable for large-scale screening in home or community settings.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1520-1533"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144698428","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal of Biomedical and Health Informatics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1