首页 > 最新文献

Biomedical Signal Processing and Control最新文献

英文 中文
A cross-session non-stationary attention-based motor imagery classification method with critic-free domain adaptation 一种基于非稳态注意力的跨会期运动图像分类方法,具有无批判域适应性
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107122
Shuai Guo , Yi Wang , Xin Zhang , Baoping Tang
Recent studies increasingly employ deep learning to decode electroencephalogram (EEG) signals. While deep learning has improved the performance of motor imagery (MI) classification to some extent, challenges remain due to significant variances in EEG data across sessions and the limitations of convolutional neural networks (CNNs). EEG signals are inherently non-stationary, traditional multi-head attention typically uses normalization methods to reduce non-stationarity and improve performance. However, non-stationary factors are crucial inherent properties of EEG signals and provide valuable guidance for decoding temporal dependencies in EEG signals. In this paper, we introduce a novel CNN combined with the Non-stationary Attention (NSA) and Critic-free Domain Adaptation Network (NSDANet), tailored for decoding MI signals. This network starts with temporal–spatial convolution devised to extract spatial–temporal features from EEG signals. It then obtains multi-modal information from average and variance perspectives. We devise a new self-attention module, the Non-stationary Attention (NSA), to capture the non-stationary temporal dependencies of MI-EEG signals. Furthermore, to align feature distributions between the source and target domains, we propose a critic-free domain adaptation network that uses the Nuclear-norm Wasserstein discrepancy (NWD) to minimize the inter-domain differences. NWD complements the original classifier by acting as a critic without a gradient penalty. This integration leverages discriminative information for feature alignment, thus enhancing EEG decoding performance. We conducted extensive cross-session experiments on both BCIC IV 2a and BCIC IV 2b dataset. Results demonstrate that the proposed method outperforms some existing approaches.
最近的研究越来越多地采用深度学习来解码脑电图(EEG)信号。虽然深度学习在一定程度上提高了运动图像(MI)分类的性能,但由于不同时段的脑电图数据存在显著差异以及卷积神经网络(CNN)的局限性,挑战依然存在。脑电信号本身具有非平稳性,传统的多头注意力通常使用归一化方法来减少非平稳性并提高性能。然而,非稳态因素是脑电信号的重要固有属性,可为解码脑电信号中的时间依赖性提供有价值的指导。在本文中,我们介绍了一种结合了非稳态注意(NSA)和无批判域自适应网络(NSDANet)的新型 CNN,专门用于解码 MI 信号。该网络从时间-空间卷积开始,旨在从脑电图信号中提取空间-时间特征。然后,它从平均值和方差角度获取多模态信息。我们设计了一个新的自我注意模块--非稳态注意(NSA),以捕捉 MI-EEG 信号的非稳态时间依赖性。此外,为了调整源域和目标域之间的特征分布,我们提出了一种无批评域适应网络,它使用核正态分布差异(Nuclear-norm Wasserstein discrepancy,NWD)来最小化域间差异。NWD 作为无梯度惩罚的批判者,对原始分类器进行了补充。这种整合利用了特征对齐的判别信息,从而提高了脑电图解码性能。我们在 BCIC IV 2a 和 BCIC IV 2b 数据集上进行了广泛的跨时段实验。结果表明,所提出的方法优于一些现有方法。
{"title":"A cross-session non-stationary attention-based motor imagery classification method with critic-free domain adaptation","authors":"Shuai Guo ,&nbsp;Yi Wang ,&nbsp;Xin Zhang ,&nbsp;Baoping Tang","doi":"10.1016/j.bspc.2024.107122","DOIUrl":"10.1016/j.bspc.2024.107122","url":null,"abstract":"<div><div>Recent studies increasingly employ deep learning to decode electroencephalogram (EEG) signals. While deep learning has improved the performance of motor imagery (MI) classification to some extent, challenges remain due to significant variances in EEG data across sessions and the limitations of convolutional neural networks (CNNs). EEG signals are inherently non-stationary, traditional multi-head attention typically uses normalization methods to reduce non-stationarity and improve performance. However, non-stationary factors are crucial inherent properties of EEG signals and provide valuable guidance for decoding temporal dependencies in EEG signals. In this paper, we introduce a novel CNN combined with the Non-stationary Attention (NSA) and Critic-free Domain Adaptation Network (NSDANet), tailored for decoding MI signals. This network starts with temporal–spatial convolution devised to extract spatial–temporal features from EEG signals. It then obtains multi-modal information from average and variance perspectives. We devise a new self-attention module, the Non-stationary Attention (NSA), to capture the non-stationary temporal dependencies of MI-EEG signals. Furthermore, to align feature distributions between the source and target domains, we propose a critic-free domain adaptation network that uses the Nuclear-norm Wasserstein discrepancy (NWD) to minimize the inter-domain differences. NWD complements the original classifier by acting as a critic without a gradient penalty. This integration leverages discriminative information for feature alignment, thus enhancing EEG decoding performance. We conducted extensive cross-session experiments on both BCIC IV 2a and BCIC IV 2b dataset. Results demonstrate that the proposed method outperforms some existing approaches.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107122"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A deep learning based approach for automatic cardiac events identification 基于深度学习的心脏事件自动识别方法
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107164
Yuanshu Li , Kexin Hong , Xiaohu Shi , Wei Pang , Yubin Xiao , Peng Zhao , Dong Xu , Chunli Song , Xu Zhou , You Zhou
Visually identifying End-Diastolic (ED) and End-Systolic (ES) frames from 2D echocardiographic videos without electrocardiogram is time-consuming but a fundamental step in routine clinical practice for assessment of cardiac structure and functionality. In recent years, several algorithms such as segmentation-based methods and regression-based methods, have been proposed to automatically identify ED and ES frames for automated cardiac function assessment. However, these methods require high quality images or large datasets which are difficult to obtain. In this work, we propose the first classification-based method, which is combined with a three-step postprocess for ED and ES identification from 2-D echocardiographic videos. We explored a deep-learning based model for this task which has lightweight structure with 175 kb parameters (in hdf5 format) and has no large demand for the number of echocardiographic videos compared with regression-based methods. In particular, we propose a weights-shared convolutional neural network module as the backbone, which is for two adjacent echocardiographic video frames feature extraction; and the backbone is combined with a classification module, which predicts the two frames’ relationship for volume-time prediction. Based on this, ED and ES can be effectively identified. We trained and evaluated the proposed method on apical 4 chamber views and apical 2 chamber views. For both cardiac views, the accuracies are above 0.95 and the AUCs (area under the ROC) are above 0.99. For ultimate ED and ES prediction, all the precisions are above 97% and all average Frame Distances(aFDs) are less than 2 frame errors. Moreover, we demonstrate that our method has considerable generalizability.
从不带心电图的二维超声心动图视频中直观地识别舒张末期(ED)和收缩末期(ES)图像虽然费时,但却是常规临床实践中评估心脏结构和功能的基本步骤。近年来,人们提出了几种算法,如基于分割的方法和基于回归的方法,用于自动识别 ED 和 ES 帧,以进行自动心脏功能评估。然而,这些方法需要高质量的图像或大型数据集,而这些都很难获得。在这项工作中,我们提出了第一种基于分类的方法,该方法与三步后处理相结合,用于从二维超声心动图视频中识别 ED 和 ES。我们为这项任务探索了一种基于深度学习的模型,该模型结构轻巧,参数为 175 kb(hdf5 格式),与基于回归的方法相比,对超声心动图视频的数量要求不高。具体而言,我们提出了一个权值共享卷积神经网络模块作为骨干,用于相邻两帧超声心动图视频的特征提取;该骨干模块与一个分类模块相结合,预测两帧视频的关系,从而进行容积-时间预测。在此基础上,可有效识别 ED 和 ES。我们在心尖 4 腔视图和心尖 2 腔视图上对所提出的方法进行了训练和评估。两种心脏视图的准确率均高于 0.95,AUC(ROC 下面积)均高于 0.99。对于最终的 ED 和 ES 预测,精确度均在 97% 以上,平均帧距(aFD)均小于 2 帧误差。此外,我们还证明了我们的方法具有相当大的通用性。
{"title":"A deep learning based approach for automatic cardiac events identification","authors":"Yuanshu Li ,&nbsp;Kexin Hong ,&nbsp;Xiaohu Shi ,&nbsp;Wei Pang ,&nbsp;Yubin Xiao ,&nbsp;Peng Zhao ,&nbsp;Dong Xu ,&nbsp;Chunli Song ,&nbsp;Xu Zhou ,&nbsp;You Zhou","doi":"10.1016/j.bspc.2024.107164","DOIUrl":"10.1016/j.bspc.2024.107164","url":null,"abstract":"<div><div>Visually identifying End-Diastolic (ED) and End-Systolic (ES) frames from 2D echocardiographic videos without electrocardiogram is time-consuming but a fundamental step in routine clinical practice for assessment of cardiac structure and functionality. In recent years, several algorithms such as segmentation-based methods and regression-based methods, have been proposed to automatically identify ED and ES frames for automated cardiac function assessment. However, these methods require high quality images or large datasets which are difficult to obtain. In this work, we propose the first classification-based method, which is combined with a three-step postprocess for ED and ES identification from 2-D echocardiographic videos. We explored a deep-learning based model for this task which has lightweight structure with 175 kb parameters (in hdf5 format) and has no large demand for the number of echocardiographic videos compared with regression-based methods. In particular, we propose a weights-shared convolutional neural network module as the backbone, which is for two adjacent echocardiographic video frames feature extraction; and the backbone is combined with a classification module, which predicts the two frames’ relationship for volume-time prediction. Based on this, ED and ES can be effectively identified. We trained and evaluated the proposed method on apical 4 chamber views and apical 2 chamber views. For both cardiac views, the accuracies are above 0.95 and the AUCs (area under the ROC) are above 0.99. For ultimate ED and ES prediction, all the precisions are above 97% and all average Frame Distances(aFDs) are less than 2 frame errors. Moreover, we demonstrate that our method has considerable generalizability.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107164"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142657666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-supervised spatial-temporal calibration and semantic refinement network for video polyp segmentation 用于视频息肉分割的半监督时空校准和语义细化网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107127
Feng Li , Zetao Huang , Lu Zhou , Haixia Peng , Yimin Chu
<div><div>Automated video polyp segmentation (VPS) was of vitality for the early prevention and diagnosis of colorectal cancer (CRC). However, existing deep learning-based automatic polyp segmentation methods mainly focused on independent static images and struggled to perform well due to neglecting spatial–temporal relationships among successive video frames, while requiring massive frame-by-frame annotations. To better alleviate these challenges, we proposed a novel semi-supervised spatial–temporal calibration and semantic refinement network (STCSR-Net) dedicated to VPS, which simultaneously considered both the inter-frame temporal consistency in video clips and intra-frame semantic-spatial information. It was composed of a segmentation pathway and a propagation pathway by use of a co-training scheme for supervising the predictions on un-annotated images in a semi-supervised learning fashion. Specifically, we proposed an adaptive sequence calibration (ASC) block in segmentation pathway and a dynamic transmission calibration (DTC) block in propagation pathway to fully take advantage of valuable temporal cues and maintain the prediction temporally consistent among consecutive frames. Meanwhile, in these two branches, we introduced residual block (RB) to suppress irrelevant noisy information and highlight rich local boundary details of polyp lesions, while constructed multi-scale context extraction (MCE) module to enhance multi-scale high-level semantic feature expression. On that basis, we designed progressive adaptive context fusion (PACF) module to gradually aggregate multi-level features under the guidance of reinforced high-level semantic information for eliminating semantic gaps among them and promoting the discrimination capacity of features for targeting polyp objects. Through synergistic combination of RB, MCE and PACF modules, semantic-spatial correlations on polyp lesions within each frame could be established. Coupled with the context-free loss, our model merged feature representations of neighboring frames to diminish the dependency on varying contexts within consecutive frames and strengthen its robustness. Extensive experiments substantiated that our model with 100% annotation ratio achieved state-of-the-art performance on challenging datasets. Even trained under 50% annotation ratio, our model exceled significantly existing state-of-the-art image-based and video-based polyp segmentation models on the newly-built local TRPolyp dataset by at least 1.3% and 0.9% enhancements in both mDice and mIoU, whilst exhibited comparable performance to top rivals attained through using fully supervised approach on publicly available CVC-612, CVC-300 and ASU-Mayo-Clinic benchmarks. Notably, our model showcased exceptionally well in videos containing complex scenarios like motion blur and occlusion. Beyond that, it also harvested approximately 0.794 mDice and 0.707 mIoU at an inference of 0.036 s per frame in endoscopist-machine competition, which
自动视频息肉分割(VPS)对于结直肠癌(CRC)的早期预防和诊断至关重要。然而,现有的基于深度学习的息肉自动分割方法主要针对独立的静态图像,由于忽略了连续视频帧之间的时空关系,同时需要大量的逐帧注释,因此难以取得良好的效果。为了更好地应对这些挑战,我们提出了一种新型的半监督时空校准和语义细化网络(STCSR-Net),专门用于 VPS,它同时考虑了视频片段帧间的时空一致性和帧内的语义空间信息。该网络由分割路径和传播路径组成,采用共同训练方案,以半监督学习的方式对未标注图像的预测进行监督。具体来说,我们在分割路径中提出了自适应序列校准(ASC)块,在传播路径中提出了动态传输校准(DTC)块,以充分利用有价值的时间线索,保持连续帧间预测的时间一致性。同时,在这两个分支中,我们引入了残差块(RB)来抑制无关的噪声信息,突出息肉病变丰富的局部边界细节,并构建了多尺度上下文提取(MCE)模块来增强多尺度高层次语义特征的表达。在此基础上,我们设计了渐进式自适应上下文融合(PACF)模块,在强化的高层次语义信息指导下逐步聚合多层次特征,消除特征间的语义空白,提高特征对息肉对象的识别能力。通过 RB、MCE 和 PACF 模块的协同组合,可以建立每帧内息肉病变的语义空间相关性。结合无上下文损失,我们的模型合并了相邻帧的特征表征,从而降低了对连续帧内不同上下文的依赖性,增强了模型的鲁棒性。广泛的实验证明,在具有挑战性的数据集上,我们的 100%注释率模型取得了最先进的性能。即使在注释率为 50% 的情况下进行训练,我们的模型在新建立的本地 TRPolyp 数据集上也明显优于现有的一流图像息肉分割模型和视频息肉分割模型,在 mDice 和 mIoU 两项指标上分别提高了至少 1.3% 和 0.9%,同时在公开的 CVC-612、CVC-300 和 ASU-Mayo-Clinic 基准上,我们的模型与采用完全监督方法的顶尖对手表现相当。值得注意的是,我们的模型在包含运动模糊和闭塞等复杂场景的视频中表现出了卓越的性能。此外,它还在内窥镜医师-机器竞赛中以每帧 0.036 秒的推理时间收获了约 0.794 mDice 和 0.707 mIoU,表现优于初级和高级内窥镜医师,几乎与专家级内窥镜医师不相上下。所提出的 STCSR-Net 的强大能力有望提高 VPS 的质量,突出了模型在真实世界临床场景中的适应性和潜力。
{"title":"Semi-supervised spatial-temporal calibration and semantic refinement network for video polyp segmentation","authors":"Feng Li ,&nbsp;Zetao Huang ,&nbsp;Lu Zhou ,&nbsp;Haixia Peng ,&nbsp;Yimin Chu","doi":"10.1016/j.bspc.2024.107127","DOIUrl":"10.1016/j.bspc.2024.107127","url":null,"abstract":"&lt;div&gt;&lt;div&gt;Automated video polyp segmentation (VPS) was of vitality for the early prevention and diagnosis of colorectal cancer (CRC). However, existing deep learning-based automatic polyp segmentation methods mainly focused on independent static images and struggled to perform well due to neglecting spatial–temporal relationships among successive video frames, while requiring massive frame-by-frame annotations. To better alleviate these challenges, we proposed a novel semi-supervised spatial–temporal calibration and semantic refinement network (STCSR-Net) dedicated to VPS, which simultaneously considered both the inter-frame temporal consistency in video clips and intra-frame semantic-spatial information. It was composed of a segmentation pathway and a propagation pathway by use of a co-training scheme for supervising the predictions on un-annotated images in a semi-supervised learning fashion. Specifically, we proposed an adaptive sequence calibration (ASC) block in segmentation pathway and a dynamic transmission calibration (DTC) block in propagation pathway to fully take advantage of valuable temporal cues and maintain the prediction temporally consistent among consecutive frames. Meanwhile, in these two branches, we introduced residual block (RB) to suppress irrelevant noisy information and highlight rich local boundary details of polyp lesions, while constructed multi-scale context extraction (MCE) module to enhance multi-scale high-level semantic feature expression. On that basis, we designed progressive adaptive context fusion (PACF) module to gradually aggregate multi-level features under the guidance of reinforced high-level semantic information for eliminating semantic gaps among them and promoting the discrimination capacity of features for targeting polyp objects. Through synergistic combination of RB, MCE and PACF modules, semantic-spatial correlations on polyp lesions within each frame could be established. Coupled with the context-free loss, our model merged feature representations of neighboring frames to diminish the dependency on varying contexts within consecutive frames and strengthen its robustness. Extensive experiments substantiated that our model with 100% annotation ratio achieved state-of-the-art performance on challenging datasets. Even trained under 50% annotation ratio, our model exceled significantly existing state-of-the-art image-based and video-based polyp segmentation models on the newly-built local TRPolyp dataset by at least 1.3% and 0.9% enhancements in both mDice and mIoU, whilst exhibited comparable performance to top rivals attained through using fully supervised approach on publicly available CVC-612, CVC-300 and ASU-Mayo-Clinic benchmarks. Notably, our model showcased exceptionally well in videos containing complex scenarios like motion blur and occlusion. Beyond that, it also harvested approximately 0.794 mDice and 0.707 mIoU at an inference of 0.036 s per frame in endoscopist-machine competition, which ","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107127"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Explainable AI-based method for brain abnormality diagnostics using MRI 利用核磁共振成像诊断大脑异常的可解释人工智能方法
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107184
Mohamed Hosny , Ahmed M. Elshenhab , Ahmed Maged
Detecting brain abnormalities using magnetic resonance imaging (MRI) is a vital frontier in neurological research. Therefore, accurate methods are essential for guiding neurologists in diagnosing enigmatic disorders such as Alzheimer’s disease (AD) and brain tumors. These methods aid in the early detection and treatment of these formidable conditions. However, traditional techniques often suffer from high computational complexity and efficiency. Additionally, existing detection models lack the ability to explain their predictions, rendering them untrustworthy for clinicians. This study presents an explainable framework for automatic brain abnormality detection in MRI images. The methodology includes a robust preprocessing pipeline that ameliorates image relevance through image thresholding, morphological operations and adaptive edge detection using the AutoCanny algorithm. AutoCanny method automatically adjusts thresholds to ensure effective edge detection across different images. Then, the MRI images are fed to efficient vision transformer model (EfficientViT) for classification. EfficientViT features a memory-efficient sandwich layout, cascaded group attention module and optimized parameter reallocation. These innovations collectively enhanced the model efficiency in terms of memory usage, computational complexity and parameter optimization. Moreover, gradient-based Shapley additive explanations is employed to explain the EfficientViT model predictions. EfficientViT achieved the highest accuracy of 99.24%, 97.1%, 99.5% and 98.87% on the AD, Tumor1, Tumor2 and merged datasets, respectively. Furthermore, the proposed model outperformed longstanding deep learning techniques. These findings have significant implications for uncovering hidden information associated with brain abnormality as well as improving the diagnostic process and treatment planning. Our model can aid neurologists in the validation of manual MRI neurological disorders screenings.
利用磁共振成像(MRI)检测大脑异常是神经学研究的一个重要前沿领域。因此,准确的方法对于指导神经学家诊断阿尔茨海默病(AD)和脑肿瘤等神秘疾病至关重要。这些方法有助于早期发现和治疗这些可怕的疾病。然而,传统技术往往存在计算复杂度高、效率低的问题。此外,现有的检测模型缺乏解释其预测结果的能力,因此无法为临床医生所信任。本研究提出了一个可解释的框架,用于自动检测核磁共振成像图像中的大脑异常。该方法包括一个强大的预处理管道,通过图像阈值、形态学操作和使用 AutoCanny 算法的自适应边缘检测来改善图像相关性。AutoCanny 方法会自动调整阈值,以确保在不同图像中进行有效的边缘检测。然后,将核磁共振图像输入高效视觉转换器模型(EfficientViT)进行分类。EfficientViT 采用了节省内存的三明治布局、级联群注意模块和优化的参数重新分配。这些创新共同提高了模型在内存使用、计算复杂度和参数优化方面的效率。此外,EfficientViT 还采用了基于梯度的夏普利加法解释来解释模型预测。在AD、Tumor1、Tumor2和合并数据集上,EfficientViT分别达到了99.24%、97.1%、99.5%和98.87%的最高准确率。此外,所提出的模型还优于长期使用的深度学习技术。这些发现对于揭示与大脑异常相关的隐藏信息以及改善诊断过程和治疗计划具有重要意义。我们的模型可以帮助神经科医生验证手动磁共振成像神经系统疾病筛查。
{"title":"Explainable AI-based method for brain abnormality diagnostics using MRI","authors":"Mohamed Hosny ,&nbsp;Ahmed M. Elshenhab ,&nbsp;Ahmed Maged","doi":"10.1016/j.bspc.2024.107184","DOIUrl":"10.1016/j.bspc.2024.107184","url":null,"abstract":"<div><div>Detecting brain abnormalities using magnetic resonance imaging (MRI) is a vital frontier in neurological research. Therefore, accurate methods are essential for guiding neurologists in diagnosing enigmatic disorders such as Alzheimer’s disease (AD) and brain tumors. These methods aid in the early detection and treatment of these formidable conditions. However, traditional techniques often suffer from high computational complexity and efficiency. Additionally, existing detection models lack the ability to explain their predictions, rendering them untrustworthy for clinicians. This study presents an explainable framework for automatic brain abnormality detection in MRI images. The methodology includes a robust preprocessing pipeline that ameliorates image relevance through image thresholding, morphological operations and adaptive edge detection using the AutoCanny algorithm. AutoCanny method automatically adjusts thresholds to ensure effective edge detection across different images. Then, the MRI images are fed to efficient vision transformer model (EfficientViT) for classification. EfficientViT features a memory-efficient sandwich layout, cascaded group attention module and optimized parameter reallocation. These innovations collectively enhanced the model efficiency in terms of memory usage, computational complexity and parameter optimization. Moreover, gradient-based Shapley additive explanations is employed to explain the EfficientViT model predictions. EfficientViT achieved the highest accuracy of 99.24%, 97.1%, 99.5% and 98.87% on the AD, Tumor1, Tumor2 and merged datasets, respectively. Furthermore, the proposed model outperformed longstanding deep learning techniques. These findings have significant implications for uncovering hidden information associated with brain abnormality as well as improving the diagnostic process and treatment planning. Our model can aid neurologists in the validation of manual MRI neurological disorders screenings.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107184"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A deep temporal network for motor imagery classification based on multi-branch feature fusion and attention mechanism 基于多分支特征融合和注意力机制的运动图像分类深度时态网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107163
Jinke Zhao, Mingliang Liu

Objective:

In recent years, the Brain-Computer Interface (BCI) technology has witnessed rapid advancements. Motor Imagery (MI), as one of the BCI paradigms, has found extensive applications in domains such as rehabilitation, entertainment, and neuroscience. How to conduct effective classification of it has emerged as one of the primary research issues. Electroencephalography (EEG) serves as an essential tool for studying the classification of MI. However, the existing models are incapable of fully extracting effective motion information from the interfered electroencephalogram data, leading to the final classification effect falling short of the expected goals. In response to this problem, we propose a deep temporal network based on multi-branch feature fusion and attention mechanism. This network incorporates a combination of multi-branch feature fusion, feature expansion, attention, and temporal decoding modules.

Methods:

First, primary features of EEG signals are extracted using a multi-branch convolutional neural network, followed by feature fusion. Subsequently, feature augmentation and attention mechanisms are employed to reduce noise interference while highlighting essential MI intentions. Finally, a temporal decoding module is utilized to deeply explore temporal information in MI data and perform classification.

Results:

The model performance was tested on the BCI_IV_2a, BCI_IV_2b, and OPenBMI datasets using both subject-specific and subject-independent experimental methods. The model achieved significant performance improvements on all three datasets, achieving accuracy of 81.21%, 93.12%, and 75.9%, respectively, better than other baseline models.

Conclusion:

Experimental results indicate that the proposed model leverages deep learning techniques for the classification of different types of MI, providing a reference framework for the development of more efficient MI-BCI systems.
目的:近年来,脑机接口(BCI)技术突飞猛进。运动想象(MI)作为BCI范式之一,已在康复、娱乐和神经科学等领域得到广泛应用。如何对其进行有效分类已成为首要研究课题之一。脑电图(EEG)是研究 MI 分类的重要工具。然而,现有的模型无法从受干扰的脑电图数据中充分提取有效的运动信息,导致最终的分类效果达不到预期目标。针对这一问题,我们提出了一种基于多分支特征融合和注意力机制的深度时态网络。方法:首先,使用多分支卷积神经网络提取脑电信号的主要特征,然后进行特征融合。随后,采用特征扩展和注意力机制来减少噪音干扰,同时突出重要的 MI 意图。最后,利用时序解码模块深入挖掘 MI 数据中的时序信息并进行分类。结果:使用特定受试者和独立于受试者的实验方法,在 BCI_IV_2a、BCI_IV_2b 和 OPenBMI 数据集上测试了该模型的性能。实验结果表明,该模型利用深度学习技术对不同类型的心肌梗塞进行了分类,为开发更高效的心肌梗塞-脑梗塞系统提供了参考框架。
{"title":"A deep temporal network for motor imagery classification based on multi-branch feature fusion and attention mechanism","authors":"Jinke Zhao,&nbsp;Mingliang Liu","doi":"10.1016/j.bspc.2024.107163","DOIUrl":"10.1016/j.bspc.2024.107163","url":null,"abstract":"<div><h3>Objective:</h3><div>In recent years, the Brain-Computer Interface (BCI) technology has witnessed rapid advancements. Motor Imagery (MI), as one of the BCI paradigms, has found extensive applications in domains such as rehabilitation, entertainment, and neuroscience. How to conduct effective classification of it has emerged as one of the primary research issues. Electroencephalography (EEG) serves as an essential tool for studying the classification of MI. However, the existing models are incapable of fully extracting effective motion information from the interfered electroencephalogram data, leading to the final classification effect falling short of the expected goals. In response to this problem, we propose a deep temporal network based on multi-branch feature fusion and attention mechanism. This network incorporates a combination of multi-branch feature fusion, feature expansion, attention, and temporal decoding modules.</div></div><div><h3>Methods:</h3><div>First, primary features of EEG signals are extracted using a multi-branch convolutional neural network, followed by feature fusion. Subsequently, feature augmentation and attention mechanisms are employed to reduce noise interference while highlighting essential MI intentions. Finally, a temporal decoding module is utilized to deeply explore temporal information in MI data and perform classification.</div></div><div><h3>Results:</h3><div>The model performance was tested on the BCI_IV_2a, BCI_IV_2b, and OPenBMI datasets using both subject-specific and subject-independent experimental methods. The model achieved significant performance improvements on all three datasets, achieving accuracy of 81.21%, 93.12%, and 75.9%, respectively, better than other baseline models.</div></div><div><h3>Conclusion:</h3><div>Experimental results indicate that the proposed model leverages deep learning techniques for the classification of different types of MI, providing a reference framework for the development of more efficient MI-BCI systems.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107163"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658272","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
2D/3D fast fine registration in minimally invasive pelvic surgery 微创盆腔手术中的 2D/3D 快速精细配准
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107145
Fujiao Ju , Yuan Li , Jingxin Zhao , Mingjie Dong
The 2D/3D rigid registration between preoperative 3D CT and intraoperative 2D X-ray is a crucial step in minimally invasive pelvic surgery. The deep learning-based 2D/3D registration methods address the inefficiencies of traditional approaches. However, the wide range of spatial transformation parameters and other complexities pose significant challenges for achieving accurate registration in a single step. Additionally, the stylistic differences between Digitally Reconstructed Radiographs (DRRs) used in training and real X-ray images limit the practical applicability of most methods. To overcome these challenges, we propose a 2D/3D fast registration framework comprising a coarse registration network, fine registration based on key point tracking and alignment, and domain adaptation. Coarse registration using plug-and-play attention is introduced to preliminarily estimate transformation parameters. Then we design a key point tracking network to match key points between different images, and leverage points alignment to achieve fine registration. To address the stylistic differences between DRR and X-ray images, we investigate a domain adaptation network. The experiments were conducted on DRR and X-ray images, respectively. Our method achieved a mean absolute error of 0.58 on DRR and a structural similarity of 78% on X-ray, outperforming baseline methods. Extensive ablation studies demonstrate that fine registration and domain adaptation significantly improve registration performance.
术前三维 CT 和术中二维 X 光片之间的二维/三维刚性配准是微创盆腔手术的关键步骤。基于深度学习的 2D/3D 配准方法解决了传统方法效率低下的问题。然而,广泛的空间变换参数和其他复杂性对一步实现精确配准构成了巨大挑战。此外,用于训练的数字重建放射照片(DRR)与真实 X 光图像之间的风格差异也限制了大多数方法的实际适用性。为了克服这些挑战,我们提出了一种 2D/3D 快速配准框架,包括粗配准网络、基于关键点跟踪和配准的精配准以及域适应。我们首先介绍了使用即插即用注意力进行粗配准的方法,以初步估计变换参数。然后,我们设计了一个关键点跟踪网络来匹配不同图像之间的关键点,并利用点对齐来实现精细配准。针对 DRR 和 X 光图像之间的风格差异,我们研究了一种域自适应网络。实验分别在 DRR 和 X 光图像上进行。我们的方法在 DRR 图像上的平均绝对误差为 0.58,在 X 光图像上的结构相似度为 78%,优于基线方法。广泛的消融研究表明,精细配准和域自适应显著提高了配准性能。
{"title":"2D/3D fast fine registration in minimally invasive pelvic surgery","authors":"Fujiao Ju ,&nbsp;Yuan Li ,&nbsp;Jingxin Zhao ,&nbsp;Mingjie Dong","doi":"10.1016/j.bspc.2024.107145","DOIUrl":"10.1016/j.bspc.2024.107145","url":null,"abstract":"<div><div>The 2D/3D rigid registration between preoperative 3D CT and intraoperative 2D X-ray is a crucial step in minimally invasive pelvic surgery. The deep learning-based 2D/3D registration methods address the inefficiencies of traditional approaches. However, the wide range of spatial transformation parameters and other complexities pose significant challenges for achieving accurate registration in a single step. Additionally, the stylistic differences between Digitally Reconstructed Radiographs (DRRs) used in training and real X-ray images limit the practical applicability of most methods. To overcome these challenges, we propose a 2D/3D fast registration framework comprising a coarse registration network, fine registration based on key point tracking and alignment, and domain adaptation. Coarse registration using plug-and-play attention is introduced to preliminarily estimate transformation parameters. Then we design a key point tracking network to match key points between different images, and leverage points alignment to achieve fine registration. To address the stylistic differences between DRR and X-ray images, we investigate a domain adaptation network. The experiments were conducted on DRR and X-ray images, respectively. Our method achieved a mean absolute error of 0.58 on DRR and a structural similarity of 78% on X-ray, outperforming baseline methods. Extensive ablation studies demonstrate that fine registration and domain adaptation significantly improve registration performance.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107145"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142652847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
2MSPK-Net: A nuclei segmentation network based on multi-scale, multi-dimensional attention, and SAM prior knowledge 2MSPK-Net:基于多尺度、多维度注意力和 SAM 先验知识的细胞核分割网络
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-13 DOI: 10.1016/j.bspc.2024.107140
Gongtao Yue , Xiaoguang Ma , Wenrui Li , Ziheng An , Chen Yang
Refined nuclei segmentation is of great significance for diagnosing the pathological conditions of tumor tissues. Although existing encoder–decoder networks have achieved remarkable progress in nuclei segmentation tasks, practical applications still encounter obstacles, especially for challenging issues such as highly dense nuclei targets and the ambiguity of boundaries between inter-class features, resulting in unsatisfactory segmentation accuracy. In this work, a novel encoder–decoder architecture was proposed to address these issues. Specifically, we first proposed a multi-scale and multi-dimension attention module to capture the contextual dependencies between individual pixels and the overall pixels, where in cross-scale learning was achieved by fusing different scale feature information of the encoding layer. Secondly, we integrated the prior knowledge of SAM into nuclei images to enhance the network’s ability to distinguish fuzzy features. To the best of our knowledge, this was the first attempt to utilize the prior knowledge of SAM to optimize nuclei segmentation tasks. Furthermore, the network was guided to supplement missing detailed features through a reverse erasing strategy and cross-layer information flow. Comprehensive experiments illustrated that the proposed method achieved MIoU improvements of 1.26% and 0.94% on the MoNuSeg and TNBC datasets, respectively, over several SOTA methods, indicating its great potential as a backbone for cancer nuclei segmentation. Code: https://github.com/ThirteenYue/2MSPK-Net.
精细的细胞核分割对诊断肿瘤组织的病理状况具有重要意义。虽然现有的编码器-解码器网络在核仁分割任务中取得了显著进展,但在实际应用中仍会遇到障碍,尤其是在核仁目标高度密集、类间特征边界模糊等挑战性问题上,导致分割精度不尽人意。在这项工作中,我们提出了一种新型编码器-解码器架构来解决这些问题。具体来说,我们首先提出了一个多尺度和多维度关注模块,以捕捉单个像素和整体像素之间的上下文依赖关系,其中跨尺度学习是通过融合编码层的不同尺度特征信息来实现的。其次,我们将 SAM 的先验知识整合到核图像中,以增强网络分辨模糊特征的能力。据我们所知,这是首次尝试利用 SAM 的先验知识来优化核仁分割任务。此外,还通过反向擦除策略和跨层信息流引导网络补充缺失的细节特征。综合实验表明,在 MoNuSeg 和 TNBC 数据集上,与几种 SOTA 方法相比,所提方法的 MIoU 分别提高了 1.26% 和 0.94%,这表明它作为癌症细胞核分割骨干的巨大潜力。代码:https://github.com/ThirteenYue/2MSPK-Net。
{"title":"2MSPK-Net: A nuclei segmentation network based on multi-scale, multi-dimensional attention, and SAM prior knowledge","authors":"Gongtao Yue ,&nbsp;Xiaoguang Ma ,&nbsp;Wenrui Li ,&nbsp;Ziheng An ,&nbsp;Chen Yang","doi":"10.1016/j.bspc.2024.107140","DOIUrl":"10.1016/j.bspc.2024.107140","url":null,"abstract":"<div><div>Refined nuclei segmentation is of great significance for diagnosing the pathological conditions of tumor tissues. Although existing encoder–decoder networks have achieved remarkable progress in nuclei segmentation tasks, practical applications still encounter obstacles, especially for challenging issues such as highly dense nuclei targets and the ambiguity of boundaries between inter-class features, resulting in unsatisfactory segmentation accuracy. In this work, a novel encoder–decoder architecture was proposed to address these issues. Specifically, we first proposed a multi-scale and multi-dimension attention module to capture the contextual dependencies between individual pixels and the overall pixels, where in cross-scale learning was achieved by fusing different scale feature information of the encoding layer. Secondly, we integrated the prior knowledge of SAM into nuclei images to enhance the network’s ability to distinguish fuzzy features. To the best of our knowledge, this was the first attempt to utilize the prior knowledge of SAM to optimize nuclei segmentation tasks. Furthermore, the network was guided to supplement missing detailed features through a reverse erasing strategy and cross-layer information flow. Comprehensive experiments illustrated that the proposed method achieved MIoU improvements of 1.26% and 0.94% on the MoNuSeg and TNBC datasets, respectively, over several SOTA methods, indicating its great potential as a backbone for cancer nuclei segmentation. Code: <span><span>https://github.com/ThirteenYue/2MSPK-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107140"},"PeriodicalIF":4.9,"publicationDate":"2024-11-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142657663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extraction of apparent BOLD components in resting state fMRI signals by a novel method called “BOLD-filter” 用一种名为 "BOLD-过滤器 "的新方法提取静息状态 fMRI 信号中的表观 BOLD 成分
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-12 DOI: 10.1016/j.bspc.2024.107151
Yul-Wan Sung , Uk-Su Choi , Seiji Ogawa
Resting-state functional magnetic resonance imaging (rs-fMRI) is widely used for studying brain diseases and cognition. Unlike task-based fMRI (tb-fMRI), which employs a task protocol to elicit the BOLD response, the rs-fMRI signal is itself considered a BOLD response and is used to examine brain function related to brain diseases and cognition because explicitly identifying the BOLD component in the rs-fMRI signal is challenging. In this study, we propose a novel method called “BOLD-filter” which aims to extract apparent BOLD (aBOLD) components from the rs-fMRI signal before further processing for cognitive and clinical applications. We confirm that applying the BOLD-filter to real data enables us to identify the aBOLD components in the rs-fMRI signal. Under the sufficient condition for aBOLD, among voxels of the whole brain, 84 % exhibited aBOLD signals after applying BOLD-filter, compared to only 14 % before the BOLD-filter, with the BOLD-filter identifying aBOLD signals in six times as many voxels (p = 0.001). Additionally, we observed that utilizing these apparent BOLD components enhances the association of functional connectivity with age in our example. The potential of BOLD filter to ensure BOLD response in rs-fMRI signal analysis is expected to bring a new dimension to rs-fMRI studies in relation to the understanding of brain disorders and cognitive function.
静息态功能磁共振成像(rs-fMRI)被广泛用于研究脑部疾病和认知。与基于任务的 fMRI(tb-fMRI)不同,tb-fMRI 采用任务协议来激发 BOLD 反应,而 rs-fMRI 信号本身被认为是 BOLD 反应,用于研究与脑疾病和认知相关的脑功能,因为明确识别 rs-fMRI 信号中的 BOLD 部分具有挑战性。在本研究中,我们提出了一种名为 "BOLD-过滤器 "的新方法,旨在从 rs-fMRI 信号中提取表观 BOLD(aBOLD)成分,然后再进行进一步处理,以用于认知和临床应用。我们证实,将 BOLD 过滤器应用于真实数据,能够识别 rs-fMRI 信号中的 aBOLD 成分。在aBOLD充分条件下,应用BOLD滤波器后,全脑体素中84%显示出aBOLD信号,而应用BOLD滤波器前仅为14%,BOLD滤波器识别出的aBOLD信号是应用前的6倍(p = 0.001)。此外,我们还观察到,在我们的例子中,利用这些明显的 BOLD 成分增强了功能连通性与年龄的关联。BOLD 过滤器在 rs-fMRI 信号分析中确保 BOLD 响应的潜力有望为 rs-fMRI 研究带来一个新的维度,有助于理解大脑疾病和认知功能。
{"title":"Extraction of apparent BOLD components in resting state fMRI signals by a novel method called “BOLD-filter”","authors":"Yul-Wan Sung ,&nbsp;Uk-Su Choi ,&nbsp;Seiji Ogawa","doi":"10.1016/j.bspc.2024.107151","DOIUrl":"10.1016/j.bspc.2024.107151","url":null,"abstract":"<div><div>Resting-state functional magnetic resonance imaging (rs-fMRI) is widely used for studying brain diseases and cognition. Unlike task-based fMRI (tb-fMRI), which employs a task protocol to elicit the BOLD response, the rs-fMRI signal is itself considered a BOLD response and is used to examine brain function related to brain diseases and cognition because explicitly identifying the BOLD component in the rs-fMRI signal is challenging. In this study, we propose a novel method called “BOLD-filter” which aims to extract apparent BOLD (aBOLD) components from the rs-fMRI signal before further processing for cognitive and clinical applications. We confirm that applying the BOLD-filter to real data enables us to identify the aBOLD components in the rs-fMRI signal. Under the sufficient condition for aBOLD, among voxels of the whole brain, 84 % exhibited aBOLD signals after applying BOLD-filter, compared to only 14 % before the BOLD-filter, with the BOLD-filter identifying aBOLD signals in six times as many voxels (p = 0.001). Additionally, we observed that utilizing these apparent BOLD components enhances the association of functional connectivity with age in our example. The potential of BOLD filter to ensure BOLD response in rs-fMRI signal analysis is expected to bring a new dimension to rs-fMRI studies in relation to the understanding of brain disorders and cognitive function.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107151"},"PeriodicalIF":4.9,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An efficient data mining technique and privacy preservation model for healthcare data using improved darts game optimizer-based weighted deep neural network and hybrid encryption 利用改进的基于飞镖游戏优化器的加权深度神经网络和混合加密技术,为医疗保健数据建立高效的数据挖掘技术和隐私保护模型
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-12 DOI: 10.1016/j.bspc.2024.107168
D. Dhinakaran , L. Srinivasan , S. Gopalakrishnan , T.P. Anish
In recent days, association rule mining techniques have been widely used in healthcare data to provide accurate records that are important to ensure the data privacy. However, making this information public leads to creating attacks on them. In this paper, a secure privacy-preservation scheme for healthcare data is implemented to protect the security of the information for disease prediction in the current healthcare applications. The health data is collected from the benchmark datasets. Initially, the data is encrypted using Fully Homomorphic Encryption and Hyperelliptic Curve Cryptography (FHE-HECC) for the privacy preservation process. This model is developed by combining the Fully Homomorphic Encryption (FHE) and Hyperelliptic Curve Cryptography (HECC). For this encryption, the optimal key is generated using the Improved Darts Game Optimizer (IDGO) leveraging the Darts Game Optimizer (DGO). In the case of data decryption, the above-mentioned cryptography is utilized. The optimally selected key encrypts the data with high security without any breaches. The stored encrypted data is monitored and the disease is recognized using the Weighted Deep Neural Network (W-DNN) method and here, Deep Neural Network (DNN) acts as the fundamental model. Finally, the privacy of health data is preserved and the type of disease is detected by the implemented model. The suggested model attained accuracy of 92.83 which is higher than the existing techniques like GRU with 86.97, RNN with 88.28, LSTM with 91.92, WDNN with 90.10, respectively. The key findings of the suggested approach Proved that it facilitates effective treatment to the patient.
近年来,关联规则挖掘技术被广泛应用于医疗保健数据中,以提供对确保数据隐私非常重要的准确记录。然而,公开这些信息会导致对它们的攻击。本文提出了一种安全的医疗数据隐私保护方案,以保护当前医疗应用中疾病预测信息的安全。健康数据是从基准数据集中收集的。在隐私保护过程中,首先使用完全同态加密和超椭圆曲线加密(FHE-HECC)对数据进行加密。该模型是结合全同态加密(FHE)和超椭圆曲线加密(HECC)技术开发的。在加密过程中,利用改进的飞镖游戏优化器(IDGO)和飞镖游戏优化器(DGO)生成最佳密钥。在数据解密时,则使用上述加密技术。经过优化选择的密钥可对数据进行高安全性加密,不会出现任何漏洞。对存储的加密数据进行监控,并使用加权深度神经网络(W-DNN)方法识别疾病,在此,深度神经网络(DNN)充当基本模型。最后,健康数据的隐私得到了保护,疾病的类型也通过实施的模型检测出来。建议模型的准确率达到 92.83,分别高于现有技术,如 GRU(86.97)、RNN(88.28)、LSTM(91.92)和 WDNN(90.10)。建议方法的主要发现证明,它有助于对患者进行有效治疗。
{"title":"An efficient data mining technique and privacy preservation model for healthcare data using improved darts game optimizer-based weighted deep neural network and hybrid encryption","authors":"D. Dhinakaran ,&nbsp;L. Srinivasan ,&nbsp;S. Gopalakrishnan ,&nbsp;T.P. Anish","doi":"10.1016/j.bspc.2024.107168","DOIUrl":"10.1016/j.bspc.2024.107168","url":null,"abstract":"<div><div>In recent days, association rule mining techniques have been widely used in healthcare data to provide accurate records that are important to ensure the data privacy. However, making this information public leads to creating attacks on them. In this paper, a secure privacy-preservation scheme for healthcare data is implemented to protect the security of the information for disease prediction in the current healthcare applications. The health data is collected from the benchmark datasets. Initially, the data is encrypted using Fully Homomorphic Encryption and Hyperelliptic Curve Cryptography (FHE-HECC) for the privacy preservation process. This model is developed by combining the Fully Homomorphic Encryption (FHE) and Hyperelliptic Curve Cryptography (HECC). For this encryption, the optimal key is generated using the Improved Darts Game Optimizer (IDGO) leveraging the Darts Game Optimizer (DGO). In the case of data decryption, the above-mentioned cryptography is utilized. The optimally selected key encrypts the data with high security without any breaches. The stored encrypted data is monitored and the disease is recognized using the Weighted Deep Neural Network (W-DNN) method and here, Deep Neural Network (DNN) acts as the fundamental model. Finally, the privacy of health data is preserved and the type of disease is detected by the implemented model. The suggested model attained accuracy of 92.83 which is higher than the existing techniques like GRU with 86.97, RNN with 88.28, LSTM with 91.92, WDNN with 90.10, respectively. The key findings of the suggested approach Proved that it facilitates effective treatment to the patient.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107168"},"PeriodicalIF":4.9,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142658294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient retinal exudates detection method using ELNet in diabetic retinopathy assessment 利用 ELNet 在糖尿病视网膜病变评估中高效检测视网膜渗出物的方法
IF 4.9 2区 医学 Q1 ENGINEERING, BIOMEDICAL Pub Date : 2024-11-11 DOI: 10.1016/j.bspc.2024.107162
G. Sasi, A. Kaleel Rahuman
Diabetic Retinopathy (DR) can be detected at earlier stage by detecting exudates in retinal fundus images. In this article, the exudates are detected and segmented using the proposed Enhanced LeNet (ELNet) classification method. The proposed exudates segmentation method consists of the following modules Retinal image classification and Exudates segmentation. In retinal image classification, the retinal images are data augmented and then ELNet classification architecture is used to classify the retinal image into either normal or abnormal. In exudates segmentation module, the exudates are detected and segmented using Kirsch edge detector. The performance of the exudate’s detection method is improved by detecting and eliminating the blood vessels in the retinal image before detecting the exudates. In this paper, Digital Retinal Images for Vessel Extraction (DRIVE) and Diabetic Retinopathy database (DIARETDB1) retinal image datasets are used for the detection of exudates in the retinal images. In this study, the proposed method showcases remarkable results, demonstrating a sensitivity of 99.31% and 99.31%, specificity of 97.44% and 95%, and an accuracy of 99.09% and 98.8% for the DRIVE and DIARETDB1 datasets, respectively. Exudates detection in both datasets without eliminating OD and retinal blood vessels, we observe similar accuracy rates, Average 96.5% for both datasets. However, when eliminating OD and retinal blood vessels, the accuracy significantly improved for both datasets, reaching approximately 99.2% average. The performance is analyzed and compared with other state of the art methods.
通过检测视网膜眼底图像中的渗出物,可以在早期检测出糖尿病视网膜病变(DR)。本文采用所提出的增强 LeNet(ELNet)分类方法来检测和分割渗出物。所提出的渗出物分割方法包括以下模块 视网膜图像分类和渗出物分割。在视网膜图像分类中,先对视网膜图像进行数据增强,然后使用 ELNet 分类架构将视网膜图像分为正常或异常。在渗出物分割模块中,使用 Kirsch 边缘检测器对渗出物进行检测和分割。在检测渗出物之前,先检测并消除视网膜图像中的血管,从而提高渗出物检测方法的性能。本文使用用于血管提取的数字视网膜图像(DRIVE)和糖尿病视网膜病变数据库(DIARETDB1)视网膜图像数据集来检测视网膜图像中的渗出物。在这项研究中,所提出的方法效果显著,在 DRIVE 和 DIARETDB1 数据集上的灵敏度分别为 99.31% 和 99.31%,特异度分别为 97.44% 和 95%,准确度分别为 99.09% 和 98.8%。在这两个数据集中,如果不剔除OD和视网膜血管,我们观察到的渗出物检测准确率相似,两个数据集的平均准确率均为96.5%。然而,当剔除外伤和视网膜血管时,两个数据集的准确率都有明显提高,平均达到约 99.2%。我们将对这些性能进行分析,并与其他先进方法进行比较。
{"title":"Efficient retinal exudates detection method using ELNet in diabetic retinopathy assessment","authors":"G. Sasi,&nbsp;A. Kaleel Rahuman","doi":"10.1016/j.bspc.2024.107162","DOIUrl":"10.1016/j.bspc.2024.107162","url":null,"abstract":"<div><div>Diabetic Retinopathy (DR) can be detected at earlier stage by detecting exudates in retinal fundus images. In this article, the exudates are detected and segmented using the proposed Enhanced LeNet (ELNet) classification method. The proposed exudates segmentation method consists of the following modules Retinal image classification and Exudates segmentation. In retinal image classification, the retinal images are data augmented and then ELNet classification architecture is used to classify the retinal image into either normal or abnormal. In exudates segmentation module, the exudates are detected and segmented using Kirsch edge detector. The performance of the exudate’s detection method is improved by detecting and eliminating the blood vessels in the retinal image before detecting the exudates. In this paper, Digital Retinal Images for Vessel Extraction (DRIVE) and Diabetic Retinopathy database (DIARETDB1) retinal image datasets are used for the detection of exudates in the retinal images. In this study, the proposed method showcases remarkable results, demonstrating a sensitivity of 99.31% and 99.31%, specificity of 97.44% and 95%, and an accuracy of 99.09% and 98.8% for the DRIVE and DIARETDB1 datasets, respectively. Exudates detection in both datasets without eliminating OD and retinal blood vessels, we observe similar accuracy rates, Average 96.5% for both datasets. However, when eliminating OD and retinal blood vessels, the accuracy significantly improved for both datasets, reaching approximately 99.2% average. The performance is analyzed and compared with other state of the art methods.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"100 ","pages":"Article 107162"},"PeriodicalIF":4.9,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142657560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Biomedical Signal Processing and Control
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1