Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109606
S. Savitha , A. Rajiv Kannan , K. Logeswaran
The study addresses the critical challenge of accurately predicting cardiovascular disease (CVD), a leading cause of mortality worldwide, where early diagnosis is crucial for effective intervention. Traditional models often struggle with high-dimensional data, imbalanced classes, and nonlinear feature interactions, limiting prediction reliability. Motivated by these gaps, this research proposes a hybrid methodology integrating Harris Hawks Search (HHS) for feature optimization with Radial Basis Function Networks (RBFN) to enhance CVD risk assessment. The HHS algorithm efficiently selects key predictive features such as chest pain type and number of vessels, reducing dimensionality while preserving vital information. Trained on optimized features, the RBFN classifier achieved superior performance with 92.1% accuracy, high sensitivity, and specificity, surpassing conventional models like Logistic Regression (81.2%) and Random Forest (86.7%). Ablation studies confirm each component’s contribution, with significant gains validated statistically (p < 0.05). The hybrid model also offers computational efficiency with training times around 31.7 s. Future work aims to validate this approach on diverse, larger datasets and integrate it into real-time clinical decision support systems, advancing personalized, interpretable, and efficient cardiovascular healthcare tools.
{"title":"Hybrid feature optimization and radial basis function networks for cardiovascular disease prediction","authors":"S. Savitha , A. Rajiv Kannan , K. Logeswaran","doi":"10.1016/j.bspc.2026.109606","DOIUrl":"10.1016/j.bspc.2026.109606","url":null,"abstract":"<div><div>The study addresses the critical challenge of accurately predicting cardiovascular disease (CVD), a leading cause of mortality worldwide, where early diagnosis is crucial for effective intervention. Traditional models often struggle with high-dimensional data, imbalanced classes, and nonlinear feature interactions, limiting prediction reliability. Motivated by these gaps, this research proposes a hybrid methodology integrating Harris Hawks Search (HHS) for feature optimization with Radial Basis Function Networks (RBFN) to enhance CVD risk assessment. The HHS algorithm efficiently selects key predictive features such as chest pain type and number of vessels, reducing dimensionality while preserving vital information. Trained on optimized features, the RBFN classifier achieved superior performance with 92.1% accuracy, high sensitivity, and specificity, surpassing conventional models like Logistic Regression (81.2%) and Random Forest (86.7%). Ablation studies confirm each component’s contribution, with significant gains validated statistically (p < 0.05). The hybrid model also offers computational efficiency with training times around 31.7 s. Future work aims to validate this approach on diverse, larger datasets and integrate it into real-time clinical decision support systems, advancing personalized, interpretable, and efficient cardiovascular healthcare tools.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109606"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Accurate prognostic evaluation of endometrial cancer relies on precise pathological grading and lymphovascular space invasion (LVSI) assessment. To overcome the inherent limitation of single-modal MRI in characterizing heterogeneous tumors, this paper proposes a Flexible Multi-modal Classification Network (FMCNet). FMCNet effectively integrates complementary MRI sequences (T1WI, T2WI, ADC, DWI) through two novel modules: an Explicit-Implicit Feature analysis module, which hierarchically disentangles shallow anatomical and deep pathological features via adaptive recalibration, and a Cross-modal Enhancement Fusion framework, which constructs unified representations through attention-guided interactions to synergize complementary information while suppressing redundancy. Evaluated on a clinical dataset of 297 patients (2,889 images), FMCNet achieves state-of-the-art performance with an overall accuracy of 93.4% (94.8% for T1WI + T2WI + ADC), significantly outperforming conventional models (VGG16: 89.2%; DenseNet: 87.6%). Ablation studies confirm the critical roles of both modules. The framework’s ability to mitigate inter-modal interference underscores its potential for improving diagnostic assessment.
{"title":"Flexible multi-modal classification network for endometrial carcinoma diagnosis","authors":"Lingling Fang, Wenhui Zhang, Yongcheng Yu, Qian Wu","doi":"10.1016/j.bspc.2026.109574","DOIUrl":"10.1016/j.bspc.2026.109574","url":null,"abstract":"<div><div>Accurate prognostic evaluation of endometrial cancer relies on precise pathological grading and lymphovascular space invasion (LVSI) assessment. To overcome the inherent limitation of single-modal MRI in characterizing heterogeneous tumors, this paper proposes a Flexible Multi-modal Classification Network (FMCNet). FMCNet effectively integrates complementary MRI sequences (T1WI, T2WI, ADC, DWI) through two novel modules: an Explicit-Implicit Feature analysis module, which hierarchically disentangles shallow anatomical and deep pathological features via adaptive recalibration, and a Cross-modal Enhancement Fusion framework, which constructs unified representations through attention-guided interactions to synergize complementary information while suppressing redundancy. Evaluated on a clinical dataset of 297 patients (2,889 images), FMCNet achieves state-of-the-art performance with an overall accuracy of 93.4% (94.8% for T1WI + T2WI + ADC), significantly outperforming conventional models (VGG16: 89.2%; DenseNet: 87.6%). Ablation studies confirm the critical roles of both modules. The framework’s ability to mitigate inter-modal interference underscores its potential for improving diagnostic assessment.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109574"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109513
Hamed Mojtahed , Ramesh R. Rao , Christopher Paolini , Mahasweta Sarkar
Background and Objective:
Heart Rate Variability (HRV) measures derived from RR-intervals serve as crucial biomarkers for cardiovascular system responses to disease, physical activity, and stress. While healthcare professionals and general users utilize these measurements to gauge an individual’s well-being and assess their stress and fitness levels, their accuracy can be compromised by low-cost wearable devices, motion artifacts, poor electrode contact, and other factors that affect the RR-interval time series data derived from electrocardiograms. This study aims to develop and evaluate a Gaussian Process framework for correcting compromised RR-intervals.
Methods:
Monte Carlo simulations were used to distribute errors across the RR-interval series, which were sectioned by physical activity. Uniformly distributed additive Gaussian noise was synthetically introduced at various error rates (10%, 20%, and 30%) and durations (1 s, 3 s, 5 s, 7 s, and variable). Multiple Gaussian Process kernels, including basic and hybrid combinations, were applied to reconstruct the erroneous series. The results are compared with the widely used Cubic Spline Interpolation (CSI). Correlation analysis was employed to study the effect of error duration and presence levels on reconstruction ability.
Results:
The Gaussian process with a Rational Quadratic kernel demonstrated substantial improvements in deviation across all 27 HRV metrics by 87.66% and 90.26% compared to CSI for fixed-duration and variable-duration conditions. When compared with raw noisy signals, the reduction reached 94% for both conditions. These figures were further improved when combining Rational Quadratic and Periodic kernels, with deviation reductions of 88.22% (fixed-duration) and 90.77% (variable-duration) relative to CSI.
Conclusions:
The Gaussian Process framework, particularly with the RQ kernel, provides a reliable method for correcting compromised RR-interval measurements and recovering HRV variables with the least deviation for most parameters. This approach could enhance the reliability of HRV measurements from consumer-grade devices, potentially improving both clinical assessments and personal health monitoring.
{"title":"Harmonizing heartbeats: RR-interval correction with Gaussian Process for reliable HRV","authors":"Hamed Mojtahed , Ramesh R. Rao , Christopher Paolini , Mahasweta Sarkar","doi":"10.1016/j.bspc.2026.109513","DOIUrl":"10.1016/j.bspc.2026.109513","url":null,"abstract":"<div><h3>Background and Objective:</h3><div>Heart Rate Variability (HRV) measures derived from RR-intervals serve as crucial biomarkers for cardiovascular system responses to disease, physical activity, and stress. While healthcare professionals and general users utilize these measurements to gauge an individual’s well-being and assess their stress and fitness levels, their accuracy can be compromised by low-cost wearable devices, motion artifacts, poor electrode contact, and other factors that affect the RR-interval time series data derived from electrocardiograms. This study aims to develop and evaluate a Gaussian Process framework for correcting compromised RR-intervals.</div></div><div><h3>Methods:</h3><div>Monte Carlo simulations were used to distribute errors across the RR-interval series, which were sectioned by physical activity. Uniformly distributed additive Gaussian noise was synthetically introduced at various error rates (10%, 20%, and 30%) and durations (1 s, 3 s, 5 s, 7 s, and variable). Multiple Gaussian Process kernels, including basic and hybrid combinations, were applied to reconstruct the erroneous series. The results are compared with the widely used Cubic Spline Interpolation (CSI). Correlation analysis was employed to study the effect of error duration and presence levels on reconstruction ability.</div></div><div><h3>Results:</h3><div>The Gaussian process with a Rational Quadratic kernel demonstrated substantial improvements in deviation across all 27 HRV metrics by 87.66% and 90.26% compared to CSI for fixed-duration and variable-duration conditions. When compared with raw noisy signals, the reduction reached 94% for both conditions. These figures were further improved when combining Rational Quadratic and Periodic kernels, with deviation reductions of 88.22% (fixed-duration) and 90.77% (variable-duration) relative to CSI.</div></div><div><h3>Conclusions:</h3><div>The Gaussian Process framework, particularly with the RQ kernel, provides a reliable method for correcting compromised RR-interval measurements and recovering HRV variables with the least deviation for most parameters. This approach could enhance the reliability of HRV measurements from consumer-grade devices, potentially improving both clinical assessments and personal health monitoring.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109513"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109571
M. Sarathkumar , J. Rajalakshmi , Jeyapandi Marimuthu , A. Solairaj
Electroencephalogram (EEG) is widely used to monitor the cerebral activities. However, artifacts arising from non-cerebral sources such as eye movements, eye blinks and muscle activities often corrupt the recorded signals and significantly reducing reliability. To address this challenge, this paper proposes an Optimized Eyeblink Artifact Removal in EEG using Attention-Based Spiking Neural Networks (EAR-ASNN-OSP). The proposed framework integrates four core modules: Signed Cumulative Distribution Transform (SCDT) for robust feature extraction, Attention Spiking Neural Network (ASNN) for artifact detection, Efficient Binary Crayfish Optimization (EBCO) for adaptive weight parameter optimization, and a Dendritic Neural Network (DNN) for effective artifact removal. The SCDT captures discriminative features such as variance, average rectified value, and peak-to-peak amplitude, which are then supplied to the ASNN for classification of EEG signals into normal, corrupt, and eyeblink categories. EBCO further enhances detection performance by optimizing the ASNN weights, while the DNN ensures accurate removal of identified artifacts. The system was implemented in Python, and its performance was rigorously evaluated using metrics such as accuracy, precision, recall, F1-score, entropy, and computational time. Experimental results demonstrate that EAR-ASNN-OSP outperforms existing techniques including AEMD-RES-KCLN, QEF-ML-EAD, and DBP-EWT-EES. Specifically, the proposed method achieves up to 15.17 % higher accuracy, 20.11 % higher precision, and 19.19 % higher F1-score compared to state-of-the-art approaches. These findings highlight the effectiveness and robustness of EAR-ASNN-OSP, making it a promising solution for reliable EEG artifact detection and removal, thereby enhancing the quality and interpretability of EEG-based clinical and research applications.
脑电图(EEG)被广泛用于监测大脑活动。然而,眼球运动、眨眼和肌肉活动等非大脑来源产生的伪影经常会破坏记录的信号,并大大降低可靠性。为了解决这一挑战,本文提出了一种基于注意力的峰值神经网络(EAR-ASNN-OSP)的优化的EEG眨眼伪迹去除方法。该框架集成了四个核心模块:用于鲁棒特征提取的签名累积分布变换(SCDT)、用于伪迹检测的注意力峰值神经网络(ASNN)、用于自适应权重参数优化的高效二进制小龙虾优化(EBCO)和用于有效去除伪迹的树突状神经网络(DNN)。SCDT捕获判别特征,如方差、平均整流值和峰对峰幅度,然后将其提供给ASNN,用于将EEG信号分类为正常、损坏和眨眼类别。EBCO通过优化ASNN权重进一步提高检测性能,而DNN则确保准确去除已识别的伪像。该系统是用Python实现的,其性能通过准确性、精密度、召回率、f1分数、熵和计算时间等指标进行了严格评估。实验结果表明,EAR-ASNN-OSP优于aemd - re - kcln、QEF-ML-EAD和DBP-EWT-EES等现有技术。具体而言,与现有方法相比,该方法的准确率提高了15.17%,精密度提高了20.11%,f1分数提高了19.19%。这些发现突出了EAR-ASNN-OSP的有效性和鲁棒性,使其成为可靠的脑电信号伪迹检测和去除的有希望的解决方案,从而提高了基于脑电图的临床和研究应用的质量和可解释性。
{"title":"Optimized eyeblink artifact removal in EEG signal using attention-based spiking neural networks","authors":"M. Sarathkumar , J. Rajalakshmi , Jeyapandi Marimuthu , A. Solairaj","doi":"10.1016/j.bspc.2026.109571","DOIUrl":"10.1016/j.bspc.2026.109571","url":null,"abstract":"<div><div>Electroencephalogram (EEG) is widely used to monitor the cerebral activities. However, artifacts arising from non-cerebral sources such as eye movements, eye blinks and muscle activities often corrupt the recorded signals and significantly reducing reliability. To address this challenge, this paper proposes an Optimized Eyeblink Artifact Removal in EEG using Attention-Based Spiking Neural Networks (EAR-ASNN-OSP). The proposed framework integrates four core modules: Signed Cumulative Distribution Transform (SCDT) for robust feature extraction, Attention Spiking Neural Network (ASNN) for artifact detection, Efficient Binary Crayfish Optimization (EBCO) for adaptive weight parameter optimization, and a Dendritic Neural Network (DNN) for effective artifact removal. The SCDT captures discriminative features such as variance, average rectified value, and peak-to-peak amplitude, which are then supplied to the ASNN for classification of EEG signals into normal, corrupt, and eyeblink categories. EBCO further enhances detection performance by optimizing the ASNN weights, while the DNN ensures accurate removal of identified artifacts. The system was implemented in Python, and its performance was rigorously evaluated using metrics such as accuracy, precision, recall, F1-score, entropy, and computational time. Experimental results demonstrate that EAR-ASNN-OSP outperforms existing techniques including AEMD-RES-KCLN, QEF-ML-EAD, and DBP-EWT-EES. Specifically, the proposed method achieves up to 15.17 % higher accuracy, 20.11 % higher precision, and 19.19 % higher F1-score compared to state-of-the-art approaches. These findings highlight the effectiveness and robustness of EAR-ASNN-OSP, making it a promising solution for reliable EEG artifact detection and removal, thereby enhancing the quality and interpretability of EEG-based clinical and research applications.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109571"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109490
Wessam M. Salama
A hybrid deep learning–machine learning framework for the automatic classification of myositis muscle illnesses from ultrasound images is presented in this paper. The suggested approach combines a Support Vector Machine (SVM) classifier and a regionally improved Fast R-CNN architecture with a VGG16 backbone to take advantage of robust decision boundaries and deep feature extraction. To increase the representation of muscle texture, a thorough preprocessing pipeline comprising edge-based noise suppression, contrast enhancement, and median filtering is put into place. A five-fold cross-validation approach is used to guarantee dependability, and performance metrics are presented with 95% Confidence Intervals (CI) and standard deviation. Strong discriminative ability is demonstrated by Receiver Operating Characteristic (ROC) analysis, which produced consistent results across folds and a high average AUC of 98.32%. The suggested framework outperforms current deep architectures with the greatest classification accuracy of 98.46%, according to comparative tests against ResNet50, DenseNet121, and EfficientNet-B0. These results demonstrate the stability, repeatability, and clinical application of the model for fine-grained distinction between the classes of Inclusion Body Myositis (IBM), Polymyositis (PM), Dermatomyositis (DM), and Normal. Therefore, the suggested Fast R-CNN–VGG16–SVM methodology offers a computationally effective and statistically verified method for automated myositis detection utilizing ultrasound imaging.
{"title":"Advanced detection of myositis muscle images based on regional enhanced deep learning models integrated with SVM and image processing techniques","authors":"Wessam M. Salama","doi":"10.1016/j.bspc.2026.109490","DOIUrl":"10.1016/j.bspc.2026.109490","url":null,"abstract":"<div><div>A hybrid deep learning–machine learning framework for the automatic classification of myositis muscle illnesses from ultrasound images is presented in this paper. The suggested approach combines a Support Vector Machine (SVM) classifier and a regionally improved Fast R-CNN architecture with a VGG16 backbone to take advantage of robust decision boundaries and deep feature extraction. To increase the representation of muscle texture, a thorough preprocessing pipeline comprising edge-based noise suppression, contrast enhancement, and median filtering is put into place. A five-fold cross-validation approach is used to guarantee dependability, and performance metrics are presented with 95% Confidence Intervals (CI) and standard deviation. Strong discriminative ability is demonstrated by Receiver Operating Characteristic (ROC) analysis, which produced consistent results across folds and a high average AUC of 98.32%. The suggested framework outperforms current deep architectures with the greatest classification accuracy of 98.46%, according to comparative tests against ResNet50, DenseNet121, and EfficientNet-B0. These results demonstrate the stability, repeatability, and clinical application of the model for fine-grained distinction between the classes of Inclusion Body Myositis (IBM), Polymyositis (PM), Dermatomyositis (DM), and Normal. Therefore, the suggested Fast R-CNN–VGG16–SVM methodology offers a computationally effective and statistically verified method for automated myositis detection utilizing ultrasound imaging.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109490"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109584
Yu Li , Jiaqing Liu , Rahul Kumar Jain , Yen-Wei Chen
Deep learning requires large and diverse datasets to effectively learn downstream tasks. However, medical imaging often suffers from limited data availability, privacy and low diversity. Generative data augmentation (GDA) offers a promising solution by synthesizing labeled samples to expand training datasets and support downstream tasks. While recent advances in generative models have shown strong potential, many existing GDA approaches rely only on lesion annotation masks, often failing to preserve fine anatomical details or capture subtle morphological variations. To address these limitations, we propose a multimodal (mask- and text-guided) diffusion framework that enables fine-grained semantic control and maintains anatomical consistency. We construct detailed text datasets for medical images using a multimodal large language model. Further, the semantic features, such as lesion color, contour and background context, are integrated into the diffusion model employing an attention-based module. Our method maintains anatomical consistency while enabling diverse semantic variations through fine-grained control, achieved via proposed attention conditional batch normalization. Experiments on the Kvasir-SEG, ISIC 2016 and 3D-IRCADb-01 datasets demonstrate that our method significantly improves segmentation performance across various architectures. Code, datasets and pretrained models are available.1
{"title":"Multimodal Brownian bridge diffusion model for controllable synthetic medical image generation","authors":"Yu Li , Jiaqing Liu , Rahul Kumar Jain , Yen-Wei Chen","doi":"10.1016/j.bspc.2026.109584","DOIUrl":"10.1016/j.bspc.2026.109584","url":null,"abstract":"<div><div>Deep learning requires large and diverse datasets to effectively learn downstream tasks. However, medical imaging often suffers from limited data availability, privacy and low diversity. Generative data augmentation (GDA) offers a promising solution by synthesizing labeled samples to expand training datasets and support downstream tasks. While recent advances in generative models have shown strong potential, many existing GDA approaches rely only on lesion annotation masks, often failing to preserve fine anatomical details or capture subtle morphological variations. To address these limitations, we propose a multimodal (mask- and text-guided) diffusion framework that enables fine-grained semantic control and maintains anatomical consistency. We construct detailed text datasets for medical images using a multimodal large language model. Further, the semantic features, such as lesion color, contour and background context, are integrated into the diffusion model employing an attention-based module. Our method maintains anatomical consistency while enabling diverse semantic variations through fine-grained control, achieved via proposed attention conditional batch normalization. Experiments on the Kvasir-SEG, ISIC 2016 and 3D-IRCADb-01 datasets demonstrate that our method significantly improves segmentation performance across various architectures. Code, datasets and pretrained models are available.<span><span><sup>1</sup></span></span></div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109584"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146038917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cardiovascular Disease (CVD) refers to a collection of heart and blood vessel disorders caused by the build-up of plaque in the arteries, which restricts the blood flow throughout the body. Existing detection methods failed to identify the disease at an early stage and may cause discomfort during certain diagnostic procedures. Hence, the DL-based Google Wide Slice Residual Network (G-WISeR-Net) model is introduced for effective CVD detection. The proposed G-WISeR-Net model employs the four most pivotal subprocesses: denoising, feature extraction, feature fusion, and detection. Initially, ECG images are sourced from the database, and binary image conversion-based denoising is performed. Then, medical features, Stationary Wavelet Transform (SWT), and statistical features, like mean, variance, relative energy, relative amplitude, entropy, kurtosis, and information gain, are extracted. Subsequently, the extracted features are fused by a Deep Residual Network (DRN) with Topsoe similarity. Then, CVD detection is performed by G-WISeR-Net, which is an integration of Google Network (GoogleNet) and Wide Slice Residual Network (WISeR). The experimental results highlight that the proposed G-WISeR-Net attained an accuracy of 91.645%, sensitivity of 92.786%, and specificity of 91.479%.
{"title":"Cardiovascular disease detection using google wide slice residual network approach by electrocardiogram images","authors":"Dugumari Siva Raja Kumar , Balajee Maram , Pravin Ramdas Kshirsagar , Telagarapu Prabhakar","doi":"10.1016/j.bspc.2026.109530","DOIUrl":"10.1016/j.bspc.2026.109530","url":null,"abstract":"<div><div>Cardiovascular Disease (CVD) refers to a collection of heart and blood vessel disorders caused by the build-up of plaque in the arteries, which restricts the blood flow throughout the body. Existing detection methods failed to identify the disease at an early stage and may cause discomfort during certain diagnostic procedures. Hence, the DL-based Google Wide Slice Residual Network (G-WISeR-Net) model is introduced for effective CVD detection. The proposed G-WISeR-Net model employs the four most pivotal subprocesses: denoising, feature extraction, feature fusion, and detection. Initially, ECG images are sourced from the database, and binary image conversion-based denoising is performed. Then, medical features, Stationary Wavelet Transform (SWT), and statistical features, like mean, variance, relative energy, relative amplitude, entropy, kurtosis, and information gain, are extracted. Subsequently, the extracted features are fused by a Deep Residual Network (DRN) with Topsoe similarity. Then, CVD detection is performed by G-WISeR-Net, which is an integration of Google Network (GoogleNet) and Wide Slice Residual Network (WISeR). The experimental results highlight that the proposed G-WISeR-Net attained an accuracy of 91.645%, sensitivity of 92.786%, and specificity of 91.479%.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109530"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109589
Runhao Liu , Ziming Chen , Guangzhen Yao , Peng Zhang
Skin cancer is one of the most prevalent and deadly forms of cancer worldwide, highlighting the critical importance of early detection and diagnosis in improving patient outcomes. Deep learning (DL) has shown significant promise in enhancing the accuracy and efficiency of automated skin disease diagnosis, particularly in detecting and classifying skin lesions. However, several challenges remain for DL-based skin cancer diagnosis, including complex features, image noise, intra-class variation, inter-class similarity, and data imbalance. This review synthesizes recent research and discusses innovative approaches to address these challenges, such as data augmentation, hybrid models, and feature fusion. Furthermore, the review highlights the integration of DL models into clinical workflows, offering insights into the potential of deep learning to revolutionize skin disease diagnosis and improve clinical decision-making. This review uniquely integrates a PRISMA-based methodology with a challenge-oriented taxonomy, providing a systematic and transparent synthesis of recent deep learning advances for skin disease diagnosis. It further highlights emerging directions such as hybrid CNN-Transformer architectures and uncertainty-aware models, emphasizing its contribution to future dermatological AI research.
{"title":"Exploring the challenge and value of deep learning in automated skin disease diagnosis","authors":"Runhao Liu , Ziming Chen , Guangzhen Yao , Peng Zhang","doi":"10.1016/j.bspc.2026.109589","DOIUrl":"10.1016/j.bspc.2026.109589","url":null,"abstract":"<div><div>Skin cancer is one of the most prevalent and deadly forms of cancer worldwide, highlighting the critical importance of early detection and diagnosis in improving patient outcomes. Deep learning (DL) has shown significant promise in enhancing the accuracy and efficiency of automated skin disease diagnosis, particularly in detecting and classifying skin lesions. However, several challenges remain for DL-based skin cancer diagnosis, including complex features, image noise, intra-class variation, inter-class similarity, and data imbalance. This review synthesizes recent research and discusses innovative approaches to address these challenges, such as data augmentation, hybrid models, and feature fusion. Furthermore, the review highlights the integration of DL models into clinical workflows, offering insights into the potential of deep learning to revolutionize skin disease diagnosis and improve clinical decision-making. This review uniquely integrates a PRISMA-based methodology with a challenge-oriented taxonomy, providing a systematic and transparent synthesis of recent deep learning advances for skin disease diagnosis. It further highlights emerging directions such as hybrid CNN-Transformer architectures and uncertainty-aware models, emphasizing its contribution to future dermatological AI research.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109589"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109459
S. Rooban , Iwin Thanakumar Joseph S , Mohamed Uvaze Ahamed Ayoobkhan , Muthukumara Rajaguru Kattiakara Muni Samy
The people are impacted by Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) virus pandemic in worldwide over last two years, commonly known as coronavirus disease (COVID 19). In an effort to stop outbreak, governments of numerous nations have implemented measures like social distance and full or partial lockdowns. There is less interaction between people as a result of these COVID-19 preventive and control measures. It reviews the contribution of robotics, enhanced with deep learning techniques, to the challenges set during the COVID-19 pandemic. Attention is placed mainly on robotic applications for diagnosis, monitoring, disinfection, patient assistance, and rehabilitation; IoT and blockchain technologies are considered only when they directly support a robotic function. Existing research is analyzed in a structured manner to identify the performance, capabilities, and limitations of different robotic models and deep learning techniques. Comparisons among the different approaches are also made to provide a better understanding of the current trends and future directions. The findings point out that integration of robotics with deep learning enhances safety, efficiency, and healthcare support in pandemic conditions.
{"title":"A comprehensive review of robotics and deep learning applications during the COVID-19 pandemic","authors":"S. Rooban , Iwin Thanakumar Joseph S , Mohamed Uvaze Ahamed Ayoobkhan , Muthukumara Rajaguru Kattiakara Muni Samy","doi":"10.1016/j.bspc.2026.109459","DOIUrl":"10.1016/j.bspc.2026.109459","url":null,"abstract":"<div><div>The people are impacted by Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) virus pandemic in worldwide over last two years, commonly known as coronavirus disease (COVID 19). In an effort to stop outbreak, governments of numerous nations have implemented measures like social distance and full or partial lockdowns. There is less interaction between people as a result of these COVID-19 preventive and control measures. It reviews the contribution of robotics, enhanced with deep learning techniques, to the challenges set during the COVID-19 pandemic. Attention is placed mainly on robotic applications for diagnosis, monitoring, disinfection, patient assistance, and rehabilitation; IoT and blockchain technologies are considered only when they directly support a robotic function. Existing research is analyzed in a structured manner to identify the performance, capabilities, and limitations of different robotic models and deep learning techniques. Comparisons among the different approaches are also made to provide a better understanding of the current trends and future directions. The findings point out that integration of robotics with deep learning enhances safety, efficiency, and healthcare support in pandemic conditions.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109459"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-19DOI: 10.1016/j.bspc.2026.109623
Lin Feng , Yongzhen Huo , Yuqiu Kong , Cheng Cheng
Emotion recognition using electroencephalography (EEG) has emerged as a key noninvasive physiological modality for human–computer interaction, intelligent healthcare, and mental-health monitoring. However, existing approaches often treat temporal and spatial features of EEG signals in isolation and lack bidirectional, deep interactive fusion mechanisms. Furthermore, multiscale analysis of temporal and spatial information remains limited in scope. To overcome these challenges, we introduce a Dual-Branch Multiscale Bi-Mamba framework (DBM-BiMamba). DBM-BiMamba is a unified, multiscale, bidirectional, spatiotemporal fusion framework. It comprises three components that are optimized together: a Multiscale Temporal Feature Learning (MTFL) branch, a Hierarchical Spatial Feature Learning (HSFL) branch and a Bidirectional Fusion Mamba (Bi-Mamba) module. The MTFL branch uses parallel depthwise separable convolutions, followed by a transformer encoder, to capture local micro-dynamics and long-range temporal dependencies. The HSFL branch performs hierarchical graph convolutions with node-wise attention to emphasize critical inter-electrode relationships and produce discriminative topological embeddings. The Bi-Mamba module then applies forward and backward state-space modeling to the temporal and spatial embeddings, fusing them at the sequence level and enabling efficient bidirectional spatiotemporal interactions rather than simply concatenating the features. Extensive ten-fold subject-independent cross-validation in the DEAP, SEED, and SEED-IV datasets demonstrates state-of-the-art accuracy of 95.55% for arousal and 95.02% for valence in DEAP, 94.77% in SEED, and 89.68% in SEED-IV. Ablation studies confirm the complementary gains of each component.
{"title":"A dual-branch multiscale model based on Bi-Mamba for EEG emotion recognition","authors":"Lin Feng , Yongzhen Huo , Yuqiu Kong , Cheng Cheng","doi":"10.1016/j.bspc.2026.109623","DOIUrl":"10.1016/j.bspc.2026.109623","url":null,"abstract":"<div><div>Emotion recognition using electroencephalography (EEG) has emerged as a key noninvasive physiological modality for human–computer interaction, intelligent healthcare, and mental-health monitoring. However, existing approaches often treat temporal and spatial features of EEG signals in isolation and lack bidirectional, deep interactive fusion mechanisms. Furthermore, multiscale analysis of temporal and spatial information remains limited in scope. To overcome these challenges, we introduce a Dual-Branch Multiscale Bi-Mamba framework (DBM-BiMamba). DBM-BiMamba is a unified, multiscale, bidirectional, spatiotemporal fusion framework. It comprises three components that are optimized together: a Multiscale Temporal Feature Learning (MTFL) branch, a Hierarchical Spatial Feature Learning (HSFL) branch and a Bidirectional Fusion Mamba (Bi-Mamba) module. The MTFL branch uses parallel depthwise separable convolutions, followed by a transformer encoder, to capture local micro-dynamics and long-range temporal dependencies. The HSFL branch performs hierarchical graph convolutions with node-wise attention to emphasize critical inter-electrode relationships and produce discriminative topological embeddings. The Bi-Mamba module then applies forward and backward state-space modeling to the temporal and spatial embeddings, fusing them at the sequence level and enabling efficient bidirectional spatiotemporal interactions rather than simply concatenating the features. Extensive ten-fold subject-independent cross-validation in the DEAP, SEED, and SEED-IV datasets demonstrates state-of-the-art accuracy of 95.55% for arousal and 95.02% for valence in DEAP, 94.77% in SEED, and 89.68% in SEED-IV. Ablation studies confirm the complementary gains of each component.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109623"},"PeriodicalIF":4.9,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146039392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}