Pub Date : 2026-01-16DOI: 10.1016/j.bspc.2026.109502
Seunghee Han , Juyeob Lee , Eunil Park
Heart disease is a major issue in modern society owing to its severity. However, to date, it heavily relies on human judgment, necessitating the need for technology that can aid in objective and rapid human diagnosis. Several studies attempted data-driven approaches to classify heart disease, but they are limited to specific diseases and may not apply to the real medical field. To address these challenges, we propose a suite of deep learning-based classifiers, including a CNN and a state-of-the-art ViT enhanced with an auxiliary UNet feature extractor. To classify the eight types of heart disease, we utilize multi-view echocardiogram images consisting of numbers that reflect the proportion of actual cardiac patients. The experimental results reveal that vanilla ViT is not suitable for the echocardiogram dataset (accuracy of 0.6451). However, the performance can be improved using the UNet auxiliary feature extraction network (achieving an accuracy of 0.8121 for EfficientUNet+ViT). Among the comparison models, our CNN achieved the highest performance with an accuracy of 0.8829 and minimal computational cost, demonstrating its efficacy for direct disease classification without the need for ViT.
{"title":"Towards objective heart disease classification: A deep neural network approach for comprehensive diagnosis","authors":"Seunghee Han , Juyeob Lee , Eunil Park","doi":"10.1016/j.bspc.2026.109502","DOIUrl":"10.1016/j.bspc.2026.109502","url":null,"abstract":"<div><div>Heart disease is a major issue in modern society owing to its severity. However, to date, it heavily relies on human judgment, necessitating the need for technology that can aid in objective and rapid human diagnosis. Several studies attempted data-driven approaches to classify heart disease, but they are limited to specific diseases and may not apply to the real medical field. To address these challenges, we propose a suite of deep learning-based classifiers, including a CNN and a state-of-the-art ViT enhanced with an auxiliary UNet feature extractor. To classify the eight types of heart disease, we utilize multi-view echocardiogram images consisting of numbers that reflect the proportion of actual cardiac patients. The experimental results reveal that vanilla ViT is not suitable for the echocardiogram dataset (accuracy of 0.6451). However, the performance can be improved using the UNet auxiliary feature extraction network (achieving an accuracy of 0.8121 for EfficientUNet+ViT). Among the comparison models, our CNN achieved the highest performance with an accuracy of 0.8829 and minimal computational cost, demonstrating its efficacy for direct disease classification without the need for ViT.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109502"},"PeriodicalIF":4.9,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145981139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-16DOI: 10.1016/j.bspc.2026.109559
Alaa Hussein Abdulaal , Morteza Valizadeh , Mehdi Chehel Amirani
Alzheimer’s disease (AD) and frontotemporal dementia (FTD) represent critical neurodegenerative disorders requiring early and accurate detection for optimal patient management. This study introduces a Hybrid Frequency-Spatial Attention Network (HFSAN) that integrates convolutional neural networks with specialized attention mechanisms for the automated classification of dementia using EEG. The proposed methodology employs decoupled frequency and spatial attention modules, enabling the targeted analysis of spectral and topographical features before adaptive fusion via a gating mechanism. Advanced signal processing techniques, including continuous wavelet transform and artifact subspace reconstruction, ensure optimal feature extraction from 19-channel EEG recordings.
Evaluation of the Miltiadous dataset, comprising 88 subjects, demonstrated exceptional performance with 96.7 % accuracy for AD versus healthy controls, 92.3 % for FTD versus healthy controls, and 87.6 % for AD versus FTD classification. The method significantly outperformed traditional machine learning (Random Forest: 84.7 %) and existing deep learning approaches (CNN-2D: 91.3 %, CNN-LSTM: 92.4 %). A strong correlation with Mini-Mental State Examination scores (r = 0.894) validates the clinical relevance of this finding. A comprehensive interpretability analysis, conducted through Grad-CAM visualization, reveals clinically meaningful attention patterns that are consistent with established neurophysiological markers. The proposed HFSAN represents a significant advancement toward clinically viable EEG-based dementia detection, offering superior accuracy, interpretability, and practical applicability for early diagnosis and disease monitoring.
{"title":"Neurophysiological biomarker extraction through decoupled attention mechanisms and EEG signal processing in Alzheimer’s disease and frontotemporal dementia","authors":"Alaa Hussein Abdulaal , Morteza Valizadeh , Mehdi Chehel Amirani","doi":"10.1016/j.bspc.2026.109559","DOIUrl":"10.1016/j.bspc.2026.109559","url":null,"abstract":"<div><div>Alzheimer’s disease (AD) and frontotemporal dementia (FTD) represent critical neurodegenerative disorders requiring early and accurate detection for optimal patient management. This study introduces a Hybrid Frequency-Spatial Attention Network (HFSAN) that integrates convolutional neural networks with specialized attention mechanisms for the automated classification of dementia using EEG. The proposed methodology employs decoupled frequency and spatial attention modules, enabling the targeted analysis of spectral and topographical features before adaptive fusion via a gating mechanism. Advanced signal processing techniques, including continuous wavelet transform and artifact subspace reconstruction, ensure optimal feature extraction from 19-channel EEG recordings.</div><div>Evaluation of the Miltiadous dataset, comprising 88 subjects, demonstrated exceptional performance with 96.7 % accuracy for AD versus healthy controls, 92.3 % for FTD versus healthy controls, and 87.6 % for AD versus FTD classification. The method significantly outperformed traditional machine learning (Random Forest: 84.7 %) and existing deep learning approaches (CNN-2D: 91.3 %, CNN-LSTM: 92.4 %). A strong correlation with Mini-Mental State Examination scores (r = 0.894) validates the clinical relevance of this finding. A comprehensive interpretability analysis, conducted through Grad-CAM visualization, reveals clinically meaningful attention patterns that are consistent with established neurophysiological markers. The proposed HFSAN represents a significant advancement toward clinically viable EEG-based dementia detection, offering superior accuracy, interpretability, and practical applicability for early diagnosis and disease monitoring.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109559"},"PeriodicalIF":4.9,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-16DOI: 10.1016/j.bspc.2026.109599
Xuning Zhang , Lu Li , Jianpeng Chen , Changlin Lv , Huan Yang , Yongming Xi
Accurate segmentation of cervical ossification and vertebrae regions in MRI is crucial for ervical ssification of the osterior ongitudinal igament (COPLL) diagnosis, as it forms the basis for deriving quantitative indicators essential for clinical decision-making. To address this need, we propose HSA-CompSeg, a novel hybrid framework for multi-view cervical MRI segmentation. The framework efficiently captures global contextual dependencies while maintaining semantic consistency across network layers. It further integrates spatial–channel collaborative attention and cross-channel mixing to enhance multi-scale feature fusion and improve boundary delineation. Beyond segmentation, we develop an automated morphological measurement pipeline, in which Landmark-based Morphology Quantification (LMQ) extracts axial-view indicators and Defect Compression Quantification (DCQ) estimates sagittal-view clinical metrics. Experiments on a COPLL MRI dataset comprising 32 patients (582 images) demonstrate that HSA-CompSeg achieves superior accuracy over state-of-the-art methods, with consistent improvements in DSC and HD95. Moreover, the resulting quantitative measurements exhibit high agreement with expert assessments, underscoring the clinical utility of the proposed end-to-end system for objective COPLL diagnosis and severity grading.
{"title":"HSA-CompSeg: Deep learning-based multi-view segmentation and automated morphological quantification for COPLL","authors":"Xuning Zhang , Lu Li , Jianpeng Chen , Changlin Lv , Huan Yang , Yongming Xi","doi":"10.1016/j.bspc.2026.109599","DOIUrl":"10.1016/j.bspc.2026.109599","url":null,"abstract":"<div><div>Accurate segmentation of cervical ossification and vertebrae regions in MRI is crucial for <span><math><mi>C</mi></math></span>ervical <span><math><mi>O</mi></math></span>ssification of the <span><math><mi>P</mi></math></span>osterior <span><math><mi>L</mi></math></span>ongitudinal <span><math><mi>L</mi></math></span>igament (COPLL) diagnosis, as it forms the basis for deriving quantitative indicators essential for clinical decision-making. To address this need, we propose HSA-CompSeg, a novel hybrid framework for multi-view cervical MRI segmentation. The framework efficiently captures global contextual dependencies while maintaining semantic consistency across network layers. It further integrates spatial–channel collaborative attention and cross-channel mixing to enhance multi-scale feature fusion and improve boundary delineation. Beyond segmentation, we develop an automated morphological measurement pipeline, in which Landmark-based Morphology Quantification (LMQ) extracts axial-view indicators and Defect Compression Quantification (DCQ) estimates sagittal-view clinical metrics. Experiments on a COPLL MRI dataset comprising 32 patients (582 images) demonstrate that HSA-CompSeg achieves superior accuracy over state-of-the-art methods, with consistent improvements in DSC and HD95. Moreover, the resulting quantitative measurements exhibit high agreement with expert assessments, underscoring the clinical utility of the proposed end-to-end system for objective COPLL diagnosis and severity grading.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109599"},"PeriodicalIF":4.9,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-16DOI: 10.1016/j.bspc.2026.109595
Changwei Wu , Yifei Chen , Yuxin Du , Jinying Zong , Jie Dong , Mingxuan Liu , Feiwei Qin , Yong Peng , Jin Fan , Chaomiao Wang
Early diagnosis of Alzheimer’s Disease (AD), particularly at the mild cognitive impairment stage, is essential for timely intervention. However, this process faces significant barriers, including reliance on subjective assessments and the high cost of advanced imaging techniques. While deep learning offers automated solutions to improve diagnostic accuracy, its widespread adoption remains constrained due to high energy requirements and computational demands, particularly in resource-limited settings. Spiking neural networks (SNNs) provide a promising alternative, as their brain-inspired design is well-suited to model the sparse and event-driven patterns characteristic of neural degeneration in AD. These networks offer the potential for developing interpretable, energy-efficient diagnostic tools. Despite their advantages, existing SNNs often suffer from limited expressiveness and challenges in stable training, which reduce their effectiveness in handling complex medical tasks. To address these shortcomings, we introduce FasterSNN, a hybrid neural architecture that combines biologically inspired Leaky Integrate-and-Fire (LIF) neurons with region-adaptive convolution and multi-scale spiking attention mechanisms. This approach facilitates efficient, sparse processing of 3D MRI data while maintaining high diagnostic accuracy. Experimental results on benchmark datasets reveal that FasterSNN delivers competitive performance with significantly enhanced efficiency and training stability, highlighting its potential for practical application in AD screening. Our source code is available at https://github.com/wuchangw/FasterSNN.
{"title":"Towards practical Alzheimer’s Disease diagnosis: A lightweight and interpretable spiking neural model","authors":"Changwei Wu , Yifei Chen , Yuxin Du , Jinying Zong , Jie Dong , Mingxuan Liu , Feiwei Qin , Yong Peng , Jin Fan , Chaomiao Wang","doi":"10.1016/j.bspc.2026.109595","DOIUrl":"10.1016/j.bspc.2026.109595","url":null,"abstract":"<div><div>Early diagnosis of Alzheimer’s Disease (AD), particularly at the mild cognitive impairment stage, is essential for timely intervention. However, this process faces significant barriers, including reliance on subjective assessments and the high cost of advanced imaging techniques. While deep learning offers automated solutions to improve diagnostic accuracy, its widespread adoption remains constrained due to high energy requirements and computational demands, particularly in resource-limited settings. Spiking neural networks (SNNs) provide a promising alternative, as their brain-inspired design is well-suited to model the sparse and event-driven patterns characteristic of neural degeneration in AD. These networks offer the potential for developing interpretable, energy-efficient diagnostic tools. Despite their advantages, existing SNNs often suffer from limited expressiveness and challenges in stable training, which reduce their effectiveness in handling complex medical tasks. To address these shortcomings, we introduce FasterSNN, a hybrid neural architecture that combines biologically inspired Leaky Integrate-and-Fire (LIF) neurons with region-adaptive convolution and multi-scale spiking attention mechanisms. This approach facilitates efficient, sparse processing of 3D MRI data while maintaining high diagnostic accuracy. Experimental results on benchmark datasets reveal that FasterSNN delivers competitive performance with significantly enhanced efficiency and training stability, highlighting its potential for practical application in AD screening. Our source code is available at <span><span>https://github.com/wuchangw/FasterSNN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109595"},"PeriodicalIF":4.9,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-16DOI: 10.1016/j.bspc.2026.109466
J Lefty Joyson , K.Ruba Soundar , P. Nancy
The non-prompt identification of lung disease causes major contribution to global mortality, leading to significant limitations of conventional diagnostic strategies. There are several challenges facing existing deep learning strategies, including, but not limited to, generalizability issues, poor integration of different features, and slow convergence. This research presents a groundbreaking diagnostic system that is centered on Multi-level Feature Gated Spatio-temporal Fusion with Siamese Tensor Transformer (MFGSF-STT), a new fusion architecture that carries both disease classification with high accuracy and radiology report generation that is clinically coherent at the same time. The main breakthrough is the multi-level gated fusion mechanism, which combines spatial, temporal, and semantic cues in a more efficient way than even the best current models. In addition to this, Animated Oat Optimization Algorithm (AOOA) is included in the system to improve performance much faster and offers better hyperparameter tuning than traditional optimization methods. Furthermore, LLaMA3 with supervised fine-tuning to produce context-filled and medically consistent radiology reports that are easier for doctors to understand and are also supportive of clinical decision-making. A significant link between language expressiveness and medical accuracy was demonstrated by the 98.5 % diagnostic accuracy and 97.3 % F1-score, BLEU score of 91.6, and clinical coherence score of 4.7/5 obtained with this framework. Overall, these frameworks have characteristics of being clinically useful solutions for diagnosis of lung disease, and solve many problems associated with other existing models to improve prediction capability and clinical applicability.
{"title":"A multi-modal fusion-based deep learning with finetuned LLaMA 3 for lung disease diagnosis using PACS radiology reports","authors":"J Lefty Joyson , K.Ruba Soundar , P. Nancy","doi":"10.1016/j.bspc.2026.109466","DOIUrl":"10.1016/j.bspc.2026.109466","url":null,"abstract":"<div><div>The non-prompt identification of lung disease causes major contribution to global mortality, leading to significant limitations of conventional diagnostic strategies. There are several challenges facing existing deep learning strategies, including, but not limited to, generalizability issues, poor integration of different features, and slow convergence. This research presents a groundbreaking diagnostic system that is centered on Multi-level Feature Gated Spatio-temporal Fusion with Siamese Tensor Transformer (MFGSF-STT), a new fusion architecture that carries both disease classification with high accuracy and radiology report generation that is clinically coherent at the same time. The main breakthrough is the multi-level gated fusion mechanism, which combines spatial, temporal, and semantic cues in a more efficient way than even the best current models. In addition to this, Animated Oat Optimization Algorithm (AOOA) is included in the system to improve performance much faster and offers better hyperparameter tuning than traditional optimization methods. Furthermore, LLaMA3 with supervised fine-tuning to produce context-filled and medically consistent radiology reports that are easier for doctors to understand and are also supportive of clinical decision-making. A significant link between language expressiveness and medical accuracy was demonstrated by the 98.5 % diagnostic accuracy and 97.3 % F1-score, BLEU score of 91.6, and clinical coherence score of 4.7/5 obtained with this framework. Overall, these frameworks have characteristics of being clinically useful solutions for diagnosis of lung disease, and solve many problems associated with other existing models to improve prediction capability and clinical applicability.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109466"},"PeriodicalIF":4.9,"publicationDate":"2026-01-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982083","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Generative models have emerged as a powerful solution to address data scarcity in deep learning for glioma diagnosis. Among these, pairwise image generation has garnered significant attention for its ability to enhance the diversity and utility of synthetic data. However, current methods are primarily limited to single-modality generation and lack precise control over crucial morphological attributes of glioma, such as shape, location, and size. This limitation hinders their broader clinical applicability. To address these challenges, we propose a novel modular generative framework. A key contribution is the introduction of a box conditioning mechanism that complements the commonly used segmentation masks and MRI images. Furthermore, we construct three aligned feature spaces through pre-training, which enable flexible and independent control over glioma morphology, surrounding anatomical context, and semantic information in multi-modal MRI. We train a total of six generators, modified from StyleGAN2, for two pixel-wise glioma segmentation map generation tasks and four modality MRI synthesis tasks. By sharing identical control vectors across the glioma and MRI generation tasks, our framework ensures superior anatomical consistency in the synthetic paired images. Extensive experiments on the BraTS 2023 dataset demonstrate the superiority of our method in terms of both synthetic image quality and utility in downstream tasks. Our code is available at https://github.com/LcQi-mic/Gli_edit.
{"title":"A modularly designed controllable generative framework for glioma and MRI editing via style representations enhancement","authors":"Liangce Qi, Zhengang Jiang, Weili Shi, Yu Miao, Guodong Wei","doi":"10.1016/j.bspc.2026.109579","DOIUrl":"10.1016/j.bspc.2026.109579","url":null,"abstract":"<div><div>Generative models have emerged as a powerful solution to address data scarcity in deep learning for glioma diagnosis. Among these, pairwise image generation has garnered significant attention for its ability to enhance the diversity and utility of synthetic data. However, current methods are primarily limited to single-modality generation and lack precise control over crucial morphological attributes of glioma, such as shape, location, and size. This limitation hinders their broader clinical applicability. To address these challenges, we propose a novel modular generative framework. A key contribution is the introduction of a box conditioning mechanism that complements the commonly used segmentation masks and MRI images. Furthermore, we construct three aligned feature spaces through pre-training, which enable flexible and independent control over glioma morphology, surrounding anatomical context, and semantic information in multi-modal MRI. We train a total of six generators, modified from StyleGAN2, for two pixel-wise glioma segmentation map generation tasks and four modality MRI synthesis tasks. By sharing identical control vectors across the glioma and MRI generation tasks, our framework ensures superior anatomical consistency in the synthetic paired images. Extensive experiments on the BraTS 2023 dataset demonstrate the superiority of our method in terms of both synthetic image quality and utility in downstream tasks. Our code is available at <span><span>https://github.com/LcQi-mic/Gli_edit</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109579"},"PeriodicalIF":4.9,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-15DOI: 10.1016/j.bspc.2026.109515
Ming Liu , Zichong Zhang , Haifeng Guo , Jianli Yang , Peng Xiong , Jieshuo Zhang , Xiuling Liu
Emotion recognition holds significant importance in fields such as biomedicine and brain-computer interfaces. Graph Convolutional Networks (GCNs) based on electroencephalographic (EEG) signals have been widely applied to emotion recognition tasks. However, existing methods often overlook the differential responses of brain regions during emotional reactions, and shallow GCN architectures struggle to fully exploit the spatial and functional correlations between EEG channels. To dynamically capture the responses of brain channels during emotional reactions and enhance the multi-scale feature extraction capability of GCNs, this study proposes an EEG emotion recognition method that combines a Graph Convolution Parameter Optimization Network (GCPONet) and a Multi-scale Emotion Recognition Network (MERNet). Specifically, GCPONet models the response differences of each channel during emotional reactions by dynamically optimizing the weight distribution of multiple EEG frequency bands ( and ), and classifies these channels into three state labels (”active”, ”stable”, and ”sluggish”) to quantify their activation levels. MERNet extracts multi-scale features by integrating label information with a graph coarsening strategy, constructing a feature representation framework from local to global levels. Subject-dependent and subject-independent experiments conducted on the SEED dataset demonstrate two key findings: (1) The channels’ state labels generated by GCPONet can accurately reflect the state of each channel, providing targeted information for emotion recognition; (2) Compared with existing popular methods, the proposed method can effectively capture multi-scale EEG features, avoid the over-smoothing issue, and achieve more favorable performance in emotion classification. This research offers a feasible new perspective for emotion recognition and brain network analysis.
{"title":"Multi-scale graph convolutional EEG emotion recognition method driven by dynamic channel state labels","authors":"Ming Liu , Zichong Zhang , Haifeng Guo , Jianli Yang , Peng Xiong , Jieshuo Zhang , Xiuling Liu","doi":"10.1016/j.bspc.2026.109515","DOIUrl":"10.1016/j.bspc.2026.109515","url":null,"abstract":"<div><div>Emotion recognition holds significant importance in fields such as biomedicine and brain-computer interfaces. Graph Convolutional Networks (GCNs) based on electroencephalographic (EEG) signals have been widely applied to emotion recognition tasks. However, existing methods often overlook the differential responses of brain regions during emotional reactions, and shallow GCN architectures struggle to fully exploit the spatial and functional correlations between EEG channels. To dynamically capture the responses of brain channels during emotional reactions and enhance the multi-scale feature extraction capability of GCNs, this study proposes an EEG emotion recognition method that combines a Graph Convolution Parameter Optimization Network (GCPONet) and a Multi-scale Emotion Recognition Network (MERNet). Specifically, GCPONet models the response differences of each channel during emotional reactions by dynamically optimizing the weight distribution of multiple EEG frequency bands (<span><math><mi>β</mi></math></span> and <span><math><mi>γ</mi></math></span>), and classifies these channels into three state labels (”active”, ”stable”, and ”sluggish”) to quantify their activation levels. MERNet extracts multi-scale features by integrating label information with a graph coarsening strategy, constructing a feature representation framework from local to global levels. Subject-dependent and subject-independent experiments conducted on the SEED dataset demonstrate two key findings: (1) The channels’ state labels generated by GCPONet can accurately reflect the state of each channel, providing targeted information for emotion recognition; (2) Compared with existing popular methods, the proposed method can effectively capture multi-scale EEG features, avoid the over-smoothing issue, and achieve more favorable performance in emotion classification. This research offers a feasible new perspective for emotion recognition and brain network analysis.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109515"},"PeriodicalIF":4.9,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-15DOI: 10.1016/j.bspc.2026.109603
Agnesh Chandra Yadav, Maheshkumar H. Kolekar
Accurate brain tumor classification and segmentation are critical for patient prognosis and clinical decision-making, as they directly guide treatment planning. This review article examines the major developments in brain tumor imaging research from 2020 to 2025, focusing on methods based on convolutional neural networks, U-Net and its extensions, attention mechanisms, transformer-based designs, hybrid models, and recent generative techniques. Studies conducted on widely used MRI datasets — such as BraTS, FeTS, TCGA, and Figshare — are discussed, covering different imaging sequences including T1, T2, FLAIR, and contrast-enhanced scans. Reported results are compared using common evaluation measures like accuracy, Dice coefficient, Hausdorff distance, and Intersection over Union. This article also discusses persistent difficulties faced by researchers, including variations between datasets, unequal availability of imaging modalities, limited annotated data, and the need for methods that preserve patient privacy during training. Current research directions include multimodal feature integration, learning representations without extensive manual labeling, distributed learning frameworks, and improved interpretability of model outputs. In addition to reviewing the current methods, this study points out their limitations and suggests future directions for building automated brain tumor detection and analysis systems that are reliable, scalable, and suitable for clinical use.
{"title":"Deep feature-based approaches for brain tumor classification and segmentation in medical imaging","authors":"Agnesh Chandra Yadav, Maheshkumar H. Kolekar","doi":"10.1016/j.bspc.2026.109603","DOIUrl":"10.1016/j.bspc.2026.109603","url":null,"abstract":"<div><div>Accurate brain tumor classification and segmentation are critical for patient prognosis and clinical decision-making, as they directly guide treatment planning. This review article examines the major developments in brain tumor imaging research from 2020 to 2025, focusing on methods based on convolutional neural networks, U-Net and its extensions, attention mechanisms, transformer-based designs, hybrid models, and recent generative techniques. Studies conducted on widely used MRI datasets — such as BraTS, FeTS, TCGA, and Figshare — are discussed, covering different imaging sequences including T1, T2, FLAIR, and contrast-enhanced scans. Reported results are compared using common evaluation measures like accuracy, Dice coefficient, Hausdorff distance, and Intersection over Union. This article also discusses persistent difficulties faced by researchers, including variations between datasets, unequal availability of imaging modalities, limited annotated data, and the need for methods that preserve patient privacy during training. Current research directions include multimodal feature integration, learning representations without extensive manual labeling, distributed learning frameworks, and improved interpretability of model outputs. In addition to reviewing the current methods, this study points out their limitations and suggests future directions for building automated brain tumor detection and analysis systems that are reliable, scalable, and suitable for clinical use.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109603"},"PeriodicalIF":4.9,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-15DOI: 10.1016/j.bspc.2026.109538
Yang Wen , Bai Chen , Wuzhen Shi , Song Wu , Xiaokang Yang , Bin Sheng
Automatic, accurate disease classification of medical images is of great significance for timely clinical diagnosis and intervention. However, most existing medical image classification methods still face many challenges, such as being misled by irrelevant information, inaccurate feature extraction from complex lesion regions, and the lack of domain-specific structure knowledge. To address these limitations, we propose DB-HDFFN, a Dual-Branch Hierarchical Dynamic Feature Fusion Network, which introduces three key innovations. Firstly, we develop a Dynamic Sparse Attention Block (DSA-Block). Unlike existing Transformer-based methods that apply full attention and are easily distracted by irrelevant regions, the proposed DSA-Block introduces a novel top-k dynamic sparsification strategy that selectively preserves the most informative regions, enabling noise-robust, computation-efficient global modeling. Secondly, we design an Intra-Inter Variance Feature Extracting Block (ICV-Block) that leverages the inherent property of significant intra-image variance and relatively slight inter-image variance in medical images. By modeling this unique property, the ICV-Block enables the network to adaptively emphasize subtle lesion regions while suppressing misleading anatomical variations, thereby addressing a long-standing challenge in accurately capturing fine-grained pathological patterns in complex lesion areas. Finally, we propose a Dynamic Fusion Block (DF-Block) that performs hierarchical cross-branch fusion, ensuring effective integration of multi-scale global and local representations. The ACC and F1 values of our proposed model were 80.28% and 80.31% on the COVID-19-CT dataset, 92.47% and 92.49% on the Chest X-ray PA dataset, and 87.68% and 87.63% on the Kvasir dataset. These results across multiple datasets thoroughly verify that the proposed dual-branch hierarchical dynamic feature fusion network outperforms other state-of-the-art models for medical image classification.
{"title":"DB-HDFFN: Dual Branch Hierarchical Dynamic Feature Fusion Network for medical image classification","authors":"Yang Wen , Bai Chen , Wuzhen Shi , Song Wu , Xiaokang Yang , Bin Sheng","doi":"10.1016/j.bspc.2026.109538","DOIUrl":"10.1016/j.bspc.2026.109538","url":null,"abstract":"<div><div>Automatic, accurate disease classification of medical images is of great significance for timely clinical diagnosis and intervention. However, most existing medical image classification methods still face many challenges, such as being misled by irrelevant information, inaccurate feature extraction from complex lesion regions, and the lack of domain-specific structure knowledge. To address these limitations, we propose DB-HDFFN, a Dual-Branch Hierarchical Dynamic Feature Fusion Network, which introduces three key innovations. Firstly, we develop a Dynamic Sparse Attention Block (DSA-Block). Unlike existing Transformer-based methods that apply full attention and are easily distracted by irrelevant regions, the proposed DSA-Block introduces a novel top-k dynamic sparsification strategy that selectively preserves the most informative regions, enabling noise-robust, computation-efficient global modeling. Secondly, we design an Intra-Inter Variance Feature Extracting Block (ICV-Block) that leverages the inherent property of significant intra-image variance and relatively slight inter-image variance in medical images. By modeling this unique property, the ICV-Block enables the network to adaptively emphasize subtle lesion regions while suppressing misleading anatomical variations, thereby addressing a long-standing challenge in accurately capturing fine-grained pathological patterns in complex lesion areas. Finally, we propose a Dynamic Fusion Block (DF-Block) that performs hierarchical cross-branch fusion, ensuring effective integration of multi-scale global and local representations. The ACC and F1 values of our proposed model were 80.28% and 80.31% on the COVID-19-CT dataset, 92.47% and 92.49% on the Chest X-ray PA dataset, and 87.68% and 87.63% on the Kvasir dataset. These results across multiple datasets thoroughly verify that the proposed dual-branch hierarchical dynamic feature fusion network outperforms other state-of-the-art models for medical image classification.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109538"},"PeriodicalIF":4.9,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ultrasound image segmentation plays a vital role in medical diagnosis. However, automatic segmentation remains a significant challenge due to the presence of noise, low contrast, and the limited availability of annotated data. This paper proposes a novel semi-supervised segmentation approach, termed WAF (Wavelet Attention Fusion). The method applies discrete wavelet transform (DWT) to decompose ultrasound images into sub-bands of different frequencies, primarily utilizing the low-frequency components for global feature representation, while the high-frequency components capture fine details and edges. To improve the model’s ability to focus on critical regions, we introduce an attention fusion module that integrates both channel and spatial attention mechanisms. This design effectively enhances the model’s perception of important frequency and spatial features on low resolution ultrasound images. Experiments on multiple ultrasound segmentation datasets demonstrate that WAF consistently outperforms traditional FixMatch and other state-of-the-art semi-supervised methods. Specifically, WAF yields Dice score improvements of +1.38%, +2.41%, and +3.88% over FixMatch on HC18, DDTI, and CCAUI, respectively. Ablation studies further confirm the essential role of wavelet decomposition and dual attention in boosting performance. Our findings suggest that WAF can significantly improve semi-supervised medical image segmentation while reducing reliance on labeled data. The code is publicly available at https://github.com/wxmadm/WAF.
{"title":"Wavelet attention fusion for semi-supervised ultrasound segmentation","authors":"Xiaming Wu , Wenbo Yue , Xinglong Wu , Qing Huang , Chang Li , Yajun Yu , Guoping Xu","doi":"10.1016/j.bspc.2026.109566","DOIUrl":"10.1016/j.bspc.2026.109566","url":null,"abstract":"<div><div>Ultrasound image segmentation plays a vital role in medical diagnosis. However, automatic segmentation remains a significant challenge due to the presence of noise, low contrast, and the limited availability of annotated data. This paper proposes a novel semi-supervised segmentation approach, termed <strong>WAF (Wavelet Attention Fusion)</strong>. The method applies discrete wavelet transform (DWT) to decompose ultrasound images into sub-bands of different frequencies, primarily utilizing the low-frequency components for global feature representation, while the high-frequency components capture fine details and edges. To improve the model’s ability to focus on critical regions, we introduce an attention fusion module that integrates both channel and spatial attention mechanisms. This design effectively enhances the model’s perception of important frequency and spatial features on low resolution ultrasound images. Experiments on multiple ultrasound segmentation datasets demonstrate that WAF consistently outperforms traditional FixMatch and other state-of-the-art semi-supervised methods. Specifically, WAF yields Dice score improvements of +1.38%, +2.41%, and +3.88% over FixMatch on HC18, DDTI, and CCAUI, respectively. Ablation studies further confirm the essential role of wavelet decomposition and dual attention in boosting performance. Our findings suggest that WAF can significantly improve semi-supervised medical image segmentation while reducing reliance on labeled data. The code is publicly available at <span><span>https://github.com/wxmadm/WAF</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"117 ","pages":"Article 109566"},"PeriodicalIF":4.9,"publicationDate":"2026-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145982051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}