Pub Date : 2025-12-27DOI: 10.1016/j.bspc.2025.109410
Tianjiao Feng , Yuanbo Shi , Xiaofeng Wang , Haoran Zhao , Weipeng Chao
Multiclass cell detection plays a crucial role in various biomedical applications, particularly in cell biology. Although the development of YOLO object detection models has advanced real-time detection capabilities, challenges such as heterogeneous staining protocols, device variability, and object occlusion continue to hinder performance in medical imaging. To address these issues, we present YOLO-MFDS, a lightweight detector for multiclass cell detection, which is built upon YOLOv11n. To handle staining heterogeneity, device variability, and occlusion, YOLO-MFDS combines four complementary components: DLK-SF for dynamic large-kernel perception with saliency-guided fusion, CSA-Rep for short cross-stage aggregation paths through re-parameterisation, CARAFE for content-aware upsampling that preserves fine boundaries, and PSA-iEMA for low-cost channel–spatial reweighting with stabilised statistics. We also used CWDLoss distillation to align the channel-wise responses in dense and overlapping regions. On BCCD, YOLO-MFDS improves YOLOv11n by 5.5% mAP at IoU 0.5% and 6.8% mAP at IoU 0.5% to 0.95, and on Br35h, by 3.9% and 8.2%. The cross-dataset validation of BCCD and Br35h indicated good generalisation. The method is designed as a clinician-in-the-loop decision-support tool and can adapt, with modest domain adaptation, to additional leukaemia subtypes and tumour cytology. The source code and dataset splits are available at: https://github.com/Fengtj123/YOLO-MFDS.git.
{"title":"YOLO-MFDS: Medical small object detection algorithm based on multi-feature fusion","authors":"Tianjiao Feng , Yuanbo Shi , Xiaofeng Wang , Haoran Zhao , Weipeng Chao","doi":"10.1016/j.bspc.2025.109410","DOIUrl":"10.1016/j.bspc.2025.109410","url":null,"abstract":"<div><div>Multiclass cell detection plays a crucial role in various biomedical applications, particularly in cell biology. Although the development of YOLO object detection models has advanced real-time detection capabilities, challenges such as heterogeneous staining protocols, device variability, and object occlusion continue to hinder performance in medical imaging. To address these issues, we present YOLO-MFDS, a lightweight detector for multiclass cell detection, which is built upon YOLOv11n. To handle staining heterogeneity, device variability, and occlusion, YOLO-MFDS combines four complementary components: DLK-SF for dynamic large-kernel perception with saliency-guided fusion, CSA-Rep for short cross-stage aggregation paths through re-parameterisation, CARAFE for content-aware upsampling that preserves fine boundaries, and PSA-iEMA for low-cost channel–spatial reweighting with stabilised statistics. We also used CWDLoss distillation to align the channel-wise responses in dense and overlapping regions. On BCCD, YOLO-MFDS improves YOLOv11n by 5.5% mAP at IoU 0.5% and 6.8% mAP at IoU 0.5% to 0.95, and on Br35h, by 3.9% and 8.2%. The cross-dataset validation of BCCD and Br35h indicated good generalisation. The method is designed as a clinician-in-the-loop decision-support tool and can adapt, with modest domain adaptation, to additional leukaemia subtypes and tumour cytology. The source code and dataset splits are available at: <span><span>https://github.com/Fengtj123/YOLO-MFDS.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109410"},"PeriodicalIF":4.9,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recent advances in spatial transcriptomics technology have facilitated the generation of increasingly diverse datasets, offering enhanced opportunities to explore organizational structure and function in a spatial context. However, the effective integration and analysis of such data remain challenging. To effectively integrate multi-slice information, we propose STBCGAE, an adversarial autoencoder-based framework for spatial domain identification in multi-slice spatial transcriptomics data. STBCGAE employs mutual nearest neighbor and nearest spot iterative algorithms to align the spatial positions across multiple slices, simultaneously establishing more precise cross-slice spatial relationships through the construction of a 3D neighbor map. To generate more effective feature embeddings, STBCGAE integrates batch information, gene expression and spatial information using a graph neural network-based autoencoder so that the model can effectively differentiate between technical variants and biological signals. Moreover, to eliminate batch effects, we introduce a batch classifier to train against the encoder. Finally, spatial clustering is performed using the Mclust method to identify spatial domains with expression profiles. By performing extensive experiments on multiple datasets, we demonstrate the capability of STBCGAE to effectively integrate multiple batches of samples in a variety of scenarios, significantly improving the accuracy of multi-slice spatial domain recognition.
{"title":"Spatial domain recognition for multi-slice spatial transcriptomics based on self encoder adversarial training","authors":"Xueqin Zhang , Xuemei Peng , Huitong Zhu , Weihong Ding , Yunlan Zhou , Zhichao Wu","doi":"10.1016/j.bspc.2025.109414","DOIUrl":"10.1016/j.bspc.2025.109414","url":null,"abstract":"<div><div>Recent advances in spatial transcriptomics technology have facilitated the generation of increasingly diverse datasets, offering enhanced opportunities to explore organizational structure and function in a spatial context. However, the effective integration and analysis of such data remain challenging. To effectively integrate multi-slice information, we propose STBCGAE, an adversarial autoencoder-based framework for spatial domain identification in multi-slice spatial transcriptomics data. STBCGAE employs mutual nearest neighbor and nearest spot iterative algorithms to align the spatial positions across multiple slices, simultaneously establishing more precise cross-slice spatial relationships through the construction of a 3D neighbor map. To generate more effective feature embeddings, STBCGAE integrates batch information, gene expression and spatial information using a graph neural network-based autoencoder so that the model can effectively differentiate between technical variants and biological signals. Moreover, to eliminate batch effects, we introduce a batch classifier to train against the encoder. Finally, spatial clustering is performed using the Mclust method to identify spatial domains with expression profiles. By performing extensive experiments on multiple datasets, we demonstrate the capability of STBCGAE to effectively integrate multiple batches of samples in a variety of scenarios, significantly improving the accuracy of multi-slice spatial domain recognition.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109414"},"PeriodicalIF":4.9,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-27DOI: 10.1016/j.bspc.2025.109432
Hongbo Guo , Shuotian Li , Guojun Long , Qiqi Liang , Yiran Wang
Parkinson’s disease (PD) diagnosis lacks objective biomarkers, leading to subjectivity and delayed treatment. This work aims to improve diagnostic accuracy through a multimodal neuroimaging–EEG framework. Method: We designed a compact CNN-based pipeline that integrates structural MRI (sMRI), functional MRI (fMRI), and electroencephalography (EEG). Modality-specific encoders were fused with a lightweight attention head and optimized using Bayesian methods. Experiments used PPMI MRI/clinical data and OpenNeuro PD EEG datasets with subject-wise train/validation/test splits. Results: The model achieved an accuracy of 0.85 and an F1-score of 0.86, outperforming single-modality baselines and traditional machine-learning methods. Frequency-domain attention enhanced β-band features, while branch masking enabled robust handling of missing modalities. Conclusion: The framework provided interpretable EEG and MRI markers with efficient offline inference. The proposed multimodal CNN demonstrates offline feasibility for PD diagnosis, improving robustness, interpretability, and diagnostic efficiency compared to conventional methods. Significance: This study introduces a scalable, lightweight neuroimaging–EEG fusion strategy compatible with brain–computer interface (BCI) pipelines. It not only enhances PD diagnostics but also provides a methodological foundation for personalized care and future applications in other neurological diseases.
{"title":"Multimodal MRI–EEG fusion for brain–computer interface applications using a lightweight CNN and attention in offline Parkinson’s disease diagnosis","authors":"Hongbo Guo , Shuotian Li , Guojun Long , Qiqi Liang , Yiran Wang","doi":"10.1016/j.bspc.2025.109432","DOIUrl":"10.1016/j.bspc.2025.109432","url":null,"abstract":"<div><div>Parkinson’s disease (PD) diagnosis lacks objective biomarkers, leading to subjectivity and delayed treatment. This work aims to improve diagnostic accuracy through a multimodal neuroimaging–EEG framework. Method: We designed a compact CNN-based pipeline that integrates structural MRI (sMRI), functional MRI (fMRI), and electroencephalography (EEG). Modality-specific encoders were fused with a lightweight attention head and optimized using Bayesian methods. Experiments used PPMI MRI/clinical data and OpenNeuro PD EEG datasets with subject-wise train/validation/test splits. Results: The model achieved an accuracy of 0.85 and an F1-score of 0.86, outperforming single-modality baselines and traditional machine-learning methods. Frequency-domain attention enhanced β-band features, while branch masking enabled robust handling of missing modalities. Conclusion: The framework provided interpretable EEG and MRI markers with efficient offline inference. The proposed multimodal CNN demonstrates offline feasibility for PD diagnosis, improving robustness, interpretability, and diagnostic efficiency compared to conventional methods. Significance: This study introduces a scalable, lightweight neuroimaging–EEG fusion strategy compatible with brain–computer interface (BCI) pipelines. It not only enhances PD diagnostics but also provides a methodological foundation for personalized care and future applications in other neurological diseases.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109432"},"PeriodicalIF":4.9,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145884666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-26DOI: 10.1016/j.bspc.2025.109425
Qi Sun , Minfeng Wu , Aorui Gou , Yibo Fan
Cross-domain joint segmentation of Optic Disc (OD) and Optic Cup (OC) on fundus images is crucial for early glaucoma detection and treatment. However, domain shifts caused by differences in imaging devices and acquisition protocols across medical centers severely degrade the generalization ability of segmentation models, and re-annotation and re-training are labor-intensive and time-consuming. Unsupervised Domain Adaptation (UDA) addresses label scarcity by leveraging source domain labels, but existing methods often overlook the fundamental image signal processing (ISP) pipeline that is a primary source of domain gaps. This paper proposes a two-stage UDA method: In the warm-up phase, we introduce a multi-level alignment strategy including a learnable ISP module that aligns cross-domain style discrepancies by simulating the imaging process. Feature- and output-level alignments further promote semantics-aware learning of domain-invariant features. In the fine-tuning phase, we adopt a progressive mean-teacher strategy combined with a confidence-guided bidirectional CutMix augmentation, which facilitates consistency learning from pseudo-labels while mitigating the impact of noisy supervision, thereby improving cross-domain generalization. Experiments on two public fundus datasets show that our ISPSeg achieves an average improvement of 2.3% in Dice scores for optic disc (OD) and optic cup (OC) segmentation compared to state-of-the-art UDA methods, which demonstrates the clinical potential of ISPSeg for glaucoma diagnosis.
{"title":"ISPSeg: Unsupervised domain adaptive fundus image segmentation via learnable image signal processing and progressive teacher","authors":"Qi Sun , Minfeng Wu , Aorui Gou , Yibo Fan","doi":"10.1016/j.bspc.2025.109425","DOIUrl":"10.1016/j.bspc.2025.109425","url":null,"abstract":"<div><div>Cross-domain joint segmentation of Optic Disc (OD) and Optic Cup (OC) on fundus images is crucial for early glaucoma detection and treatment. However, domain shifts caused by differences in imaging devices and acquisition protocols across medical centers severely degrade the generalization ability of segmentation models, and re-annotation and re-training are labor-intensive and time-consuming. Unsupervised Domain Adaptation (UDA) addresses label scarcity by leveraging source domain labels, but existing methods often overlook the fundamental image signal processing (ISP) pipeline that is a primary source of domain gaps. This paper proposes a two-stage UDA method: In the warm-up phase, we introduce a multi-level alignment strategy including a learnable ISP module that aligns cross-domain style discrepancies by simulating the imaging process. Feature- and output-level alignments further promote semantics-aware learning of domain-invariant features. In the fine-tuning phase, we adopt a progressive mean-teacher strategy combined with a confidence-guided bidirectional CutMix augmentation, which facilitates consistency learning from pseudo-labels while mitigating the impact of noisy supervision, thereby improving cross-domain generalization. Experiments on two public fundus datasets show that our ISPSeg achieves an average improvement of 2.3% in Dice scores for optic disc (OD) and optic cup (OC) segmentation compared to state-of-the-art UDA methods, which demonstrates the clinical potential of ISPSeg for glaucoma diagnosis.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109425"},"PeriodicalIF":4.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-26DOI: 10.1016/j.bspc.2025.109436
Juanfei Li , Pazilya Yusantay , Kunru Wang , Huiyu Zhou , Shuiping Gou , Gen Li
Medical image classification technology holds significant clinical implications for the early screening, diagnosis, and treatment of diseases. However, most existing medical image classification models focus on single-organ analysis, which presents limitations in task generalization. Their scalability and broader applicability remain underexplored. This study aims to develop a generalizable multi-organ lesion classification framework to overcome the challenges of small lesion-to-background ratios, indistinct morphological boundaries, and heterogeneous lesion manifestations. This enables robust screening of pathological abnormalities across multiple abdominal organs.
We present OSRE-MLC, an innovative framework integrating two key components: (1) an organ-specific feature perception module that dynamically adapts to anatomical variations while preventing feature degradation, and (2) a region-specific enhancement module that optimizes discriminative lesion representation prior to classification. The framework integrates multi-organ abdominal CT images, including liver, kidney, and pancreas, from five separate datasets. The architecture uniquely combines multi-organ segmentation with attention-based feature refinement, enabling simultaneous organ localization and pathology characterization through an end-to-end trainable network.
Comprehensive evaluation on abdominal CT datasets has demonstrated OSRE-MLC’s superior performance, achieving 95.0% accuracy, 94.44% F1-score, 95.0% precision, and 93.96% recall in liver, kidney, and pancreas lesion screening, significantly outperforming existing methods. The proposed framework establishes a new paradigm for multi-organ pathological analysis by effectively addressing feature degradation and inter-organ variability. Its clinically interpretable architecture and robust performance demonstrate significant potential for improving diagnostic accuracy in complex abdominal imaging, offering promising applications in precision medicine and computer-aided diagnosis systems.
{"title":"Abdominal multi-organ lesion recognition via organ-specific feature perception and regionally enhanced feature learning","authors":"Juanfei Li , Pazilya Yusantay , Kunru Wang , Huiyu Zhou , Shuiping Gou , Gen Li","doi":"10.1016/j.bspc.2025.109436","DOIUrl":"10.1016/j.bspc.2025.109436","url":null,"abstract":"<div><div>Medical image classification technology holds significant clinical implications for the early screening, diagnosis, and treatment of diseases. However, most existing medical image classification models focus on single-organ analysis, which presents limitations in task generalization. Their scalability and broader applicability remain underexplored. This study aims to develop a generalizable multi-organ lesion classification framework to overcome the challenges of small lesion-to-background ratios, indistinct morphological boundaries, and heterogeneous lesion manifestations. This enables robust screening of pathological abnormalities across multiple abdominal organs.</div><div>We present OSRE-MLC, an innovative framework integrating two key components: (1) an organ-specific feature perception module that dynamically adapts to anatomical variations while preventing feature degradation, and (2) a region-specific enhancement module that optimizes discriminative lesion representation prior to classification. The framework integrates multi-organ abdominal CT images, including liver, kidney, and pancreas, from five separate datasets. The architecture uniquely combines multi-organ segmentation with attention-based feature refinement, enabling simultaneous organ localization and pathology characterization through an end-to-end trainable network.</div><div>Comprehensive evaluation on abdominal CT datasets has demonstrated OSRE-MLC’s superior performance, achieving 95.0% accuracy, 94.44% F1-score, 95.0% precision, and 93.96% recall in liver, kidney, and pancreas lesion screening, significantly outperforming existing methods. The proposed framework establishes a new paradigm for multi-organ pathological analysis by effectively addressing feature degradation and inter-organ variability. Its clinically interpretable architecture and robust performance demonstrate significant potential for improving diagnostic accuracy in complex abdominal imaging, offering promising applications in precision medicine and computer-aided diagnosis systems.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109436"},"PeriodicalIF":4.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-24DOI: 10.1016/j.bspc.2025.109430
Yazhou Li , Kairu Li , Xiaoxin Wang , Yixuan Sheng
Gesture recognition based on surface electromyography (sEMG) has emerged as a promising approach for human–machine interaction systems, particularly in applications such as prosthetic hand control. Nevertheless, achieving an optimal balance between computational complexity and classification accuracy remains a persistent challenge for recognition networks. Thus, this paper proposes a multi-domain feature fusion (MDFF) methodology coupled with a lightweight Vanilla network (LVNet) to reduce computational demands whilst maintaining satisfactory classification performance, thereby enabling its direct deployment on terminal devices with limited computing resources, such as laptops or intelligent prosthetic hands. The proposed MDFF-LVNet model establishes an end-to-end fully convolutional classification architecture: the MDFF extracts and fuses time-domain and time–frequency domain features, and the LVNet improves inference speed and recognition capability by using dynamic convolution and activation functions to train dual convolutional operations. Experimental results demonstrate that the MDFF-LVNet model achieves classification accuracies of 95.78%, 92.77%, floating-point operations of 4.52 GFLOPs and 8.05 GFLOPs, and inference time of 18.11 ms and 45.69 ms on public gesture datasets NinaPro DB2 and DB5, respectively. To evaluate its online recognition performance, experiments conducted on a bionic prosthetic hand using a self-constructed sEMG dataset of 6 gestures achieved an offline accuracy of 99.57%.
{"title":"sEMG-based gesture recognition using multi-domain feature fusion with a lightweight Vanilla network","authors":"Yazhou Li , Kairu Li , Xiaoxin Wang , Yixuan Sheng","doi":"10.1016/j.bspc.2025.109430","DOIUrl":"10.1016/j.bspc.2025.109430","url":null,"abstract":"<div><div>Gesture recognition based on surface electromyography (sEMG) has emerged as a promising approach for human–machine interaction systems, particularly in applications such as prosthetic hand control. Nevertheless, achieving an optimal balance between computational complexity and classification accuracy remains a persistent challenge for recognition networks. Thus, this paper proposes a multi-domain feature fusion (MDFF) methodology coupled with a lightweight Vanilla network (LVNet) to reduce computational demands whilst maintaining satisfactory classification performance, thereby enabling its direct deployment on terminal devices with limited computing resources, such as laptops or intelligent prosthetic hands. The proposed MDFF-LVNet model establishes an end-to-end fully convolutional classification architecture: the MDFF extracts and fuses time-domain and time–frequency domain features, and the LVNet improves inference speed and recognition capability by using dynamic convolution and activation functions to train dual convolutional operations. Experimental results demonstrate that the MDFF-LVNet model achieves classification accuracies of 95.78%, 92.77%, floating-point operations of 4.52 GFLOPs and 8.05 GFLOPs, and inference time of 18.11 ms and 45.69 ms on public gesture datasets NinaPro DB2 and DB5, respectively. To evaluate its online recognition performance, experiments conducted on a bionic prosthetic hand using a self-constructed sEMG dataset of 6 gestures achieved an offline accuracy of 99.57%.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109430"},"PeriodicalIF":4.9,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841373","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-24DOI: 10.1016/j.bspc.2025.109369
Feng Wu , Enhao Wang , Binqiang Xue , Yinhua Liu
Emotion recognition based on regional electroencephalography (EEG) signals can effectively mitigate practical deployment barriers, but limited data from sparse channels constrains the comprehensive representation of emotional states. This paper proposes an emotion recognition method based on functional connectivity of local brain regions. By simulating dynamic interactions in brain networks under emotional states, the method effectively enhances the representation of local information, which in turn improves recognition performance. Firstly, we introduce a cross-temporal connectivity feature modeling method based on spatial brain connectivity, and feature boosting maps are constructed by computing linear and nonlinear dependencies to improve connectivity representation. Secondly, Spatio-temporal Attention Transformer and Convolutional Neural Network are employed for encoding global and local temporal features, respectively. A feature fusion block is designed to integrate features from two encoders, fully leveraging their complementarity. Finally, the fused features are passed through a fully connected layer and subsequently fed into a softmax classifier for emotion classification. We conducted various experiments on two public EEG emotion datasets, SEED and DEAP. The results demonstrate that our method effectively captures emotional information from local brain regions, achieving significant recognition performance.
{"title":"Emotion recognition based on spatio-temporal connectivity of prefrontal EEG signals","authors":"Feng Wu , Enhao Wang , Binqiang Xue , Yinhua Liu","doi":"10.1016/j.bspc.2025.109369","DOIUrl":"10.1016/j.bspc.2025.109369","url":null,"abstract":"<div><div>Emotion recognition based on regional electroencephalography (EEG) signals can effectively mitigate practical deployment barriers, but limited data from sparse channels constrains the comprehensive representation of emotional states. This paper proposes an emotion recognition method based on functional connectivity of local brain regions. By simulating dynamic interactions in brain networks under emotional states, the method effectively enhances the representation of local information, which in turn improves recognition performance. Firstly, we introduce a cross-temporal connectivity feature modeling method based on spatial brain connectivity, and feature boosting maps are constructed by computing linear and nonlinear dependencies to improve connectivity representation. Secondly, Spatio-temporal Attention Transformer and Convolutional Neural Network are employed for encoding global and local temporal features, respectively. A feature fusion block is designed to integrate features from two encoders, fully leveraging their complementarity. Finally, the fused features are passed through a fully connected layer and subsequently fed into a softmax classifier for emotion classification. We conducted various experiments on two public EEG emotion datasets, SEED and DEAP. The results demonstrate that our method effectively captures emotional information from local brain regions, achieving significant recognition performance.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109369"},"PeriodicalIF":4.9,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-24DOI: 10.1016/j.bspc.2025.109389
M. Radhika , P. Sivakumar , D. Somasundaram , T. Sivakami
This paper presents a power efficient signal conversion and compression on Electroencephalogram (EEG) and Electrocardiogram (ECG) signals using a low power dual slope analog to digital converter (LDS-ADC) and hybrid discrete cosine transform with improved emperor penguin optimization (hybrid DCT-IEPO). Initially, ECG and EEG signals are acquired and converted into a digital signal using a power efficient LDS-ADC. After the process of energy efficient signal conversion, ECG and EEG signal compression is performed. Here, the hybrid discrete cosine transform (hybrid DCT) is applied to converted digital signals, resulting in the number of coefficients. Subsequently, optimal coefficients are selected by the emperor penguin optimization algorithm. Finally, inverse hybrid DCT reconstructs the signal with the selected optimal coefficients. Thus, the reconstructed power efficient output signal is attained and it is utilized for various applications. The outcome of the proposed work is examined with the prevailing methods compression ratio and it is implemented in the MATLAB. Experimental results show that the proposed technique obtained better performance in compression ratio, percentage root mean square error difference (PRD), quality score (QS), and mean square error (MSE).
{"title":"Power efficient signal conversion and quality signal compression using LDS-ADC and hybrid DCT for biomedical signals","authors":"M. Radhika , P. Sivakumar , D. Somasundaram , T. Sivakami","doi":"10.1016/j.bspc.2025.109389","DOIUrl":"10.1016/j.bspc.2025.109389","url":null,"abstract":"<div><div>This paper presents a power efficient signal conversion and compression on Electroencephalogram (EEG) and Electrocardiogram (ECG) signals using a low power dual slope analog to digital converter (LDS-ADC) and hybrid discrete cosine transform with improved emperor penguin optimization (hybrid DCT-IEPO). Initially, ECG and EEG signals are acquired and converted into a digital signal using a power efficient LDS-ADC. After the process of energy efficient signal conversion, ECG and EEG signal compression is performed. Here, the hybrid discrete cosine transform (hybrid DCT) is applied to converted digital signals, resulting in the number of coefficients. Subsequently, optimal coefficients are selected by the emperor penguin optimization algorithm. Finally, inverse hybrid DCT reconstructs the signal with the selected optimal coefficients. Thus, the reconstructed power efficient output signal is attained and it is utilized for various applications. The outcome of the proposed work is examined with the prevailing methods compression ratio and it is implemented in the MATLAB. Experimental results show that the proposed technique obtained better performance in compression ratio, percentage root mean square error difference (PRD), quality score (QS), and mean square error (MSE).</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109389"},"PeriodicalIF":4.9,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-23DOI: 10.1016/j.bspc.2025.109427
Mostafa M. Abdel-Aziz , Nabil A. Lashin , Hanaa M. Hamza , Khalid M. Hosny
This study introduces a new data-hiding method that combines fractal box encryption (FBE), deep learning-driven feature extraction, and a modified embedding strategy utilizing the spread spectrum method. This approach enables the secure incorporation and extraction of hidden data within medical images. The procedure commences with preliminary preprocessing of the host image, followed by morphological restoration to augment its structural attributes. A pre-trained ResNet-50 model is utilized to extract sophisticated image features, which are then encrypted with fractal box encryption to augment security. The secret image is then integrated into the encrypted feature vector using a spread spectrum embedding technique, ensuring that the concealed data remains robust against typical image processing attacks. During decryption, the obscured image is effectively retrieved by associating the watermarked characteristics with a noise pattern. This method ensures secure data concealment through fractal encryption while preserving the integrity of the concealed image, rendering it tamper-resistant. The proposed method offers a robust and efficient solution for data concealment and validation in medical imaging, where the protection of integrity and confidentiality is paramount.
{"title":"Enhanced security of medical images through fractal box encryption and CNN-driven data hiding method","authors":"Mostafa M. Abdel-Aziz , Nabil A. Lashin , Hanaa M. Hamza , Khalid M. Hosny","doi":"10.1016/j.bspc.2025.109427","DOIUrl":"10.1016/j.bspc.2025.109427","url":null,"abstract":"<div><div>This study introduces a new data-hiding method that combines fractal box encryption (FBE), deep learning-driven feature extraction, and a modified embedding strategy utilizing the spread spectrum method. This approach enables the secure incorporation and extraction of hidden data within medical images. The procedure commences with preliminary preprocessing of the host image, followed by morphological restoration to augment its structural attributes. A pre-trained ResNet-50 model is utilized to extract sophisticated image features, which are then encrypted with fractal box encryption to augment security. The secret image is then integrated into the encrypted feature vector using a spread spectrum embedding technique, ensuring that the concealed data remains robust against typical image processing attacks. During decryption, the obscured image is effectively retrieved by associating the watermarked characteristics with a noise pattern. This method ensures secure data concealment through fractal encryption while preserving the integrity of the concealed image, rendering it tamper-resistant. The proposed method offers a robust and efficient solution for data concealment and validation in medical imaging, where the protection of integrity and confidentiality is paramount.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109427"},"PeriodicalIF":4.9,"publicationDate":"2025-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-22DOI: 10.1016/j.bspc.2025.109444
Mariana Castro-Montano , Andy Petros , Ling Li , Enayetur Rahman , Simon Hannam , Grant Clow , Panayiotis A Kyriacou , Jim McLaughlin , Meha Qassem
Generalised oedema is common in neonatal intensive care units (NICUs), particularly in preterm and low-birth-weight infants. Characterised by tissue swelling from excess water accumulation, it can reflect systematic illness such as congestive heart failure, hepatic cirrhosis, nephrotic syndrome, sepsis, and acute kidney injury. Current clinical assessment methods, including formulas based on weight and fluid input/output and visual skin observation, lack accuracy and sensitivity, especially in critically ill infants. Techniques such as bioimpedance and ultrasound have been explored but are unsuitable for neonates and do not provide direct water content measurements. Spectroscopy, a non-invasive optical method, offers a promising solution by measuring tissue water content through light interactions in the Near Infrared (NIR) spectrum. This study investigates oedema in neonates using an NIR hyperspectral system in the NICU. Data was collected from 20 neonates, both with and without oedema over the course of three consecutive days. Spectral analysis revealed significant differences, notably at water absorption peaks around 1200 nm (p = 0.012). A Partial Least Squares Discriminant Analysis (PLS-DA) model effectively differentiated between oedematous and non-oedematous infants using spectral and standard clinical features, achieving 85.56 % recall and 100 % precision in testing. These findings suggest NIR spectroscopy combined with PLS-DA offers a reliable, non-contact method for early oedema detection in neonates, potentially enhancing monitoring and outcomes in the NICU.
{"title":"Generalised oedema monitoring utilising a NIR hyperspectral camera in critically ill neonates: A feasibility study","authors":"Mariana Castro-Montano , Andy Petros , Ling Li , Enayetur Rahman , Simon Hannam , Grant Clow , Panayiotis A Kyriacou , Jim McLaughlin , Meha Qassem","doi":"10.1016/j.bspc.2025.109444","DOIUrl":"10.1016/j.bspc.2025.109444","url":null,"abstract":"<div><div>Generalised oedema is common in neonatal intensive care units (NICUs), particularly in preterm and low-birth-weight infants. Characterised by tissue swelling from excess water accumulation, it can reflect systematic illness such as congestive heart failure, hepatic cirrhosis, nephrotic syndrome, sepsis, and acute kidney injury. Current clinical assessment methods, including formulas based on weight and fluid input/output and visual skin observation, lack accuracy and sensitivity, especially in critically ill infants. Techniques such as bioimpedance and ultrasound have been explored but are unsuitable for neonates and do not provide direct water content measurements. Spectroscopy, a non-invasive optical method, offers a promising solution by measuring tissue water content through light interactions in the Near Infrared (NIR) spectrum. This study investigates oedema in neonates using an NIR hyperspectral system in the NICU. Data was collected from 20 neonates, both with and without oedema over the course of three consecutive days. Spectral analysis revealed significant differences, notably at water absorption peaks around 1200 nm (p = 0.012). A Partial Least Squares Discriminant Analysis (PLS-DA) model effectively differentiated between oedematous and non-oedematous infants using spectral and standard clinical features, achieving 85.56 % recall and 100 % precision in testing. These findings suggest NIR spectroscopy combined with PLS-DA offers a reliable, non-contact method for early oedema detection in neonates, potentially enhancing monitoring and outcomes in the NICU.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"115 ","pages":"Article 109444"},"PeriodicalIF":4.9,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145841371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}