Pub Date : 2026-02-07DOI: 10.1016/j.bspc.2026.109739
Yingfa Li , Jialin Shi , Yufei Wang , Jiping Wei , Yaru Wei , Liang Wu , Meihao Wang , Zhifang Pan
Integrating radiology and histopathology images provides critical complementary perspectives for cancer survival prediction. However, current research faces two main challenges: (1) significant discrepancies in spatial scale and feature dimensionality between modalities; and (2) limited clinical generalizability due to existing methods being restricted to single cancer types or tasks. To overcome these barriers, we propose the Dual-Branch Multimodal Attention-based Feature Fusion Network (DBMAF). This framework employs an enhanced multi-scale channel attention mechanism for intra-modal feature extraction and an attention-guided cross-modal module to learn discriminative correlations between modalities. We validated DBMAF on four cancer cohorts, comprising three public datasets (TCGA-OV, TCGA-KIRC, TCGA-LIHC) and one private institutional dataset (WMU-CRC). Quantitative evaluations demonstrate that our method consistently outperforms all compared methods, achieving a maximum C-index of 0.910 on the TCGA-LIHC cohort. Furthermore, DBMAF showed robust performance across multiple survival endpoints (OS, TTP, and PFS) on the TCGA-OV dataset, highlighting its clinical utility for precise treatment stratification.
{"title":"DBMAF: Dual-branch multimodal attention-based feature fusion network for fusing histopathology and radiology images","authors":"Yingfa Li , Jialin Shi , Yufei Wang , Jiping Wei , Yaru Wei , Liang Wu , Meihao Wang , Zhifang Pan","doi":"10.1016/j.bspc.2026.109739","DOIUrl":"10.1016/j.bspc.2026.109739","url":null,"abstract":"<div><div>Integrating radiology and histopathology images provides critical complementary perspectives for cancer survival prediction. However, current research faces two main challenges: (1) significant discrepancies in spatial scale and feature dimensionality between modalities; and (2) limited clinical generalizability due to existing methods being restricted to single cancer types or tasks. To overcome these barriers, we propose the Dual-Branch Multimodal Attention-based Feature Fusion Network (DBMAF). This framework employs an enhanced multi-scale channel attention mechanism for intra-modal feature extraction and an attention-guided cross-modal module to learn discriminative correlations between modalities. We validated DBMAF on four cancer cohorts, comprising three public datasets (TCGA-OV, TCGA-KIRC, TCGA-LIHC) and one private institutional dataset (WMU-CRC). Quantitative evaluations demonstrate that our method consistently outperforms all compared methods, achieving a maximum C-index of 0.910 on the TCGA-LIHC cohort. Furthermore, DBMAF showed robust performance across multiple survival endpoints (OS, TTP, and PFS) on the TCGA-OV dataset, highlighting its clinical utility for precise treatment stratification.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109739"},"PeriodicalIF":4.9,"publicationDate":"2026-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146191910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-07DOI: 10.1016/j.bspc.2026.109754
Qingfeng Tang , Huihui Hu , Chao Tao , Pengcheng Ding , Guowei Dai , Guangjun Wang , Xiaojuan Hu , Benyue Su , Jiatuo Xu , Hui An
Although concatenating knowledge features (KF) and data features (DF) of photoplethysmography (PPG) can improve the predictive performance of blood pressure monitoring models, this approach inevitably increases the dimensionality of feature space. To address this limitation, we propose an innovative feature extraction method that deeply integrate KF and DF, rather than simply concatenating them.
Our method employs functional data analysis to extract DF by treating PPG as continuously functional curve. Subsequently, the distribution patterns of KF are thoroughly analyzed to construct a KF-based constrained space, which serves as a guide during DF extraction, to yield novel data-knowledge features (DKF).
The experimental results on blood pressure prediction showed that, without the need for additional dimensions, 9-dimensional DKF delivered superior predictive performance compared to both 9-dimensional DF and 8-dimensional KF. Specifically, for systolic blood pressure prediction, DKF reduces the mean absolute error (MAE) to 11.41, outperforming KF (MAE=12.11) and DF (MAE=13.24). Similarly, for diastolic blood pressure, DKF achieves an MAE of 7.27, lower than that of KF (7.41) and DF (7.84).
The proposed feature extraction method effectively overcomes the drawbacks of feature concatenation, offering a novel and effective approach to extracting low-dimensional, highly discriminative features from PPG for accurate blood pressure estimation.
{"title":"Data-knowledge feature fusion for PPG-based blood pressure prediction: Low-dimensional extraction via functional data analysis and knowledge constraint","authors":"Qingfeng Tang , Huihui Hu , Chao Tao , Pengcheng Ding , Guowei Dai , Guangjun Wang , Xiaojuan Hu , Benyue Su , Jiatuo Xu , Hui An","doi":"10.1016/j.bspc.2026.109754","DOIUrl":"10.1016/j.bspc.2026.109754","url":null,"abstract":"<div><div>Although concatenating knowledge features (KF) and data features (DF) of photoplethysmography (PPG) can improve the predictive performance of blood pressure monitoring models, this approach inevitably increases the dimensionality of feature space. To address this limitation, we propose an innovative feature extraction method that deeply integrate KF and DF, rather than simply concatenating them.</div><div>Our method employs functional data analysis to extract DF by treating PPG as continuously functional curve. Subsequently, the distribution patterns of KF are thoroughly analyzed to construct a KF-based constrained space, which serves as a guide during DF extraction, to yield novel data-knowledge features (DKF).</div><div>The experimental results on blood pressure prediction showed that, without the need for additional dimensions, 9-dimensional DKF delivered superior predictive performance compared to both 9-dimensional DF and 8-dimensional KF. Specifically, for systolic blood pressure prediction, DKF reduces the mean absolute error (MAE) to 11.41, outperforming KF (MAE=12.11) and DF (MAE=13.24). Similarly, for diastolic blood pressure, DKF achieves an MAE of 7.27, lower than that of KF (7.41) and DF (7.84).</div><div>The proposed feature extraction method effectively overcomes the drawbacks of feature concatenation, offering a novel and effective approach to extracting low-dimensional, highly discriminative features from PPG for accurate blood pressure estimation.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109754"},"PeriodicalIF":4.9,"publicationDate":"2026-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146191946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109714
Kun Wu , Xiang Chen , Nina Cheng , Nishant Ravikumar , Alejandro F. Frangi
Objective:
Unsupervised cardiac motion estimation often confronts complex scenarios, which include a lack of explicit reference for the deformation fields, and intramodal anatomical gaps. These factors introduce substantial obstacles in the effective representation of both smooth and accurate cardiac motion, thereby hindering the prediction of intricate structural details after registration. However, existing approaches have not sufficiently explored the explicit spatial correlations encompassed in multi-range displacements.
Methods:
To overcome the challenges, we propose a novel multi-attention-guided network, MAPC-Net, a pyramidal network with spatial correlation normalisation compensated in the mechanism for high-quality cardiac motion estimation.
Results:
The extensive experimental results from quantitative and qualitative aspects indicate that MAPC-Net achieves exceptional performance in the generalisation of the effective deformation field on the private dataset UKBiobank and publicly available ACDC in terms of cardiac cine-MRI datasets. Our model achieves an average Dice score over 75 (77.2), a 95 - Hausdorff Distance less than 4.50 mm and a Negative Jacobian Determinant value of 0.20 without segmentation label guided over UKBioBank dataset. We highlighted the significant potential of the proposed framework in clinical relevance by demonstrating the downstream analysis in terms of cardiac peak strain signal. We improve the estimated peak radial strain value from 41.72 to 43.28.
Conclusion:
A novel framework was proposed for the refinement of motion estimation by introducing attention-guided correlation between warped and fixed frames.
Significance:
The architecture of our proposed model offers a new solution for predicting high-quality cardiac deformation fields, leveraging an attention-aware cost volume calculation embedded in a pyramidal network for motion estimation.
{"title":"Multi-attention-aware motion estimation for cardiac MR imaging based on a feature pyramid network","authors":"Kun Wu , Xiang Chen , Nina Cheng , Nishant Ravikumar , Alejandro F. Frangi","doi":"10.1016/j.bspc.2026.109714","DOIUrl":"10.1016/j.bspc.2026.109714","url":null,"abstract":"<div><h3>Objective:</h3><div>Unsupervised cardiac motion estimation often confronts complex scenarios, which include a lack of explicit reference for the deformation fields, and intramodal anatomical gaps. These factors introduce substantial obstacles in the effective representation of both smooth and accurate cardiac motion, thereby hindering the prediction of intricate structural details after registration. However, existing approaches have not sufficiently explored the explicit spatial correlations encompassed in multi-range displacements.</div></div><div><h3>Methods:</h3><div>To overcome the challenges, we propose a novel multi-attention-guided network, MAPC-Net, a pyramidal network with spatial correlation normalisation compensated in the mechanism for high-quality cardiac motion estimation.</div></div><div><h3>Results:</h3><div>The extensive experimental results from quantitative and qualitative aspects indicate that MAPC-Net achieves exceptional performance in the generalisation of the effective deformation field on the private dataset UKBiobank and publicly available ACDC in terms of cardiac cine-MRI datasets. Our model achieves an average Dice score over 75<span><math><mtext>%</mtext></math></span> (77.2<span><math><mtext>%</mtext></math></span>), a 95<span><math><mtext>%</mtext></math></span> - Hausdorff Distance less than 4.50 mm and a Negative Jacobian Determinant value of 0.20<span><math><mtext>%</mtext></math></span> without segmentation label guided over UKBioBank dataset. We highlighted the significant potential of the proposed framework in clinical relevance by demonstrating the downstream analysis in terms of cardiac peak strain signal. We improve the estimated peak radial strain value from 41.72<span><math><mtext>%</mtext></math></span> to 43.28<span><math><mtext>%</mtext></math></span>.</div></div><div><h3>Conclusion:</h3><div>A novel framework was proposed for the refinement of motion estimation by introducing attention-guided correlation between warped and fixed frames.</div></div><div><h3>Significance:</h3><div>The architecture of our proposed model offers a new solution for predicting high-quality cardiac deformation fields, leveraging an attention-aware cost volume calculation embedded in a pyramidal network for motion estimation.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109714"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109746
Jie Zhou , Li Wang , Fengji Li , Shaochuan Zhang , Fei Shen , Fan Fan , Tao Liu , Xiaohong Chen , Haijun Niu
The application of Electrolarynx (EL) for tonal language laryngectomees remains challenging due to the difficulty in achieving tonal completion without useful fundamental frequency (F0) information. This study proposes a novel Mandarin EL speech enhancement framework by integrating the prior F0 information provided by finger movements, combined with the Cycle-Consistent Adversarial Network (CycleGAN) and Continuous Wavelet Transform (CWT). For prosody modeling, we exploit the hierarchical structure inherent in Mandarin prosody by using CWT decomposition coefficients as a feature representation of F0. For spectral conversion, we extract Mel-frequency cepstral coefficients (MCEP) as spectral features. These two feature sets were trained separately using the CycleGAN model. In results, acoustic feature analysis indicates that the four tones after converted are closer to normal tones in both F0 value and F0 contour. The spectrogram of the converted speech is also more similar to that of normal speech, and compensates for low-frequency energy missing below 500 Hz. Both subjective and objective evaluations demonstrate the effectiveness of the proposed method in Mandarin EL speech enhancement. This study also provides a novel approach for EL speech enhancement in other tonal languages. And it may provide valuable insights and guidance for future improvement in tonal EL devices development and EL speech enhancement.
{"title":"CycleGAN-based prosody and spectrum modeling for Mandarin touch-controlled Electrolaryngeal speech enhancement","authors":"Jie Zhou , Li Wang , Fengji Li , Shaochuan Zhang , Fei Shen , Fan Fan , Tao Liu , Xiaohong Chen , Haijun Niu","doi":"10.1016/j.bspc.2026.109746","DOIUrl":"10.1016/j.bspc.2026.109746","url":null,"abstract":"<div><div>The application of Electrolarynx (EL) for tonal language laryngectomees remains challenging due to the difficulty in achieving tonal completion without useful fundamental frequency (F0) information. This study proposes a novel Mandarin EL speech enhancement framework by integrating the prior F0 information provided by finger movements, combined with the Cycle-Consistent Adversarial Network (CycleGAN) and Continuous Wavelet Transform (CWT). For prosody modeling, we exploit the hierarchical structure inherent in Mandarin prosody by using CWT decomposition coefficients as a feature representation of F0. For spectral conversion, we extract Mel-frequency cepstral coefficients (MCEP) as spectral features. These two feature sets were trained separately using the CycleGAN model. In results, acoustic feature analysis indicates that the four tones after converted are closer to normal tones in both F0 value and F0 contour. The spectrogram of the converted speech is also more similar to that of normal speech, and compensates for low-frequency energy missing below 500 Hz. Both subjective and objective evaluations demonstrate the effectiveness of the proposed method in Mandarin EL speech enhancement. This study also provides a novel approach for EL speech enhancement in other tonal languages. And it may provide valuable insights and guidance for future improvement in tonal EL devices development and EL speech enhancement.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109746"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109685
Yicai Bai , Yucheng Zhou , Jinqi Dong , Dengjiujiu He , Chao Jiang , Jinglu Hu , Yingjie Li
Emotion is fundamental to human cognition and behavior, and electroencephalography (EEG), with its high temporal resolution, provides a powerful approach to investigate the neural activity underlying emotional processing. We combined EEG with Virtual Reality technology to conduct an EEG-based emotion experiment in a more immersive environment. Moreover, EEG microstate and functional connectivity features are closely related yet capturing their complex nonlinear interactions remains challenging. To address this challenge, we proposed a deep learning framework that integrates Cross-Attention mechanisms with convolutional neural networks (CNN) to model these interactions. Specifically, Cross-Attention captures inter-feature dependencies, while CNN performs hierarchical feature extraction. Experimental results demonstrate that our framework significantly outperforms the baseline CNN model, particularly in recognizing subtle emotional states such as neutral emotion. Notably, this improvement is driven by the synergistic interaction between microstate and functional connectivity features, thereby improving model interpretability. These findings highlight the potential of Cross-Attention CNN to elucidate the complex nonlinear neural mechanisms underlying emotional processing.
{"title":"Integrating EEG microstates and functional connectivity via cross-attention for emotion recognition in virtual reality","authors":"Yicai Bai , Yucheng Zhou , Jinqi Dong , Dengjiujiu He , Chao Jiang , Jinglu Hu , Yingjie Li","doi":"10.1016/j.bspc.2026.109685","DOIUrl":"10.1016/j.bspc.2026.109685","url":null,"abstract":"<div><div>Emotion is fundamental to human cognition and behavior, and electroencephalography (EEG), with its high temporal resolution, provides a powerful approach to investigate the neural activity underlying emotional processing. We combined EEG with Virtual Reality technology to conduct an EEG-based emotion experiment in a more immersive environment. Moreover, EEG microstate and functional connectivity features are closely related yet capturing their complex nonlinear interactions remains challenging. To address this challenge, we proposed a deep learning framework that integrates Cross-Attention mechanisms with convolutional neural networks (CNN) to model these interactions. Specifically, Cross-Attention captures inter-feature dependencies, while CNN performs hierarchical feature extraction. Experimental results demonstrate that our framework significantly outperforms the baseline CNN model, particularly in recognizing subtle emotional states such as neutral emotion. Notably, this improvement is driven by the synergistic interaction between microstate and functional connectivity features, thereby improving model interpretability. These findings highlight the potential of Cross-Attention CNN to elucidate the complex nonlinear neural mechanisms underlying emotional processing.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109685"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109736
Huanqing Zhang , Jun Xie , Kaixuan Liu , Yan Liu , Wenxiang Dong , Guanghua Xu
Steady state motion auditory evoked potential (SSMAEP) is neural responses elicited by rhythmic auditory stimuli with periodic spatial motion. SSMAEP brain computer interface (BCI) relies on auditory selective attention to decode user intent in multi-source environments. However, the complex temporal and spectral structure of SSMAEP presents challenges for effective feature extraction from electroencephalogram (EEG). Time-frequency transforms are suited for extracting the joint time–frequency features of SSMAEP. Notably, these transforms share structural similarities with convolution operations in convolution neural networks. In this study, we propose a novel time frequency convolutional layer that incorporates structured kernels based on the S transform, continuous wavelet transform (CWT), and short-time Fourier transform (STFT). These time frequency kernels are embedded as learnable filters and replace the conventional first convolutional layer of ShallowConvNet. This design enables the model to more effectively capture SSMAEP signal dynamics across both time and frequency domains. The proposed method was evaluated on two SSMAEP-BCI datasets with two and three auditory targets. Experimental results demonstrate consistent improvements in classification accuracy and robustness compared to baseline models. Furthermore, analysis of the learned kernels revealed that the time frequency filters retained their interpretable structure after training, with task-relevant shifts in center frequency and bandwidth. These findings highlight not only the performance advantage but also the improved interpretability of the proposed model, offering insights into the spectral encoding of SSMAEP-BCI.
{"title":"Time frequency transform kernel enhanced ShallowConvNet for auditory selective attention decoding with steady state motion auditory evoked potential","authors":"Huanqing Zhang , Jun Xie , Kaixuan Liu , Yan Liu , Wenxiang Dong , Guanghua Xu","doi":"10.1016/j.bspc.2026.109736","DOIUrl":"10.1016/j.bspc.2026.109736","url":null,"abstract":"<div><div>Steady state motion auditory evoked potential (SSMAEP) is neural responses elicited by rhythmic auditory stimuli with periodic spatial motion. SSMAEP brain computer interface (BCI) relies on auditory selective attention to decode user intent in multi-source environments. However, the complex temporal and spectral structure of SSMAEP presents challenges for effective feature extraction from electroencephalogram (EEG). Time-frequency transforms are suited for extracting the joint time–frequency features of SSMAEP. Notably, these transforms share structural similarities with convolution operations in convolution neural networks. In this study, we propose a novel time frequency convolutional layer that incorporates structured kernels based on the S transform, continuous wavelet transform (CWT), and short-time Fourier transform (STFT). These time frequency kernels are embedded as learnable filters and replace the conventional first convolutional layer of ShallowConvNet. This design enables the model to more effectively capture SSMAEP signal dynamics across both time and frequency domains. The proposed method was evaluated on two SSMAEP-BCI datasets with two and three auditory targets. Experimental results demonstrate consistent improvements in classification accuracy and robustness compared to baseline models. Furthermore, analysis of the learned kernels revealed that the time frequency filters retained their interpretable structure after training, with task-relevant shifts in center frequency and bandwidth. These findings highlight not only the performance advantage but also the improved interpretability of the proposed model, offering insights into the spectral encoding of SSMAEP-BCI.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109736"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109792
Altaf Hussain , Shuaiyong Li , Tariq Hussain
Wireless Body Area Networks (WBANs) support continuous patient monitoring in clinical and remote settings by enabling low-power sensors to collect and forward physiological data. However, WBAN deployments are constrained by limited battery capacity and challenging on-body propagation, which increase path loss, degrade link reliability, and shorten network lifetime. Moreover, many existing routing/clustering solutions treat these issues separately rather than jointly. To address this, we propose MacroNet-Enhanced Energy-aware Node Clustering Protocol (MEE-NCP), which integrates dual energy-efficiency models with a MacroNet-based clustering design using four cluster heads (CHs). MEE-NCP forwards data to the coordinator node through CHs using a cost function that prioritizes higher residual energy and shorter distance to balance energy consumption and improve delivery reliability. We evaluate MEE-NCP in MATLAB against representative WBAN routing schemes using Packet Error Rate (PER), Packet Generation Rate (PGR), Data Generation Rate (DGR), RSSI, SNR, residual energy, throughput, end-to-end delay, and network lifetime. Simulation results indicate improved energy distribution and lifetime, reduced delay and packet errors, and stronger link quality via better path-loss handling.
{"title":"MacroNet-enhanced energy-aware node clustering protocol for wireless body area networks","authors":"Altaf Hussain , Shuaiyong Li , Tariq Hussain","doi":"10.1016/j.bspc.2026.109792","DOIUrl":"10.1016/j.bspc.2026.109792","url":null,"abstract":"<div><div>Wireless Body Area Networks (WBANs) support continuous patient monitoring in clinical and remote settings by enabling low-power sensors to collect and forward physiological data. However, WBAN deployments are constrained by limited battery capacity and challenging on-body propagation, which increase path loss, degrade link reliability, and shorten network lifetime. Moreover, many existing routing/clustering solutions treat these issues separately rather than jointly. To address this, we propose MacroNet-Enhanced Energy-aware Node Clustering Protocol (MEE-NCP), which integrates dual energy-efficiency models with a MacroNet-based clustering design using four cluster heads (CHs). MEE-NCP forwards data to the coordinator node through CHs using a cost function that prioritizes higher residual energy and shorter distance to balance energy consumption and improve delivery reliability. We evaluate MEE-NCP in MATLAB against representative WBAN routing schemes using Packet Error Rate (PER), Packet Generation Rate (PGR), Data Generation Rate (DGR), RSSI, SNR, residual energy, throughput, end-to-end delay, and network lifetime. Simulation results indicate improved energy distribution and lifetime, reduced delay and packet errors, and stronger link quality via better path-loss handling.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"118 ","pages":"Article 109792"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109795
Jiancheng Han , Heqing Wang , Yifan Feng , Qi Yang , Jingtan Li , Haojie Zhang , Yihua He , Jiang Liu , Toru Nakamura , Yang Cao , Naidi Sun , Kun Qian , Bin Hu , Xinru Gao , Yan Xia , Zongjie Weng , Björn W. Schuller , Yoshiharu Yamamoto
This study presents the first application of federated learning (FL) for prenatal detection of Interrupted Aortic Arch (IAA) using fetal ultrasound images. To address the challenges of data scarcity, privacy constraints, and inter-institutional variability, we develop a federated learning IAA detection method and systematically evaluate three representative strategies (FedAvg, FedProx, and FedBABU) across five clinical centres. Results show that FL improves model performance over local training in recall and F1-score in data-scarce centres. Among FL algorithms, FedAvg and FedProx consistently outperform FedBABU in stability and generalisation. Among the three CNN architectures compared — ResNet-50, EfficientNet-B3, and DenseNet-121 — DenseNet-121 demonstrates superior overall performance, particularly in non-independent and identically distributed (Non-IID) scenarios. Our framework demonstrates the feasibility of collaborative AI for rare disease detection without data sharing, laying the foundation for scalable, real-world prenatal screening of congenital heart defects.
{"title":"Federated learning for prenatal detection of interrupted aortic arch using fetal ultrasound imaging","authors":"Jiancheng Han , Heqing Wang , Yifan Feng , Qi Yang , Jingtan Li , Haojie Zhang , Yihua He , Jiang Liu , Toru Nakamura , Yang Cao , Naidi Sun , Kun Qian , Bin Hu , Xinru Gao , Yan Xia , Zongjie Weng , Björn W. Schuller , Yoshiharu Yamamoto","doi":"10.1016/j.bspc.2026.109795","DOIUrl":"10.1016/j.bspc.2026.109795","url":null,"abstract":"<div><div>This study presents the first application of federated learning (FL) for prenatal detection of Interrupted Aortic Arch (IAA) using fetal ultrasound images. To address the challenges of data scarcity, privacy constraints, and inter-institutional variability, we develop a federated learning IAA detection method and systematically evaluate three representative strategies (FedAvg, FedProx, and FedBABU) across five clinical centres. Results show that FL improves model performance over local training in recall and F1-score in data-scarce centres. Among FL algorithms, FedAvg and FedProx consistently outperform FedBABU in stability and generalisation. Among the three CNN architectures compared — ResNet-50, EfficientNet-B3, and DenseNet-121 — DenseNet-121 demonstrates superior overall performance, particularly in non-independent and identically distributed (Non-IID) scenarios. Our framework demonstrates the feasibility of collaborative AI for rare disease detection without data sharing, laying the foundation for scalable, real-world prenatal screening of congenital heart defects.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109795"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109726
Yi Wang , Pei Deng , Tinghui Zheng , Haoyao Cao
Automatic and accurate segmentation of coronary arteries (CA) is a prerequisite for high-precision reconstruction of three-dimensional CA models. However, the complex structure of CA, including low contrast, significant variation in vessel diameter, and high curvature, poses significant challenges for segmentation and reconstruction. In addition, coronary computed tomography angiography (CCTA) images contain abundant background information (such as other tissues, organs, or vessels), further increasing the difficulty of segmentation. These factors often lead to vessel discontinuity and incomplete segmentation. Therefore, accurate CA segmentation remains a challenging task. In this study, we propose the GMMA-Net network to improve the continuity, robustness, and noise resistance of CA segmentation. GMMA-Net employs a grouped multi-path feature fusion module (GMFFM) in the encoder to capture richer multi-scale feature information. Furthermore, by introducing a multi-scale attention module (MAM) into the bottleneck layer of GMMA-Net, we achieve dynamic weight adjustment, capture long-range dependencies, and suppress redundant features. Experimental results show that GMMA-Net outperforms existing methods in the task of CA segmentation from CCTA images, effectively overcoming challenges caused by scale sensitivity and noise interference. GMMA-Net demonstrates superior performance on metrics such as IoU, Dice coefficient, recall rate, and , especially exhibiting stronger segmentation capability when handling cases with poor image quality and large variations in vessel diameter. The code of the proposed method is available at https://github.com/DengPei-C/GMMA-Net.
{"title":"GMMA-Net: A CCTA image segmentation algorithm based on grouped multi-path feature fusion and multi-scale attention","authors":"Yi Wang , Pei Deng , Tinghui Zheng , Haoyao Cao","doi":"10.1016/j.bspc.2026.109726","DOIUrl":"10.1016/j.bspc.2026.109726","url":null,"abstract":"<div><div>Automatic and accurate segmentation of coronary arteries (CA) is a prerequisite for high-precision reconstruction of three-dimensional CA models. However, the complex structure of CA, including low contrast, significant variation in vessel diameter, and high curvature, poses significant challenges for segmentation and reconstruction. In addition, coronary computed tomography angiography (CCTA) images contain abundant background information (such as other tissues, organs, or vessels), further increasing the difficulty of segmentation. These factors often lead to vessel discontinuity and incomplete segmentation. Therefore, accurate CA segmentation remains a challenging task. In this study, we propose the GMMA-Net network to improve the continuity, robustness, and noise resistance of CA segmentation. GMMA-Net employs a grouped multi-path feature fusion module (GMFFM) in the encoder to capture richer multi-scale feature information. Furthermore, by introducing a multi-scale attention module (MAM) into the bottleneck layer of GMMA-Net, we achieve dynamic weight adjustment, capture long-range dependencies, and suppress redundant features. Experimental results show that GMMA-Net outperforms existing methods in the task of CA segmentation from CCTA images, effectively overcoming challenges caused by scale sensitivity and noise interference. GMMA-Net demonstrates superior performance on metrics such as IoU, Dice coefficient, recall rate, and <span><math><mrow><mi>H</mi><msub><mrow><mi>D</mi></mrow><mrow><mn>95</mn></mrow></msub></mrow></math></span>, especially exhibiting stronger segmentation capability when handling cases with poor image quality and large variations in vessel diameter. The code of the proposed method is available at <span><span>https://github.com/DengPei-C/GMMA-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109726"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-06DOI: 10.1016/j.bspc.2026.109639
Yuting Guo , Shuai Li , Wenfeng Song , Aimin Hao
Automated diagnostic report generation is a challenging vision-and-language bridging task aimed at accurately describing medical images and performing cross-modal causal inference. Despite its significant clinical importance, widespread application remains challenging. Existing methods often rely on pre-trained models with large-scale medical report datasets, leading to data shifts between training and testing sets, resulting in irrelevant contextual biases in the visual domain and correlation biases within the knowledge graph. To address these issues, we propose a novel multimodal causal inference approach called Multimodal Counterfactual Unbiased Report Generation (MCURG), which incorporates causal inference to exploit invariant rationales. Our key innovation lies in leveraging counterfactual inference to reduce visual and knowledge biases. MCURG employs a Structural Causal Model (SCM) to elucidate the complex relationships among images, knowledge graphs, reports, confounders, and personalized features. We design two multimodal debiasing modules: a visual debiasing module and a knowledge graph debiasing module. The visual debiasing module focuses on the Total Direct Effect of image features, mitigating confounding factors, while the knowledge graph debiasing module identifies individualized treatments within the graph, reducing spurious generations. We conducted extensive experiments and comprehensive evaluations on multiple datasets, demonstrating that MCURG effectively reduces bias and improves the accuracy of generated reports. This multimodal causal inference approach, through the use of SCM and counterfactual reasoning, successfully addresses bias in automated diagnostic report generation, marking a significant innovation in the field. The codes are available at https://github.com/stellating/MCURG.
{"title":"Unbiased diagnostic report generation via multi-modal counterfactual inference","authors":"Yuting Guo , Shuai Li , Wenfeng Song , Aimin Hao","doi":"10.1016/j.bspc.2026.109639","DOIUrl":"10.1016/j.bspc.2026.109639","url":null,"abstract":"<div><div>Automated diagnostic report generation is a challenging vision-and-language bridging task aimed at accurately describing medical images and performing cross-modal causal inference. Despite its significant clinical importance, widespread application remains challenging. Existing methods often rely on pre-trained models with large-scale medical report datasets, leading to data shifts between training and testing sets, resulting in irrelevant contextual biases in the visual domain and correlation biases within the knowledge graph. To address these issues, we propose a novel multimodal causal inference approach called Multimodal Counterfactual Unbiased Report Generation (MCURG), which incorporates causal inference to exploit invariant rationales. Our key innovation lies in leveraging counterfactual inference to reduce visual and knowledge biases. MCURG employs a Structural Causal Model (SCM) to elucidate the complex relationships among images, knowledge graphs, reports, confounders, and personalized features. We design two multimodal debiasing modules: a visual debiasing module and a knowledge graph debiasing module. The visual debiasing module focuses on the Total Direct Effect of image features, mitigating confounding factors, while the knowledge graph debiasing module identifies individualized treatments within the graph, reducing spurious generations. We conducted extensive experiments and comprehensive evaluations on multiple datasets, demonstrating that MCURG effectively reduces bias and improves the accuracy of generated reports. This multimodal causal inference approach, through the use of SCM and counterfactual reasoning, successfully addresses bias in automated diagnostic report generation, marking a significant innovation in the field. The codes are available at <span><span>https://github.com/stellating/MCURG</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55362,"journal":{"name":"Biomedical Signal Processing and Control","volume":"119 ","pages":"Article 109639"},"PeriodicalIF":4.9,"publicationDate":"2026-02-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146193097","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}