Magnetocardiography (MCG) enables passive detection of weak magnetic fields generated by the heart with high sensitivity, which can offer valuable information for diagnosing and treating heart conditions. Due to the limitations of the geomagnetic field and unknown magnetic interference, the MCG signals are often overwhelmed by high levels of magnetic noise. In this paper, we propose the design of a high-resolution and movable MCG system comprised of an active-passive coupling magnetic control (AP-CMC) system and a wearable multi-channel signal detection array. The system realizes the MCG measurement at the same time as the AP-CMC system eliminates interference in real time, i.e., simultaneous control and simultaneous measurement. Dynamic MCG signal measurements were successfully conducted, obtaining typical characteristic features of MCG signals. Our method shows promise in enhancing the accuracy and expanding the scope of MCG measurement applications, thereby offering valuable insights for the early diagnosis and precise localization of heart diseases.
{"title":"High-Resolution and Wearable Magnetocardiography (MCG) Measurement With Active-Passive Coupling Magnetic Control Method.","authors":"Shuai Dou, Xikai Liu, Pengfei Song, Yidi Cao, Tong Wen, Rui Feng, Bangcheng Han","doi":"10.1109/JBHI.2025.3584984","DOIUrl":"10.1109/JBHI.2025.3584984","url":null,"abstract":"<p><p>Magnetocardiography (MCG) enables passive detection of weak magnetic fields generated by the heart with high sensitivity, which can offer valuable information for diagnosing and treating heart conditions. Due to the limitations of the geomagnetic field and unknown magnetic interference, the MCG signals are often overwhelmed by high levels of magnetic noise. In this paper, we propose the design of a high-resolution and movable MCG system comprised of an active-passive coupling magnetic control (AP-CMC) system and a wearable multi-channel signal detection array. The system realizes the MCG measurement at the same time as the AP-CMC system eliminates interference in real time, i.e., simultaneous control and simultaneous measurement. Dynamic MCG signal measurements were successfully conducted, obtaining typical characteristic features of MCG signals. Our method shows promise in enhancing the accuracy and expanding the scope of MCG measurement applications, thereby offering valuable insights for the early diagnosis and precise localization of heart diseases.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1178-1186"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144540054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2025.3587639
Pengfei Wang, Danyang Li, Yaoduo Zhang, Gaofeng Chen, Yongbo Wang, Jianhua Ma, Ji He
Although supervised deep learning methods have made significant advances in low-dose computed tomography (LDCT) image denoising, these approaches typically require pairs of low-dose and normal-dose CT images for training, which are often unavailable in clinical settings. Self-supervised deep learning (SSDL) has great potential to cast off the dependence on paired training datasets. However, existing SSDL methods are limited by the neighboring noise independence assumptions, making them ineffective for handling spatially correlated noises in LDCT images. To address this issue, this paper introduces a novel SSDL approach, named, Noise-Aware Blind Spot Network (NA-BSN), for high-quality LDCT imaging, while mitigating the dependence on the assumption of neighboring noise independence. NA-BSN achieves high-quality image reconstruction without referencing clean data through its explicit noise-aware constraint mechanism during the self-supervised learning process. Specifically, it is experimentally observed and theoretical proven that the $l1$ norm value of CT images in a downsampled space follows a certain descend trend with increasing of the radiation dose, which is then used to construct the explicit noise-aware constraint in the architecture of BSN for self-supervised LDCT image denoising. Various clinical datasets are adopted to validate the performance of the presented NA-BSN method. Experimental results reveal that NA-BSN significantly reduces the spatially correlated CT noises and retains crucial image details in various complex scenarios, such as different types of scanning machines, scanning positions, dose-level settings, and reconstruction kernels.
{"title":"BSN With Explicit Noise-Aware Constraint for Self-Supervised Low-Dose CT Denoising.","authors":"Pengfei Wang, Danyang Li, Yaoduo Zhang, Gaofeng Chen, Yongbo Wang, Jianhua Ma, Ji He","doi":"10.1109/JBHI.2025.3587639","DOIUrl":"10.1109/JBHI.2025.3587639","url":null,"abstract":"<p><p>Although supervised deep learning methods have made significant advances in low-dose computed tomography (LDCT) image denoising, these approaches typically require pairs of low-dose and normal-dose CT images for training, which are often unavailable in clinical settings. Self-supervised deep learning (SSDL) has great potential to cast off the dependence on paired training datasets. However, existing SSDL methods are limited by the neighboring noise independence assumptions, making them ineffective for handling spatially correlated noises in LDCT images. To address this issue, this paper introduces a novel SSDL approach, named, Noise-Aware Blind Spot Network (NA-BSN), for high-quality LDCT imaging, while mitigating the dependence on the assumption of neighboring noise independence. NA-BSN achieves high-quality image reconstruction without referencing clean data through its explicit noise-aware constraint mechanism during the self-supervised learning process. Specifically, it is experimentally observed and theoretical proven that the $l1$ norm value of CT images in a downsampled space follows a certain descend trend with increasing of the radiation dose, which is then used to construct the explicit noise-aware constraint in the architecture of BSN for self-supervised LDCT image denoising. Various clinical datasets are adopted to validate the performance of the presented NA-BSN method. Experimental results reveal that NA-BSN significantly reduces the spatially correlated CT noises and retains crucial image details in various complex scenarios, such as different types of scanning machines, scanning positions, dose-level settings, and reconstruction kernels.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1286-1299"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144608189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2025.3636169
Yasaman Baradaran, Raul Fernandez Rojas, Roland Goecke, Maryam Ghahramani
The prefrontal cortex (PFC) of the brain is involved in processing visual, vestibular, and somatosensory inputs to stabilise postural balance. However, the PFC's activation map for a standing person and different sensory inputs remains unclear. This study aimed to explore the PFC activity map and distinct haemodynamic responses during postural control when sensory inputs change. To this end, functional near-infrared spectroscopy (fNIRS) was employed to capture the haemodynamic responses throughout the PFC from a group of young adults standing in four sensory conditions. The results revealed distinct PFC activation patterns supporting sensory processing, motor planning, and cognitive control to maintain balance under different degraded sensory conditions. Furthermore, by applying machine learning classifiers and multivariate feature selection, the PFC locations and haemodynamic responses indicative of different sensory conditions were identified. The findings of this study offer valuable insights for optimising rehabilitation approaches, enhancing the design of fNIRS studies, and advancing brain-computer interface technologies for balance assessment and training.
{"title":"Exploring Prefrontal Cortex Involvement in Postural Control Across Degraded Sensory Conditions Using fNIRS and Classification.","authors":"Yasaman Baradaran, Raul Fernandez Rojas, Roland Goecke, Maryam Ghahramani","doi":"10.1109/JBHI.2025.3636169","DOIUrl":"10.1109/JBHI.2025.3636169","url":null,"abstract":"<p><p>The prefrontal cortex (PFC) of the brain is involved in processing visual, vestibular, and somatosensory inputs to stabilise postural balance. However, the PFC's activation map for a standing person and different sensory inputs remains unclear. This study aimed to explore the PFC activity map and distinct haemodynamic responses during postural control when sensory inputs change. To this end, functional near-infrared spectroscopy (fNIRS) was employed to capture the haemodynamic responses throughout the PFC from a group of young adults standing in four sensory conditions. The results revealed distinct PFC activation patterns supporting sensory processing, motor planning, and cognitive control to maintain balance under different degraded sensory conditions. Furthermore, by applying machine learning classifiers and multivariate feature selection, the PFC locations and haemodynamic responses indicative of different sensory conditions were identified. The findings of this study offer valuable insights for optimising rehabilitation approaches, enhancing the design of fNIRS studies, and advancing brain-computer interface technologies for balance assessment and training.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1418-1428"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145596307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2024.3505955
Yizhen Luo, Jiahuan Zhang, Siqi Fan, Kai Yang, Massimo Hong, Yushuai Wu, Mu Qiao, Zaiqing Nie
Recent advances in large language models (LLMs) like ChatGPT have shed light on the development of knowledgeable and versatile AI research assistants in various scientific domains. However, they fall short in biomedical applications due to a lack of proprietary biomedical knowledge and deficiencies in handling biological sequences for molecules and proteins. To address these issues, we present BioMedGPT, a multimodal large language model for assisting biomedical research. We first incorporate domain expertise into LLMs by incremental pre-training on large-scale biomedical literature. Then, we harmonize 2D molecular graphs, protein sequences, and natural language within a unified, parameter-efficient fusion architecture by fine-tuning on multimodal question-answering datasets. Through comprehensive experiments, we show that BioMedGPT performs on par with human experts in comprehending biomedical documents and answering research questions. It also exhibits promising capability in analyzing intricate functions and properties of novel molecules and proteins, surpassing state-of-the-art LLMs by 17.1% and 49.8% absolute gains respectively in ROUGE-L on molecule and protein question-answering.
{"title":"BioMedGPT: An Open Multimodal Large Language Model for BioMedicine.","authors":"Yizhen Luo, Jiahuan Zhang, Siqi Fan, Kai Yang, Massimo Hong, Yushuai Wu, Mu Qiao, Zaiqing Nie","doi":"10.1109/JBHI.2024.3505955","DOIUrl":"10.1109/JBHI.2024.3505955","url":null,"abstract":"<p><p>Recent advances in large language models (LLMs) like ChatGPT have shed light on the development of knowledgeable and versatile AI research assistants in various scientific domains. However, they fall short in biomedical applications due to a lack of proprietary biomedical knowledge and deficiencies in handling biological sequences for molecules and proteins. To address these issues, we present BioMedGPT, a multimodal large language model for assisting biomedical research. We first incorporate domain expertise into LLMs by incremental pre-training on large-scale biomedical literature. Then, we harmonize 2D molecular graphs, protein sequences, and natural language within a unified, parameter-efficient fusion architecture by fine-tuning on multimodal question-answering datasets. Through comprehensive experiments, we show that BioMedGPT performs on par with human experts in comprehending biomedical documents and answering research questions. It also exhibits promising capability in analyzing intricate functions and properties of novel molecules and proteins, surpassing state-of-the-art LLMs by 17.1% and 49.8% absolute gains respectively in ROUGE-L on molecule and protein question-answering.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"981-992"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143604651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In modern medicine, the widespread use of medical imaging has greatly improved diagnostic and treatment efficiency. However, these images contain sensitive personal information, and any leakage could seriously compromise patient privacy, leading to ethical and legal issues. Federated learning (FL), an emerging privacy-preserving technique, transmits gradients rather than raw data for model training. Yet, recent studies reveal that gradient inversion attacks can exploit this information to reconstruct private data, posing a significant threat to FL. Current attacks remain limited in image resolution, similarity, and batch processing, and thus do not yet pose a significant risk to FL. To address this, we propose a novel gradient inversion attack based on sparsified gradient matching and segmentation reorganization (SR) to reconstruct high-resolution, high-similarity medical images in batch mode. Specifically, an $L_{1}$ loss function optimises the gradient sparsification process, while the SR strategy enhances image resolution. An adaptive learning rate adjustment mechanism is also employed to improve optimisation stability and avoid local optima. Experimental results demonstrate that our method significantly outperforms state-of-the-art approaches in both visual quality and quantitative metrics, achieving up to a 146% improvement in similarity.
{"title":"Medical Image Privacy in Federated Learning: Segmentation-Reorganization and Sparsified Gradient Matching Attacks.","authors":"Kaimin Wei, Jin Qian, Chengkun Jia, Jinpeng Chen, Jilian Zhang, Yongdong Wu, Jinyu Zhu, Yuhan Guo","doi":"10.1109/JBHI.2025.3593631","DOIUrl":"10.1109/JBHI.2025.3593631","url":null,"abstract":"<p><p>In modern medicine, the widespread use of medical imaging has greatly improved diagnostic and treatment efficiency. However, these images contain sensitive personal information, and any leakage could seriously compromise patient privacy, leading to ethical and legal issues. Federated learning (FL), an emerging privacy-preserving technique, transmits gradients rather than raw data for model training. Yet, recent studies reveal that gradient inversion attacks can exploit this information to reconstruct private data, posing a significant threat to FL. Current attacks remain limited in image resolution, similarity, and batch processing, and thus do not yet pose a significant risk to FL. To address this, we propose a novel gradient inversion attack based on sparsified gradient matching and segmentation reorganization (SR) to reconstruct high-resolution, high-similarity medical images in batch mode. Specifically, an $L_{1}$ loss function optimises the gradient sparsification process, while the SR strategy enhances image resolution. An adaptive learning rate adjustment mechanism is also employed to improve optimisation stability and avoid local optima. Experimental results demonstrate that our method significantly outperforms state-of-the-art approaches in both visual quality and quantitative metrics, achieving up to a 146% improvement in similarity.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1443-1451"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144952133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2025.3595371
Patricia A Apellaniz, Borja Arroyo Galende, Ana Jimenez, Juan Parras, Santiago Zazo
The scarcity of medical data, particularly in Survival Analysis (SA) for cancer-related diseases, challenges data-driven healthcare research. While Synthetic Tabular Data Generation (STDG) models have been proposed to address this issue, most rely on datasets with abundant samples, which do not reflect real-world limitations. We suggest using an STDG approach that leverages transfer learning and meta-learning techniques to create an artificial inductive bias, guiding generative models trained on limited samples. Experiments on classification datasets across varying sample sizes validated the method's robustness, with further clinical utility assessment on cancer-related SA data. While divergence-based similarity validation proved effective in capturing improvements in generation quality, clinical utility validation showed limited sensitivity to sample size, highlighting its shortcomings. In SA experiments, we observed that altering the task can reveal if relationships among variables are accurately generated, with most cases benefiting from the proposed methodology. Our findings confirm the method's ability to generate high-quality synthetic data under constrained conditions. We emphasize the need to complement utility-based validation with similarity metrics, particularly in low-data settings, to assess STDG performance reliably.
{"title":"Advancing Cancer Research With Synthetic Data Generation in Low-Data Scenarios.","authors":"Patricia A Apellaniz, Borja Arroyo Galende, Ana Jimenez, Juan Parras, Santiago Zazo","doi":"10.1109/JBHI.2025.3595371","DOIUrl":"10.1109/JBHI.2025.3595371","url":null,"abstract":"<p><p>The scarcity of medical data, particularly in Survival Analysis (SA) for cancer-related diseases, challenges data-driven healthcare research. While Synthetic Tabular Data Generation (STDG) models have been proposed to address this issue, most rely on datasets with abundant samples, which do not reflect real-world limitations. We suggest using an STDG approach that leverages transfer learning and meta-learning techniques to create an artificial inductive bias, guiding generative models trained on limited samples. Experiments on classification datasets across varying sample sizes validated the method's robustness, with further clinical utility assessment on cancer-related SA data. While divergence-based similarity validation proved effective in capturing improvements in generation quality, clinical utility validation showed limited sensitivity to sample size, highlighting its shortcomings. In SA experiments, we observed that altering the task can reveal if relationships among variables are accurately generated, with most cases benefiting from the proposed methodology. Our findings confirm the method's ability to generate high-quality synthetic data under constrained conditions. We emphasize the need to complement utility-based validation with similarity metrics, particularly in low-data settings, to assess STDG performance reliably.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1666-1679"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144784194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2025.3592024
Tong Jin, Jin Liu, Diandian Wang, Kun Wang, Chenlong Miao, Yikun Zhang, Dianlin Hu, Zhan Wu, Yang Chen
In computed tomography (CT), metal artifacts pose a persistent challenge to achieving high-quality imaging. Despite advancements in metal artifact reduction (MAR) techniques, many existing approaches have not fully leveraged the intrinsic a priori knowledge related to metal artifacts, improved model interpretability, or addressed the complex texture of CT images effectively. To address these limitations, we propose a novel and interpretable framework, the wavelet-inspired oriented adaptive dictionary network (WOADNet). WOADNet builds on sparse coding with orientational information in the wavelet domain. By exploring the discriminative features of artifacts and anatomical tissues, we adopt a high-precision filter parameterization strategy that incorporates multiangle rotations. Furthermore, we integrate a reweighted sparse constraint framework into the convolutional dictionary learning process and employ a cross-space, multiscale attention mechanism to construct an adaptive convolutional dictionary unit for the artifact feature encoder. This innovative design allows for flexible adjustment of weights and convolutional representations, resulting in significant image quality improvements. The experimental results using synthetic and clinical datasets demonstrate that WOADNet outperforms both traditional and state-of-the-art MAR methods in terms of suppressing artifacts.
{"title":"WOADNet: A Wavelet-Inspired Orientational Adaptive Dictionary Network for CT Metal Artifact Reduction.","authors":"Tong Jin, Jin Liu, Diandian Wang, Kun Wang, Chenlong Miao, Yikun Zhang, Dianlin Hu, Zhan Wu, Yang Chen","doi":"10.1109/JBHI.2025.3592024","DOIUrl":"10.1109/JBHI.2025.3592024","url":null,"abstract":"<p><p>In computed tomography (CT), metal artifacts pose a persistent challenge to achieving high-quality imaging. Despite advancements in metal artifact reduction (MAR) techniques, many existing approaches have not fully leveraged the intrinsic a priori knowledge related to metal artifacts, improved model interpretability, or addressed the complex texture of CT images effectively. To address these limitations, we propose a novel and interpretable framework, the wavelet-inspired oriented adaptive dictionary network (WOADNet). WOADNet builds on sparse coding with orientational information in the wavelet domain. By exploring the discriminative features of artifacts and anatomical tissues, we adopt a high-precision filter parameterization strategy that incorporates multiangle rotations. Furthermore, we integrate a reweighted sparse constraint framework into the convolutional dictionary learning process and employ a cross-space, multiscale attention mechanism to construct an adaptive convolutional dictionary unit for the artifact feature encoder. This innovative design allows for flexible adjustment of weights and convolutional representations, resulting in significant image quality improvements. The experimental results using synthetic and clinical datasets demonstrate that WOADNet outperforms both traditional and state-of-the-art MAR methods in terms of suppressing artifacts.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1452-1465"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144698429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2025.3594219
Jiaheng Wang, Zhenyu Wang, Tianheng Xu, Ang Li, Yuan Si, Ting Zhou, Xi Zhao, Honglin Hu
In recent years, the diverse applications of electroencephalography (EEG) - based affective brain-computer interfaces (aBCIs) are being extensively explored. However, due to adverse factors like noise and physiological variability, the recognition capability of aBCIs can unforeseeably suffer abrupt declines. Since the timing of these aBCI failures is unknown, placing trust in aBCIs without scrutiny can lead to undesirable consequences. To alleviate this issue, we propose an algorithm for estimating the reliability of aBCI (primarily Graph Convolutional Network), synchronously delivering a probabilistic confidence score upon aBCI decision completion, thereby reflecting the aBCI's real-time recognition capabilities. Methodologically, we use the Maximum Softmax Probability (MSP) from EEG recognition networks as confidence scores and leverage the Scaling Operator to calibrate them. Then, the Projection Operator is employed to address confidence estimation biases caused by noise and subject variability. For the numerical concentration of MSP, we provide fresh insights into its causes and propose corresponding solutions. The derivation of the estimator from the Maximum Entropy Principle is also substantiated for robust theoretical underpinnings. Finally, we confirm theoretically that the estimator does not compromise BCI performance. In experiments conducted on public datasets SEED and SEED-IV, the proposed algorithm demonstrates superior performance in estimating aBCIs reliability compared to other benchmarks, and commendable adaptability to new subjects. This research has the potential to lead to more trustworthy aBCIs and advance their broader application in complex real-world scenarios.
{"title":"Enhancing the Reliability of Affective Brain-Computer Interfaces by Using Specifically Designed Confidence Estimator.","authors":"Jiaheng Wang, Zhenyu Wang, Tianheng Xu, Ang Li, Yuan Si, Ting Zhou, Xi Zhao, Honglin Hu","doi":"10.1109/JBHI.2025.3594219","DOIUrl":"10.1109/JBHI.2025.3594219","url":null,"abstract":"<p><p>In recent years, the diverse applications of electroencephalography (EEG) - based affective brain-computer interfaces (aBCIs) are being extensively explored. However, due to adverse factors like noise and physiological variability, the recognition capability of aBCIs can unforeseeably suffer abrupt declines. Since the timing of these aBCI failures is unknown, placing trust in aBCIs without scrutiny can lead to undesirable consequences. To alleviate this issue, we propose an algorithm for estimating the reliability of aBCI (primarily Graph Convolutional Network), synchronously delivering a probabilistic confidence score upon aBCI decision completion, thereby reflecting the aBCI's real-time recognition capabilities. Methodologically, we use the Maximum Softmax Probability (MSP) from EEG recognition networks as confidence scores and leverage the Scaling Operator to calibrate them. Then, the Projection Operator is employed to address confidence estimation biases caused by noise and subject variability. For the numerical concentration of MSP, we provide fresh insights into its causes and propose corresponding solutions. The derivation of the estimator from the Maximum Entropy Principle is also substantiated for robust theoretical underpinnings. Finally, we confirm theoretically that the estimator does not compromise BCI performance. In experiments conducted on public datasets SEED and SEED-IV, the proposed algorithm demonstrates superior performance in estimating aBCIs reliability compared to other benchmarks, and commendable adaptability to new subjects. This research has the potential to lead to more trustworthy aBCIs and advance their broader application in complex real-world scenarios.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1073-1086"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144764849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2024.3432195
Andy Yiu-Chau Tam, Ye-Jiao Mao, Derek Ka-Hei Lai, Andy Chi-Ho Chan, Daphne Sze Ki Cheung, William Kearns, Duo Wai-Chi Wong, James Chung-Wai Cheung
The accuracy of sleep posture assessment in standard polysomnography might be compromised by the unfamiliar sleep lab environment. In this work, we aim to develop a depth camera-based sleep posture monitoring and classification system for home or community usage and tailor a deep learning model that can account for blanket interference. Our model included a joint coordinate estimation network (JCE) and sleep posture classification network (SPC). SaccpaNet (Separable Atrous Convolution-based Cascade Pyramid Attention Network) was developed using a combination of pyramidal structure of residual separable atrous convolution unit to reduce computational cost and enlarge receptive field. The Saccpa attention unit served as the core of JCE and SPC, while different backbones for SPC were also evaluated. The model was cross-modally pretrained by RGB images from the COCO whole body dataset and then trained/tested using dept image data collected from 150 participants performing seven sleep postures across four blanket conditions. Besides, we applied a data augmentation technique that used intra-class mix-up to synthesize blanket conditions; and an overlaid flip-cut to synthesize partially covered blanket conditions for a robustness that we referred to as the Post-hoc Data Augmentation Robustness Test (PhD-ART). Our model achieved an average precision of estimated joint coordinate (in terms of PCK@0.1) of 0.652 and demonstrated adequate robustness. The overall classification accuracy of sleep postures (F1-score) was 0.885 and 0.940, for 7- and 6-class classification, respectively. Our system was resistant to the interference of blanket, with a spread difference of 2.5%.
{"title":"SaccpaNet: A Separable Atrous Convolution- Based Cascade Pyramid Attention Network to Estimate Body Landmarks Using Cross-Modal Knowledge Transfer for Under-Blanket Sleep Posture Classification.","authors":"Andy Yiu-Chau Tam, Ye-Jiao Mao, Derek Ka-Hei Lai, Andy Chi-Ho Chan, Daphne Sze Ki Cheung, William Kearns, Duo Wai-Chi Wong, James Chung-Wai Cheung","doi":"10.1109/JBHI.2024.3432195","DOIUrl":"10.1109/JBHI.2024.3432195","url":null,"abstract":"<p><p>The accuracy of sleep posture assessment in standard polysomnography might be compromised by the unfamiliar sleep lab environment. In this work, we aim to develop a depth camera-based sleep posture monitoring and classification system for home or community usage and tailor a deep learning model that can account for blanket interference. Our model included a joint coordinate estimation network (JCE) and sleep posture classification network (SPC). SaccpaNet (Separable Atrous Convolution-based Cascade Pyramid Attention Network) was developed using a combination of pyramidal structure of residual separable atrous convolution unit to reduce computational cost and enlarge receptive field. The Saccpa attention unit served as the core of JCE and SPC, while different backbones for SPC were also evaluated. The model was cross-modally pretrained by RGB images from the COCO whole body dataset and then trained/tested using dept image data collected from 150 participants performing seven sleep postures across four blanket conditions. Besides, we applied a data augmentation technique that used intra-class mix-up to synthesize blanket conditions; and an overlaid flip-cut to synthesize partially covered blanket conditions for a robustness that we referred to as the Post-hoc Data Augmentation Robustness Test (PhD-ART). Our model achieved an average precision of estimated joint coordinate (in terms of PCK@0.1) of 0.652 and demonstrated adequate robustness. The overall classification accuracy of sleep postures (F1-score) was 0.885 and 0.940, for 7- and 6-class classification, respectively. Our system was resistant to the interference of blanket, with a spread difference of 2.5%.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1593-1604"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141751582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01DOI: 10.1109/JBHI.2025.3592897
Yifei Zhao, Xiaoying Wang, Junping Yin
Accurate and efficient Video Polyp Segmentation (VPS) is vital for the early detection of colorectal cancer and the effectivetreatment of polyps. However, achieving this remains highly challenging due to the inherent difficulty in modeling the spatial-temporal relationships within colonoscopy videos. Existing methods that directly associate video frames frequently fail to account for variations in polyp or background motion, leading to excessive noise and reduced segmentation accuracy. Conversely, approaches that rely on optical flow models to estimate motion and align frames incur significant computational overhead. To address these limitations, we propose a novel VPS framework, termed Deformable Alignment and Local Attention (DALA). In this framework, we first construct a shared encoder to jointly encode the feature representations of paired video frames. Subsequently, we introduce a Multi-Scale Frame Alignment (MSFA) module based on deformable convolution to estimate the motion between reference and anchor frames. The multi-scale architecture is designed to accommodate the scale variations of polyps arising from differing viewing angles and speeds during colonoscopy. Furthermore, Local Attention (LA) is employed to selectively aggregate the aligned features, yielding more precise spatial-temporal feature representations. Extensive experiments conducted on the challenging SUN-SEG dataset and PolypGen dataset demonstrate that DALA achieves superior performance compared to state-of-the-art models.
{"title":"Efficient Video Polyp Segmentation by Deformable Alignment and Local Attention.","authors":"Yifei Zhao, Xiaoying Wang, Junping Yin","doi":"10.1109/JBHI.2025.3592897","DOIUrl":"10.1109/JBHI.2025.3592897","url":null,"abstract":"<p><p>Accurate and efficient Video Polyp Segmentation (VPS) is vital for the early detection of colorectal cancer and the effectivetreatment of polyps. However, achieving this remains highly challenging due to the inherent difficulty in modeling the spatial-temporal relationships within colonoscopy videos. Existing methods that directly associate video frames frequently fail to account for variations in polyp or background motion, leading to excessive noise and reduced segmentation accuracy. Conversely, approaches that rely on optical flow models to estimate motion and align frames incur significant computational overhead. To address these limitations, we propose a novel VPS framework, termed Deformable Alignment and Local Attention (DALA). In this framework, we first construct a shared encoder to jointly encode the feature representations of paired video frames. Subsequently, we introduce a Multi-Scale Frame Alignment (MSFA) module based on deformable convolution to estimate the motion between reference and anchor frames. The multi-scale architecture is designed to accommodate the scale variations of polyps arising from differing viewing angles and speeds during colonoscopy. Furthermore, Local Attention (LA) is employed to selectively aggregate the aligned features, yielding more precise spatial-temporal feature representations. Extensive experiments conducted on the challenging SUN-SEG dataset and PolypGen dataset demonstrate that DALA achieves superior performance compared to state-of-the-art models.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":"1534-1543"},"PeriodicalIF":6.8,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144715061","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}