Pub Date : 2024-11-12DOI: 10.1109/JBHI.2024.3486251
Ching-Yi Lin, Marc Dandin
Cell culture assays play a vital role in various fields of biology. Conventional assay techniques like immunohistochemistry, immunofluorescence, and flow cytometry offer valuable insights into cell phenotype and behavior. However, each of these techniques requires labeling or staining, and this is a major drawback, specifically in applications that require compact and integrated analytical devices. To address this shortcoming, CMOS capacitance sensors capable of conducting label-free cell culture assays have been proposed. In this paper, we present a computational framework for further augmenting the capabilities of these capacitance sensors. In our framework, identification and classification of mitosis and migration are achieved by leveraging observations from measured capacitance time series data. Specifically, we engineered two time series features that enable discriminating cell behaviors at the single-cell level. Our feature representation achieves an area under curve (AUC) of 0.719 in the receiver operating characteristic (ROC) curve. Additionally, we show that our feature representation technique is applicable across arbitrary experiments, as validated by a leave-one-run-out test yielding an F-1 score of 0.803 and a G-Mean of 0.647.
{"title":"Machine Learning Identification and Classification of Mitosis and Migration of Cancer Cells in a Lab-on-CMOS Capacitance Sensing platform.","authors":"Ching-Yi Lin, Marc Dandin","doi":"10.1109/JBHI.2024.3486251","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3486251","url":null,"abstract":"<p><p>Cell culture assays play a vital role in various fields of biology. Conventional assay techniques like immunohistochemistry, immunofluorescence, and flow cytometry offer valuable insights into cell phenotype and behavior. However, each of these techniques requires labeling or staining, and this is a major drawback, specifically in applications that require compact and integrated analytical devices. To address this shortcoming, CMOS capacitance sensors capable of conducting label-free cell culture assays have been proposed. In this paper, we present a computational framework for further augmenting the capabilities of these capacitance sensors. In our framework, identification and classification of mitosis and migration are achieved by leveraging observations from measured capacitance time series data. Specifically, we engineered two time series features that enable discriminating cell behaviors at the single-cell level. Our feature representation achieves an area under curve (AUC) of 0.719 in the receiver operating characteristic (ROC) curve. Additionally, we show that our feature representation technique is applicable across arbitrary experiments, as validated by a leave-one-run-out test yielding an F-1 score of 0.803 and a G-Mean of 0.647.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Low-dose digital radiography (DR) and computed tomography (CT) become increasingly popular due to reduced radiation dose. However, they often result in degraded images with lower signal-to-noise ratios, creating an urgent need for effective denoising techniques. The recent advancement of the single-image-based denoising approach provides a promising solution without requirement of pairwise training data, which are scarce in medical imaging. These methods typically rely on sampling image pairs from a noisy image for inter-supervised denoising. Although enjoying simplicity, the generated image pairs are at the same noise level and only include partial information about the input images. This study argues that generating image pairs at different noise levels while fully using the information of the input image is preferable since it could provide richer multi-perspective clues to guide the denoising process. To this end, we present a novel Multi-Level Noise Sampling (MNS) method for low-dose tomography denoising. Specifically, MNS method generates multi-level noisy sub-images by partitioning the highdimensional input space into multiple low-dimensional subspaces with a simple yet effective strategy. The superiority of the MNS method in single-image-based denoising over the competing methods has been investigated and verified theoretically. Moreover, to bridge the gap between selfsupervised and supervised denoising networks, we introduce an optimization function that leverages prior knowledge of multi-level noisy sub-images to guide the training process. Through extensive quantitative and qualitative experiments conducted on large-scale clinical low-dose CT and DR datasets, we validate the effectiveness and superiority of our MNS approach over other state-of-the-art supervised and self-supervised methods.
低剂量数字放射摄影(DR)和计算机断层扫描(CT)因辐射剂量减少而越来越受欢迎。然而,它们通常会产生信噪比较低的退化图像,因此迫切需要有效的去噪技术。最近,基于单图像的去噪方法取得了进展,提供了一种不需要成对训练数据的有前途的解决方案,而这种数据在医学成像中非常稀缺。这些方法通常依赖于从噪声图像中抽取图像对进行监督间去噪。这些方法虽然简单,但生成的图像对噪声水平相同,而且只包含输入图像的部分信息。本研究认为,在充分利用输入图像信息的同时生成不同噪声水平的图像对更可取,因为它能提供更丰富的多角度线索来指导去噪过程。为此,我们提出了一种用于低剂量断层扫描去噪的新型多级噪声采样(MNS)方法。具体来说,MNS 方法通过将高维输入空间分割成多个低维子空间,以简单而有效的策略生成多级噪声子图像。我们从理论上研究并验证了 MNS 方法在基于单图像的去噪方面优于其他竞争方法。此外,为了缩小自我监督和监督去噪网络之间的差距,我们引入了一个优化函数,利用多级噪声子图像的先验知识来指导训练过程。通过在大规模临床低剂量 CT 和 DR 数据集上进行广泛的定量和定性实验,我们验证了 MNS 方法的有效性和优越性,超过了其他最先进的监督和自我监督方法。
{"title":"Multi-level Noise Sampling from Single Image for Low-dose Tomography Reconstruction.","authors":"Weiwen Wu, Yifei Long, Zhifan Gao, Guang Yang, Fangxiao Cheng, Jianjia Zhang","doi":"10.1109/JBHI.2024.3486726","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3486726","url":null,"abstract":"<p><p>Low-dose digital radiography (DR) and computed tomography (CT) become increasingly popular due to reduced radiation dose. However, they often result in degraded images with lower signal-to-noise ratios, creating an urgent need for effective denoising techniques. The recent advancement of the single-image-based denoising approach provides a promising solution without requirement of pairwise training data, which are scarce in medical imaging. These methods typically rely on sampling image pairs from a noisy image for inter-supervised denoising. Although enjoying simplicity, the generated image pairs are at the same noise level and only include partial information about the input images. This study argues that generating image pairs at different noise levels while fully using the information of the input image is preferable since it could provide richer multi-perspective clues to guide the denoising process. To this end, we present a novel Multi-Level Noise Sampling (MNS) method for low-dose tomography denoising. Specifically, MNS method generates multi-level noisy sub-images by partitioning the highdimensional input space into multiple low-dimensional subspaces with a simple yet effective strategy. The superiority of the MNS method in single-image-based denoising over the competing methods has been investigated and verified theoretically. Moreover, to bridge the gap between selfsupervised and supervised denoising networks, we introduce an optimization function that leverages prior knowledge of multi-level noisy sub-images to guide the training process. Through extensive quantitative and qualitative experiments conducted on large-scale clinical low-dose CT and DR datasets, we validate the effectiveness and superiority of our MNS approach over other state-of-the-art supervised and self-supervised methods.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3495835
Jing Bai, Zhengyang Zhang, Yue Yin, Weikang Jin, Talal Ahmed Ali Ali, Yong Xiong, Zhu Xiao
Incurable Alzheimer's disease (AD) plagues many elderly people and families. It is important to accurately diagnose and predict it at an early stage. However, the existing methods have shortcomings, such as inability to learn local and global information and the inability to extract effective features. In this paper, we propose a lightweight classification network Local and Global Graph ConvNeXt. This model has a hybrid architecture of convolutional neural network and Transformers. We build the Global NeXt Block and the Local NeXt Block to extract the local and global features of the structural magnetic resonance imaging (sMRI). These two blocks are optimized by adding global multilayer perceptron and locally grouped attention, respectively. Then, the features are fed into the pixel graph neural network to aggregate the valid pixel features using mask attention. In addition, we decoupled the loss by category to optimize the calculation of the loss. This method was tested on slices of the processed sMRI datasets from ADNI and achieved excellent performance. Our model achieves 95.81% accuracy with fewer parameters and floating point operations per second (FLOPS) than other classical efficient models in the diagnosis of AD.
无法治愈的阿尔茨海默病(AD)困扰着许多老年人和家庭。在早期阶段对其进行准确诊断和预测非常重要。然而,现有方法存在无法学习局部和全局信息、无法提取有效特征等缺点。在本文中,我们提出了一种轻量级分类网络本地和全局图 ConvNeXt。该模型采用卷积神经网络和变形器的混合架构。我们构建了全局 NeXt 块和局部 NeXt 块,以提取结构性磁共振成像(sMRI)的局部和全局特征。这两个区块分别通过添加全局多层感知器和局部分组注意进行优化。然后,将这些特征输入像素图神经网络,利用掩码注意力聚合有效的像素特征。此外,我们还将损失按类别解耦,以优化损失的计算。这种方法在 ADNI 处理过的 sMRI 数据集切片上进行了测试,取得了优异的性能。与其他诊断 AD 的经典高效模型相比,我们的模型以更少的参数和每秒浮点运算 (FLOPS) 达到了 95.81% 的准确率。
{"title":"LGG-NeXt: A Next Generation CNN and Transformer Hybrid Model for the Diagnosis of Alzheimer's Disease Using 2D Structural MRI.","authors":"Jing Bai, Zhengyang Zhang, Yue Yin, Weikang Jin, Talal Ahmed Ali Ali, Yong Xiong, Zhu Xiao","doi":"10.1109/JBHI.2024.3495835","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3495835","url":null,"abstract":"<p><p>Incurable Alzheimer's disease (AD) plagues many elderly people and families. It is important to accurately diagnose and predict it at an early stage. However, the existing methods have shortcomings, such as inability to learn local and global information and the inability to extract effective features. In this paper, we propose a lightweight classification network Local and Global Graph ConvNeXt. This model has a hybrid architecture of convolutional neural network and Transformers. We build the Global NeXt Block and the Local NeXt Block to extract the local and global features of the structural magnetic resonance imaging (sMRI). These two blocks are optimized by adding global multilayer perceptron and locally grouped attention, respectively. Then, the features are fed into the pixel graph neural network to aggregate the valid pixel features using mask attention. In addition, we decoupled the loss by category to optimize the calculation of the loss. This method was tested on slices of the processed sMRI datasets from ADNI and achieved excellent performance. Our model achieves 95.81% accuracy with fewer parameters and floating point operations per second (FLOPS) than other classical efficient models in the diagnosis of AD.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3496074
Jun Li, Yuzhu Guo
Objective: Freezing of gait (FOG) in Parkinson's disease has a complex neurological mechanism. Compared with other modalities, electroencephalogram (EEG) can reflect FOG-related brain activity of both motor and non-motor symptoms. However, EEG-based FOG prediction methods often extract time, spatial, frequency, time-frequency, or phase information separately, which fragments the coupling among these heterogeneous features and cannot completely characterize the brain dynamics when FOG occurs.
Methods: In this study, dynamic spatiotemporal coherent modes of EEG were studied and used for FOG detection and prediction. A dynamic mode decomposition (DMD) method was first applied to extract the spatiotemporal coherent modes. Dynamic changes of the spatiotemporal modes, in both amplitude and phase of motor-related frequency bands, were evaluated with analytic common spatial patterns (ACSP) to extract the essential differences among normal, freezing, and transitional gaits.
Results: The proposed method was verified in practical clinical data. Results showed that, in the detection task, the DMD-ACSP achieved an accuracy of 86.4 ± 3.6% and a sensitivity of 83.5 ± 4.3%. In the prediction task, 86.5 ± 3.2% accuracy and 86.7 ± 7.8% sensitivity were achieved.
Conclusion: Comparative studies showed that the DMD-ACSP method significantly improves FOG detection and prediction performance. Moreover, the DMD-ACSP reveals the spatial patterns of dynamic brain functional connectivity, which best discriminate the different gaits.
Significance: The spatiotemporal coherent modes may provide a useful indication for personalized intervention and transcranial magnetic stimulation neuromodulation in medical practices.
目的:帕金森病的步态冻结(FOG)具有复杂的神经机制。与其他模式相比,脑电图(EEG)可以反映与 FOG 相关的运动症状和非运动症状的大脑活动。然而,基于脑电图的 FOG 预测方法通常是分别提取时间、空间、频率、时频或相位信息,这就割裂了这些异构特征之间的耦合,无法完全描述 FOG 发生时的脑动力学特征:本研究对脑电图的动态时空相干模式进行了研究,并将其用于 FOG 的检测和预测。首先应用动态模式分解(DMD)方法提取时空相干模式。通过分析共同空间模式(ACSP)评估时空模式在运动相关频段的振幅和相位上的动态变化,以提取正常步态、冻结步态和过渡步态之间的本质区别:结果:所提出的方法在实际临床数据中得到了验证。结果表明,在检测任务中,DMD-ACSP 的准确率为 86.4 ± 3.6%,灵敏度为 83.5 ± 4.3%。在预测任务中,准确率为 86.5 ± 3.2%,灵敏度为 86.7 ± 7.8%:比较研究表明,DMD-ACSP 方法显著提高了 FOG 检测和预测性能。此外,DMD-ACSP 还揭示了动态脑功能连接的空间模式,这种模式最能区分不同的步态:时空相干模式可为医疗实践中的个性化干预和经颅磁刺激神经调控提供有用的指示。
{"title":"EEG Detection and Prediction of Freezing of Gait in Parkinson's Disease Based on Spatiotemporal Coherent Modes.","authors":"Jun Li, Yuzhu Guo","doi":"10.1109/JBHI.2024.3496074","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3496074","url":null,"abstract":"<p><strong>Objective: </strong>Freezing of gait (FOG) in Parkinson's disease has a complex neurological mechanism. Compared with other modalities, electroencephalogram (EEG) can reflect FOG-related brain activity of both motor and non-motor symptoms. However, EEG-based FOG prediction methods often extract time, spatial, frequency, time-frequency, or phase information separately, which fragments the coupling among these heterogeneous features and cannot completely characterize the brain dynamics when FOG occurs.</p><p><strong>Methods: </strong>In this study, dynamic spatiotemporal coherent modes of EEG were studied and used for FOG detection and prediction. A dynamic mode decomposition (DMD) method was first applied to extract the spatiotemporal coherent modes. Dynamic changes of the spatiotemporal modes, in both amplitude and phase of motor-related frequency bands, were evaluated with analytic common spatial patterns (ACSP) to extract the essential differences among normal, freezing, and transitional gaits.</p><p><strong>Results: </strong>The proposed method was verified in practical clinical data. Results showed that, in the detection task, the DMD-ACSP achieved an accuracy of 86.4 ± 3.6% and a sensitivity of 83.5 ± 4.3%. In the prediction task, 86.5 ± 3.2% accuracy and 86.7 ± 7.8% sensitivity were achieved.</p><p><strong>Conclusion: </strong>Comparative studies showed that the DMD-ACSP method significantly improves FOG detection and prediction performance. Moreover, the DMD-ACSP reveals the spatial patterns of dynamic brain functional connectivity, which best discriminate the different gaits.</p><p><strong>Significance: </strong>The spatiotemporal coherent modes may provide a useful indication for personalized intervention and transcranial magnetic stimulation neuromodulation in medical practices.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3496700
Baiying Lei, Gege Cai, Yun Zhu, Tianfu Wang, Lei Dong, Cheng Zhao, Xinzhi Hu, Huijun Zhu, Lin Lu, Feng Feng, Ming Feng, Renzhi Wang
The sellar region tumor is a brain tumor that only exists in the brain sellar, which affects the central nervous system. The early diagnosis of the sellar region tumor subtypes helps clinicians better understand the best treatment and recovery of pa-tients. Magnetic resonance imaging (MRI) has proven to be an effective tool for the early detection of sellar region tumors. However, the existing sellar region tumor diagnosis still remains challenging due to the small amount of dataset and data imbalance. To overcome these challenges, we propose a novel self-supervised multi-scale multi-modal graph pool Transformer (MMGPT) network that can enhance the multi-modal fusion of small and imbalanced MRI data of sellar region tumors. MMGPT can strengthen feature interaction between multi-modal images, which makes our model more robust. A contrastive learning equipped auto-encoder (CAE) via self-supervised learning (SSL) is adopted to learn more detailed information between different samples. The proposed CAE transfers the pre-trained knowledge to the downstream tasks. Finally, a hybrid loss is equipped to relieve the performance degradation caused by data imbalance. The experimental results show that the proposed method outperforms state-of-the-art methods and obtains higher accuracy and AUC in the classification of sellar region tumors.
{"title":"Self-supervised Multi-scale Multi-modal Graph Pool Transformer for Sellar Region Tumor Diagnosis.","authors":"Baiying Lei, Gege Cai, Yun Zhu, Tianfu Wang, Lei Dong, Cheng Zhao, Xinzhi Hu, Huijun Zhu, Lin Lu, Feng Feng, Ming Feng, Renzhi Wang","doi":"10.1109/JBHI.2024.3496700","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3496700","url":null,"abstract":"<p><p>The sellar region tumor is a brain tumor that only exists in the brain sellar, which affects the central nervous system. The early diagnosis of the sellar region tumor subtypes helps clinicians better understand the best treatment and recovery of pa-tients. Magnetic resonance imaging (MRI) has proven to be an effective tool for the early detection of sellar region tumors. However, the existing sellar region tumor diagnosis still remains challenging due to the small amount of dataset and data imbalance. To overcome these challenges, we propose a novel self-supervised multi-scale multi-modal graph pool Transformer (MMGPT) network that can enhance the multi-modal fusion of small and imbalanced MRI data of sellar region tumors. MMGPT can strengthen feature interaction between multi-modal images, which makes our model more robust. A contrastive learning equipped auto-encoder (CAE) via self-supervised learning (SSL) is adopted to learn more detailed information between different samples. The proposed CAE transfers the pre-trained knowledge to the downstream tasks. Finally, a hybrid loss is equipped to relieve the performance degradation caused by data imbalance. The experimental results show that the proposed method outperforms state-of-the-art methods and obtains higher accuracy and AUC in the classification of sellar region tumors.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cloud computing and Internet of Things (IoT) technologies are gradually becoming the technological changemakers in cancer diagnosis. Blood cancer is an aggressive disease affecting the blood, bone marrow, and lymphatic system, and its early detection is crucial for subsequent treatment. Flow cytometry has been widely studied as a commonly used method for detecting blood cancer. However, the high computation and resource consumption severely limit its practical application, especifically in regions with limited medical and computational resources. In this study, with the help of cloud computing and IoT technologies, we develop a novel blood cancer dynamic monitoring diagnostic model named BloodPatrol based on an intelligent feature weight fusion mechanism. The proposed model is capable of capturing the dual-view importance relationship between cell samples and features, greatly improving prediction accuracy and significantly surpassing previous models. Besides, benefiting from the powerful processing ability of cloud computing, BloodPatrol can run on a distributed network to efficiently process large-scale cell data, which provides immediate and scalable blood cancer diagnostic services. We have also created a cloud diagnostic platform to facilitate access to our work, the latest access link and updates are available at: https://github.com/kkkayle/BloodPatrol.
{"title":"BloodPatrol: Revolutionizing Blood Cancer Diagnosis - Advanced Real-Time Detection Leveraging Deep Learning & Cloud Technologies.","authors":"Jinhang Wei, Longyue Wang, Zhecheng Zhou, Linlin Zhuo, Xiangxiang Zeng, Xiangzheng Fu, Quan Zou, Keqin Li, Zhongjun Zhou","doi":"10.1109/JBHI.2024.3496294","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3496294","url":null,"abstract":"<p><p>Cloud computing and Internet of Things (IoT) technologies are gradually becoming the technological changemakers in cancer diagnosis. Blood cancer is an aggressive disease affecting the blood, bone marrow, and lymphatic system, and its early detection is crucial for subsequent treatment. Flow cytometry has been widely studied as a commonly used method for detecting blood cancer. However, the high computation and resource consumption severely limit its practical application, especifically in regions with limited medical and computational resources. In this study, with the help of cloud computing and IoT technologies, we develop a novel blood cancer dynamic monitoring diagnostic model named BloodPatrol based on an intelligent feature weight fusion mechanism. The proposed model is capable of capturing the dual-view importance relationship between cell samples and features, greatly improving prediction accuracy and significantly surpassing previous models. Besides, benefiting from the powerful processing ability of cloud computing, BloodPatrol can run on a distributed network to efficiently process large-scale cell data, which provides immediate and scalable blood cancer diagnostic services. We have also created a cloud diagnostic platform to facilitate access to our work, the latest access link and updates are available at: https://github.com/kkkayle/BloodPatrol.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3495975
Clara Garcia-Vicente, Gonzalo C Gutierrez-Tobal, Fernando Vaquerizo-Villar, Adrian Martin-Montero, David Gozal, Roberto Hornero
Obstructive sleep apnea (OSA) in children is a prevalent and serious respiratory condition linked to cardiovascular morbidity. Polysomnography, the standard diagnostic approach, faces challenges in accessibility and complexity, leading to underdiagnosis. To simplify OSA diagnosis, deep learning (DL) algorithms have been developed using cardiac signals, but they often lack interpretability. Our study introduces a novel interpretable DL approach (SleepECG-Net) for directly estimating OSA severity in at-risk children. A combination of convolutional and recurrent neural networks (CNN-RNN) was trained on overnight electrocardiogram (ECG) signals. Gradient-weighted Class Activation Mapping (Grad-CAM), an eXplainable Artificial Intelligence (XAI) algorithm, was applied to explain model decisions and extract ECG patterns relevant to pediatric OSA. Accordingly, ECG signals from the semi-public Childhood Adenotonsillectomy Trial (CHAT, n = 1610) and Cleveland Family Study (CFS, n = 64), and the private University of Chicago (UofC, n = 981) databases were used. OSA diagnostic performance reached 4-class Cohen's Kappa of 0.410, 0.335, and 0.249 in CHAT, UofC, and CFS, respectively. The proposal demonstrated improved performance with increased severity along with heightened cardiovascular risk. XAI findings highlighted the detection of established ECG features linked to OSA, such as bradycardia-tachycardia events and delayed ECG patterns during apnea/hypopnea occurrences, focusing on clusters of events. Furthermore, Grad-CAM heatmaps identified potential ECG patterns indicating cardiovascular risk, such as P, T, and U waves, QT intervals, and QRS complex variations. Hence, SleepECG-Net approach may improve pediatric OSA diagnosis by also offering cardiac risk factor information, thereby increasing clinician confidence in automated systems, and promoting their effective adoption in clinical practice.
儿童阻塞性睡眠呼吸暂停(OSA)是一种普遍而严重的呼吸系统疾病,与心血管疾病的发病率有关。多导睡眠图是一种标准的诊断方法,但在可及性和复杂性方面面临挑战,导致诊断不足。为了简化 OSA 诊断,人们利用心脏信号开发了深度学习(DL)算法,但这些算法往往缺乏可解释性。我们的研究引入了一种新颖的可解释深度学习方法(SleepECG-Net),用于直接估计高危儿童的 OSA 严重程度。我们在隔夜心电图(ECG)信号上训练了卷积神经网络和递归神经网络(CNN-RNN)的组合。梯度加权类活化映射(Grad-CAM)是一种可解释的人工智能(XAI)算法,用于解释模型决策和提取与小儿 OSA 相关的心电图模式。因此,我们使用了来自半公开的儿童腺样体切除术试验(CHAT,n = 1610)和克利夫兰家庭研究(CFS,n = 64)以及私人的芝加哥大学(UofC,n = 981)数据库的心电图信号。在 CHAT、UofC 和 CFS 中,OSA 诊断性能的 4 级 Cohen's Kappa 分别为 0.410、0.335 和 0.249。该建议表明,随着严重程度的增加以及心血管风险的增加,诊断性能也会提高。XAI 的研究结果突出了与 OSA 相关的既定心电图特征的检测,如心动过缓-心动过速事件和呼吸暂停/呼吸暂停发生时的延迟心电图模式,重点是事件群。此外,Grad-CAM 热图还能识别显示心血管风险的潜在心电图模式,如 P 波、T 波和 U 波、QT 间期和 QRS 波群变化。因此,SleepECG-Net 方法可以通过提供心脏风险因素信息来改善儿科 OSA 诊断,从而增强临床医生对自动系统的信心,并促进其在临床实践中的有效应用。
{"title":"SleepECG-Net: explainable deep learning approach with ECG for pediatric sleep apnea diagnosis.","authors":"Clara Garcia-Vicente, Gonzalo C Gutierrez-Tobal, Fernando Vaquerizo-Villar, Adrian Martin-Montero, David Gozal, Roberto Hornero","doi":"10.1109/JBHI.2024.3495975","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3495975","url":null,"abstract":"<p><p>Obstructive sleep apnea (OSA) in children is a prevalent and serious respiratory condition linked to cardiovascular morbidity. Polysomnography, the standard diagnostic approach, faces challenges in accessibility and complexity, leading to underdiagnosis. To simplify OSA diagnosis, deep learning (DL) algorithms have been developed using cardiac signals, but they often lack interpretability. Our study introduces a novel interpretable DL approach (SleepECG-Net) for directly estimating OSA severity in at-risk children. A combination of convolutional and recurrent neural networks (CNN-RNN) was trained on overnight electrocardiogram (ECG) signals. Gradient-weighted Class Activation Mapping (Grad-CAM), an eXplainable Artificial Intelligence (XAI) algorithm, was applied to explain model decisions and extract ECG patterns relevant to pediatric OSA. Accordingly, ECG signals from the semi-public Childhood Adenotonsillectomy Trial (CHAT, n = 1610) and Cleveland Family Study (CFS, n = 64), and the private University of Chicago (UofC, n = 981) databases were used. OSA diagnostic performance reached 4-class Cohen's Kappa of 0.410, 0.335, and 0.249 in CHAT, UofC, and CFS, respectively. The proposal demonstrated improved performance with increased severity along with heightened cardiovascular risk. XAI findings highlighted the detection of established ECG features linked to OSA, such as bradycardia-tachycardia events and delayed ECG patterns during apnea/hypopnea occurrences, focusing on clusters of events. Furthermore, Grad-CAM heatmaps identified potential ECG patterns indicating cardiovascular risk, such as P, T, and U waves, QT intervals, and QRS complex variations. Hence, SleepECG-Net approach may improve pediatric OSA diagnosis by also offering cardiac risk factor information, thereby increasing clinician confidence in automated systems, and promoting their effective adoption in clinical practice.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619328","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3496495
Xingsi Xue, Mu-En Wu, Fazlullah Khan
Integrating diverse biomedical knowledge information is essential to enhance the accuracy and efficiency of medical diagnoses, facilitate personalized treatment plans, and ultimately improve patient outcomes. However, Biomedical Information Integration (BII) faces significant challenges due to variations in terminology and the complex structure of entity descriptions across different datasets. A critical step in BII is biomedical entity alignment, which involves accurately identifying and matching equivalent entities across diverse datasets to ensure seamless data integration. In recent years, Large Language Model (LLMs), such as Bidirectional Encoder Representations from Transformers (BERTs), have emerged as valuable tools for discerning heterogeneous biomedical data due to their deep contextual embeddings and bidirectionality. However, different LLMs capture various nuances and complexity levels within the biomedical data, and none of them can ensure their effectiveness in all heterogeneous entity matching tasks. To address this issue, we propose a novel Two-Stage LLM construction (TSLLM) framework to adaptively select and combine LLMs for Biomedical Information Integration (BII). First, a Multi-Objective Genetic Programming (MOGP) algorithm is proposed for generating versatile high-level LLMs, and then, a Single-Objective Genetic Algorithm (SOGA) employs a confidence-based strategy is presented to combine the built LLMs, which can further improve the discriminative power of distinguishing heterogeneous entities. The experiment utilizes OAEI's entity matching datasets, i.e., Benchmark and Conference, along with LargeBio, Disease and Phenotype datasets to test the performance of TSLLM. The experimental findings validate the efficiency of TSLLM in adaptively differentiating heterogeneous biomedical entities, which significantly outperforms the leading entity matching techniques.
{"title":"Biomedical Information Integration via Adaptive Large Language Model Construction.","authors":"Xingsi Xue, Mu-En Wu, Fazlullah Khan","doi":"10.1109/JBHI.2024.3496495","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3496495","url":null,"abstract":"<p><p>Integrating diverse biomedical knowledge information is essential to enhance the accuracy and efficiency of medical diagnoses, facilitate personalized treatment plans, and ultimately improve patient outcomes. However, Biomedical Information Integration (BII) faces significant challenges due to variations in terminology and the complex structure of entity descriptions across different datasets. A critical step in BII is biomedical entity alignment, which involves accurately identifying and matching equivalent entities across diverse datasets to ensure seamless data integration. In recent years, Large Language Model (LLMs), such as Bidirectional Encoder Representations from Transformers (BERTs), have emerged as valuable tools for discerning heterogeneous biomedical data due to their deep contextual embeddings and bidirectionality. However, different LLMs capture various nuances and complexity levels within the biomedical data, and none of them can ensure their effectiveness in all heterogeneous entity matching tasks. To address this issue, we propose a novel Two-Stage LLM construction (TSLLM) framework to adaptively select and combine LLMs for Biomedical Information Integration (BII). First, a Multi-Objective Genetic Programming (MOGP) algorithm is proposed for generating versatile high-level LLMs, and then, a Single-Objective Genetic Algorithm (SOGA) employs a confidence-based strategy is presented to combine the built LLMs, which can further improve the discriminative power of distinguishing heterogeneous entities. The experiment utilizes OAEI's entity matching datasets, i.e., Benchmark and Conference, along with LargeBio, Disease and Phenotype datasets to test the performance of TSLLM. The experimental findings validate the efficiency of TSLLM in adaptively differentiating heterogeneous biomedical entities, which significantly outperforms the leading entity matching techniques.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3496122
Amir Mehdi Shayan, David B Hitchcock, Simar Singh, Jianxin Gao, Richard E Groff, Ravikiran B Singapogu
This study explores the application of functional data analysis (FDA) to hand roll velocity during radial suturing on the SutureCoach bench simulator for evaluating open suturing performance. By treating temporal sensor data as mathematical functions, FDA provides a holistic view of the dynamic changes in hand roll, offering comprehensive assessments that are easily interpretable and clinically relevant. Cluster analysis was performed on hand roll profiles from 96 subjects, categorized into advanced surgeons, trainee surgeons, and novices. Functional k-means, using dynamic time-warping to align curves, were used to partition the data into two preset numbers of clusters (3 and 6). Both clustering models (3-cluster and 6-cluster) effectively clustered performance into groups with distinct characteristics and levels of skill (evident from visual inspection of cluster centroids). The relationship between cluster membership and suturing skills was corroborated using proxy measures of skill: expert global rating scale ratings, clinical status and expertise, and simulator-derived metrics. The findings of this study offer valuable insight into essential components of suturing skill and can improve the autonomy and efficiency of simulation-based suturing training. The clinical relevance of our results is immediately pertinent to the field of surgical skill assessment, where FDA-based methods could potentially be employed for objective feedback and training.
本研究探讨了功能数据分析(FDA)在 SutureCoach 工作台模拟器上径向缝合过程中手滚动速度的应用,以评估开放式缝合性能。通过将时间传感器数据视为数学函数,FDA 提供了手滚动动态变化的整体视图,提供了易于解释且与临床相关的综合评估。对 96 名受试者的手滚动曲线进行了聚类分析,分为高级外科医生、实习外科医生和新手。使用功能 k 均值法,利用动态时间平移对齐曲线,将数据划分为两个预设数目的聚类(3 和 6)。两种聚类模型(3 个聚类和 6 个聚类)都能有效地将成绩聚类为具有不同特征和技能水平的组别(通过目测聚类中心点可以明显看出)。群组成员资格与缝合技能之间的关系可通过技能的替代衡量标准得到证实,这些衡量标准包括:专家全球评分量表评分、临床状态和专业知识以及模拟器衍生指标。这项研究的结果为了解缝合技能的基本要素提供了宝贵的见解,并能提高模拟缝合培训的自主性和效率。我们研究结果的临床意义与外科技能评估领域密切相关,基于 FDA 的方法可用于客观反馈和培训。
{"title":"Functional Data Analysis of Hand Rotation for Open Surgical Suturing Skill Assessment.","authors":"Amir Mehdi Shayan, David B Hitchcock, Simar Singh, Jianxin Gao, Richard E Groff, Ravikiran B Singapogu","doi":"10.1109/JBHI.2024.3496122","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3496122","url":null,"abstract":"<p><p>This study explores the application of functional data analysis (FDA) to hand roll velocity during radial suturing on the SutureCoach bench simulator for evaluating open suturing performance. By treating temporal sensor data as mathematical functions, FDA provides a holistic view of the dynamic changes in hand roll, offering comprehensive assessments that are easily interpretable and clinically relevant. Cluster analysis was performed on hand roll profiles from 96 subjects, categorized into advanced surgeons, trainee surgeons, and novices. Functional k-means, using dynamic time-warping to align curves, were used to partition the data into two preset numbers of clusters (3 and 6). Both clustering models (3-cluster and 6-cluster) effectively clustered performance into groups with distinct characteristics and levels of skill (evident from visual inspection of cluster centroids). The relationship between cluster membership and suturing skills was corroborated using proxy measures of skill: expert global rating scale ratings, clinical status and expertise, and simulator-derived metrics. The findings of this study offer valuable insight into essential components of suturing skill and can improve the autonomy and efficiency of simulation-based suturing training. The clinical relevance of our results is immediately pertinent to the field of surgical skill assessment, where FDA-based methods could potentially be employed for objective feedback and training.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-11-11DOI: 10.1109/JBHI.2024.3496757
Yuxuan Shi, Aimin Jiang, Ju Zhong, Min Li, Yanping Zhu
In motor imagery (MI) tasks for brain computer interfaces (BCIs), the spatial covariance matrix (SCM) of electroencephalogram (EEG) signals plays a critical role in accurate classification. Given that SCMs are symmetric positive definite (SPD), Riemannian geometry is widely utilized to extract classification features. However, calculating distances between SCMs is computationally intensive due to operations like eigenvalue decomposition, and classical optimization techniques, such as gradient descent, cannot be directly applied to Riemannian manifolds, making the computation of the Riemannian mean more complex and reliant on iterative methods or approximations. In this paper, we propose a novel multiclass classification framework that integrates Riemannian geometry and neural networks to mitigate these challenges. The framework comprises two modules: a Riemannian module with multiple branches and a classification module. During training, a fusion loss function is introduced to update the branch corresponding to the true label, while other branches are updated using different loss functions along with the classification module. Comprehensive experiments on four sets of MI EEG data demonstrate the efficiency and effectiveness of the proposed model.
在脑计算机接口(BCI)的运动想象(MI)任务中,脑电图(EEG)信号的空间协方差矩阵(SCM)对准确分类起着至关重要的作用。鉴于空间协方差矩阵是对称正定(SPD)的,黎曼几何被广泛用于提取分类特征。然而,由于特征值分解等操作,计算单片机之间的距离需要大量计算,而且梯度下降等经典优化技术不能直接应用于黎曼流形,这使得黎曼均值的计算变得更加复杂,并依赖于迭代法或近似法。在本文中,我们提出了一个新颖的多类分类框架,将黎曼几何和神经网络整合在一起,以减轻这些挑战。该框架由两个模块组成:具有多个分支的黎曼模块和分类模块。在训练过程中,引入一个融合损失函数来更新与真实标签相对应的分支,而其他分支则与分类模块一起使用不同的损失函数进行更新。四组 MI EEG 数据的综合实验证明了所提模型的效率和有效性。
{"title":"Multiclass Classification Framework of Motor Imagery EEG by Riemannian Geometry Networks.","authors":"Yuxuan Shi, Aimin Jiang, Ju Zhong, Min Li, Yanping Zhu","doi":"10.1109/JBHI.2024.3496757","DOIUrl":"https://doi.org/10.1109/JBHI.2024.3496757","url":null,"abstract":"<p><p>In motor imagery (MI) tasks for brain computer interfaces (BCIs), the spatial covariance matrix (SCM) of electroencephalogram (EEG) signals plays a critical role in accurate classification. Given that SCMs are symmetric positive definite (SPD), Riemannian geometry is widely utilized to extract classification features. However, calculating distances between SCMs is computationally intensive due to operations like eigenvalue decomposition, and classical optimization techniques, such as gradient descent, cannot be directly applied to Riemannian manifolds, making the computation of the Riemannian mean more complex and reliant on iterative methods or approximations. In this paper, we propose a novel multiclass classification framework that integrates Riemannian geometry and neural networks to mitigate these challenges. The framework comprises two modules: a Riemannian module with multiple branches and a classification module. During training, a fusion loss function is introduced to update the branch corresponding to the true label, while other branches are updated using different loss functions along with the classification module. Comprehensive experiments on four sets of MI EEG data demonstrate the efficiency and effectiveness of the proposed model.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142619321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}