Predicting a patient's future health state through the analysis of their clinical records is an emerging area in the field of intelligent medicine. It has the potential to assist healthcare professionals in prescribing treatments safely, making more accurate diagnoses, and improving patient care. However, clinical notes have been underutilized due to their complexity, high dimensionality, and sparsity. Nevertheless, these clinical records hold significant promise for enhancing clinical decision. To tackle these problems, a novel feedback attention-based bidirectional long short-term memory (FABiLSTM) model has been proposed for more effective diagnosis using clinical records. This model incorporates PubMedBERT for filtering irrelevant information, enhances global vector word embeddings for numerical representations and K-means clustering, and performs to explore term frequency and inverse document frequency intricacies. The proposed approach excels in capturing information, aiding accurate disease prediction. The predictive capability is further enhanced with the help of a billiards-inspired optimization algorithm. The effectiveness of the FABiLSTM method has been assessed with the MIMIC-III dataset, yielding impressive results in accuracy, precision, F1 score, and recall score of 98.52%, 98%, 98.2%, and 98.2% individually. These results reveal ways in which the proposed technique excels in comparison with current practices.
{"title":"Processing of clinical notes for efficient diagnosis with feedback attention-based BiLSTM.","authors":"Nitalaksheswara Rao Kolukula, Sreekanth Puli, Chandaka Babi, Rajendra Prasad Kalapala, Gandhi Ongole, Venkata Murali Krishna Chinta","doi":"10.1007/s11517-024-03126-8","DOIUrl":"10.1007/s11517-024-03126-8","url":null,"abstract":"<p><p>Predicting a patient's future health state through the analysis of their clinical records is an emerging area in the field of intelligent medicine. It has the potential to assist healthcare professionals in prescribing treatments safely, making more accurate diagnoses, and improving patient care. However, clinical notes have been underutilized due to their complexity, high dimensionality, and sparsity. Nevertheless, these clinical records hold significant promise for enhancing clinical decision. To tackle these problems, a novel feedback attention-based bidirectional long short-term memory (FABiLSTM) model has been proposed for more effective diagnosis using clinical records. This model incorporates PubMedBERT for filtering irrelevant information, enhances global vector word embeddings for numerical representations and K-means clustering, and performs to explore term frequency and inverse document frequency intricacies. The proposed approach excels in capturing information, aiding accurate disease prediction. The predictive capability is further enhanced with the help of a billiards-inspired optimization algorithm. The effectiveness of the FABiLSTM method has been assessed with the MIMIC-III dataset, yielding impressive results in accuracy, precision, F1 score, and recall score of 98.52%, 98%, 98.2%, and 98.2% individually. These results reveal ways in which the proposed technique excels in comparison with current practices.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3193-3208"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141155573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-24DOI: 10.1007/s11517-024-03127-7
Shilpa Bajaj, Manju Bala, Mohit Angurala
Deep learning (DL) requires a large amount of training data to improve performance and prevent overfitting. To overcome these difficulties, we need to increase the size of the training dataset. This can be done by augmentation on a small dataset. The augmentation approaches must enhance the model's performance during the learning period. There are several types of transformations that can be applied to medical images. These transformations can be applied to the entire dataset or to a subset of the data, depending on the desired outcome. In this study, we categorize data augmentation methods into four groups: Absent augmentation, where no modifications are made; basic augmentation, which includes brightness and contrast adjustments; intermediate augmentation, encompassing a wider array of transformations like rotation, flipping, and shifting in addition to brightness and contrast adjustments; and advanced augmentation, where all transformation layers are employed. We plan to conduct a comprehensive analysis to determine which group performs best when applied to brain CT images. This evaluation aims to identify the augmentation group that produces the most favorable results in terms of improving model accuracy, minimizing diagnostic errors, and ensuring the robustness of the model in the context of brain CT image analysis.
{"title":"A comparative analysis of different augmentations for brain images.","authors":"Shilpa Bajaj, Manju Bala, Mohit Angurala","doi":"10.1007/s11517-024-03127-7","DOIUrl":"10.1007/s11517-024-03127-7","url":null,"abstract":"<p><p>Deep learning (DL) requires a large amount of training data to improve performance and prevent overfitting. To overcome these difficulties, we need to increase the size of the training dataset. This can be done by augmentation on a small dataset. The augmentation approaches must enhance the model's performance during the learning period. There are several types of transformations that can be applied to medical images. These transformations can be applied to the entire dataset or to a subset of the data, depending on the desired outcome. In this study, we categorize data augmentation methods into four groups: Absent augmentation, where no modifications are made; basic augmentation, which includes brightness and contrast adjustments; intermediate augmentation, encompassing a wider array of transformations like rotation, flipping, and shifting in addition to brightness and contrast adjustments; and advanced augmentation, where all transformation layers are employed. We plan to conduct a comprehensive analysis to determine which group performs best when applied to brain CT images. This evaluation aims to identify the augmentation group that produces the most favorable results in terms of improving model accuracy, minimizing diagnostic errors, and ensuring the robustness of the model in the context of brain CT image analysis.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3123-3150"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141087747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-09DOI: 10.1007/s11517-024-03103-1
Zhaohui Li, Xiaohui Tan, Xinyu Li, Liyong Yin
Motor imagery (MI) based brain-computer interfaces (BCIs) decode the users' intentions from electroencephalography (EEG) to achieve information control and interaction between the brain and external devices. In this paper, firstly, we apply Riemannian geometry to the covariance matrix extracted by spatial filtering to obtain robust and distinct features. Then, a multiscale temporal-spectral segmentation scheme is developed to enrich the feature dimensionality. In order to determine the optimal feature configurations, we utilize a linear learning-based temporal window and spectral band (TWSB) selection method to evaluate the feature contributions, which efficiently reduces the redundant features and improves the decoding efficiency without excessive loss of accuracy. Finally, support vector machines are used to predict the classification labels based on the selected MI features. To evaluate the performance of our model, we test it on the publicly available BCI Competition IV dataset 2a and 2b. The results show that the method has an average accuracy of 79.1% and 83.1%, which outperforms other existing methods. Using TWSB feature selection instead of selecting all features improves the accuracy by up to about 6%. Moreover, the TWSB selection method can effectively reduce the computational burden. We believe that the framework reveals more interpretable feature information of motor imagery EEG signals, provides neural responses discriminative with high accuracy, and facilitates the performance of real-time MI-BCI.
基于运动图像(MI)的脑机接口(BCI)能从脑电图(EEG)中解码用户的意图,从而实现大脑与外部设备之间的信息控制和交互。在本文中,我们首先对空间滤波提取的协方差矩阵进行黎曼几何处理,以获得鲁棒且独特的特征。然后,我们开发了一种多尺度时间-光谱分割方案,以丰富特征维度。为了确定最佳特征配置,我们采用了一种基于线性学习的时窗和频谱带(TWSB)选择方法来评估特征贡献,从而有效地减少了冗余特征,提高了解码效率,同时不会损失过多的精度。最后,我们使用支持向量机来预测基于所选 MI 特征的分类标签。为了评估模型的性能,我们在公开的 BCI Competition IV 数据集 2a 和 2b 上进行了测试。结果表明,该方法的平均准确率分别为 79.1% 和 83.1%,优于其他现有方法。使用 TWSB 特征选择代替选择所有特征,可将准确率提高约 6%。此外,TWSB 选择方法还能有效减轻计算负担。我们认为,该框架揭示了运动意象脑电信号中更多可解释的特征信息,提供了高准确度的神经反应判别,有助于实时 MI-BCI 的实现。
{"title":"Multiclass motor imagery classification with Riemannian geometry and temporal-spectral selection.","authors":"Zhaohui Li, Xiaohui Tan, Xinyu Li, Liyong Yin","doi":"10.1007/s11517-024-03103-1","DOIUrl":"10.1007/s11517-024-03103-1","url":null,"abstract":"<p><p>Motor imagery (MI) based brain-computer interfaces (BCIs) decode the users' intentions from electroencephalography (EEG) to achieve information control and interaction between the brain and external devices. In this paper, firstly, we apply Riemannian geometry to the covariance matrix extracted by spatial filtering to obtain robust and distinct features. Then, a multiscale temporal-spectral segmentation scheme is developed to enrich the feature dimensionality. In order to determine the optimal feature configurations, we utilize a linear learning-based temporal window and spectral band (TWSB) selection method to evaluate the feature contributions, which efficiently reduces the redundant features and improves the decoding efficiency without excessive loss of accuracy. Finally, support vector machines are used to predict the classification labels based on the selected MI features. To evaluate the performance of our model, we test it on the publicly available BCI Competition IV dataset 2a and 2b. The results show that the method has an average accuracy of 79.1% and 83.1%, which outperforms other existing methods. Using TWSB feature selection instead of selecting all features improves the accuracy by up to about 6%. Moreover, the TWSB selection method can effectively reduce the computational burden. We believe that the framework reveals more interpretable feature information of motor imagery EEG signals, provides neural responses discriminative with high accuracy, and facilitates the performance of real-time MI-BCI.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"2961-2973"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140899396","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We aimed to investigate the electrocardiogram (ECG) features in persons with chronic disorders of consciousness (DOC, ≥ 29 days since injury, DSI) resulted from the most severe brain damages. The ECG data from 30 patients with chronic DOC and 18 healthy controls (HCs) were recorded during resting wakefulness state for about five minutes. The patients were classified into vegetative state (VS) and minimally conscious state (MCS). Eight ECG metrics were extracted for comparisons between the subject subgroups, and regression analysis of the metrics were conducted on the DSI (29-593 days). The DOC patients exhibit a significantly higher heart rate (HR, p = 0.009) and lower values for SDNN (p = 0.001), CVRR (p = 0.009), and T-wave amplitude (p < 0.001) compared to the HCs. However, there're no significant differences in QRS, QT, QTc, or ST amplitude between the two groups (p > 0.05). Three ECG metrics of the DOC patients-HR, SDNN, and CVRR-are significantly correlated with the DSI. The ECG abnormalities persist in chronic DOC patients. The abnormalities are mainly manifested in the rhythm features HR, SDNN and CVRR, but not the waveform features such as QRS width, QT, QTc, ST and T-wave amplitudes.
{"title":"The ECG abnormalities in persons with chronic disorders of consciousness.","authors":"Xiaodan Tan, Minmin Luo, Qiuyi Xiao, Xiaochun Zheng, Jiajia Kang, Daogang Zha, Qiuyou Xie, Chang'an A Zhan","doi":"10.1007/s11517-024-03129-5","DOIUrl":"10.1007/s11517-024-03129-5","url":null,"abstract":"<p><p>We aimed to investigate the electrocardiogram (ECG) features in persons with chronic disorders of consciousness (DOC, ≥ 29 days since injury, DSI) resulted from the most severe brain damages. The ECG data from 30 patients with chronic DOC and 18 healthy controls (HCs) were recorded during resting wakefulness state for about five minutes. The patients were classified into vegetative state (VS) and minimally conscious state (MCS). Eight ECG metrics were extracted for comparisons between the subject subgroups, and regression analysis of the metrics were conducted on the DSI (29-593 days). The DOC patients exhibit a significantly higher heart rate (HR, p = 0.009) and lower values for SDNN (p = 0.001), CVRR (p = 0.009), and T-wave amplitude (p < 0.001) compared to the HCs. However, there're no significant differences in QRS, QT, QTc, or ST amplitude between the two groups (p > 0.05). Three ECG metrics of the DOC patients-HR, SDNN, and CVRR-are significantly correlated with the DSI. The ECG abnormalities persist in chronic DOC patients. The abnormalities are mainly manifested in the rhythm features HR, SDNN and CVRR, but not the waveform features such as QRS width, QT, QTc, ST and T-wave amplitudes.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3013-3023"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140946346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-17DOI: 10.1007/s11517-024-03117-9
Mohammad Reza Rezazadeh, Alireza Dastan, Sasan Sadrizadeh, Omid Abouali
The impact of drug delivery and particulate matter exposure on the human respiratory tract is influenced by various anatomical and physiological factors, particularly the structure of the respiratory tract and its fluid dynamics. This study employs computational fluid dynamics (CFD) to investigate airflow in two 3D models of the human air conducting zone. The first model uses a combination of CT-scan images and geometrical data from human cadaver to extract the upper and central airways down to the ninth generation, while the second model develops the lung airways from the first Carina to the end of the ninth generation using Kitaoka's deterministic algorithm. The study examines the differences in geometrical characteristics, airflow rates, velocity, Reynolds number, and pressure drops of both models in the inhalation and exhalation phases for different lobes and generations of the airways. From trachea to the ninth generation, the average air flowrates and Reynolds numbers exponentially decay in both models during inhalation and exhalation. The steady drop is the case for the average air velocity in Kitaoka's model, while that experiences a maximum in the 3rd or 4th generation in the quasi-realistic model. Besides, it is shown that the flow field remains laminar in the upper and central airways up to the total flow rate of 15 l/min. The results of this work can contribute to the understanding of flow behavior in upper respiratory tract.
{"title":"A quasi-realistic computational model development and flow field study of the human upper and central airways.","authors":"Mohammad Reza Rezazadeh, Alireza Dastan, Sasan Sadrizadeh, Omid Abouali","doi":"10.1007/s11517-024-03117-9","DOIUrl":"10.1007/s11517-024-03117-9","url":null,"abstract":"<p><p>The impact of drug delivery and particulate matter exposure on the human respiratory tract is influenced by various anatomical and physiological factors, particularly the structure of the respiratory tract and its fluid dynamics. This study employs computational fluid dynamics (CFD) to investigate airflow in two 3D models of the human air conducting zone. The first model uses a combination of CT-scan images and geometrical data from human cadaver to extract the upper and central airways down to the ninth generation, while the second model develops the lung airways from the first Carina to the end of the ninth generation using Kitaoka's deterministic algorithm. The study examines the differences in geometrical characteristics, airflow rates, velocity, Reynolds number, and pressure drops of both models in the inhalation and exhalation phases for different lobes and generations of the airways. From trachea to the ninth generation, the average air flowrates and Reynolds numbers exponentially decay in both models during inhalation and exhalation. The steady drop is the case for the average air velocity in Kitaoka's model, while that experiences a maximum in the 3rd or 4th generation in the quasi-realistic model. Besides, it is shown that the flow field remains laminar in the upper and central airways up to the total flow rate of 15 l/min. The results of this work can contribute to the understanding of flow behavior in upper respiratory tract.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3025-3041"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140959116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In clinical practice, the morphology of the left atrial appendage (LAA) plays an important role in the selection of LAA closure devices for LAA closure procedures. The morphology determination is influenced by the segmentation results. The LAA occupies only a small part of the entire 3D medical image, and the segmentation results are more likely to be biased towards the background region, making the segmentation of the LAA challenging. In this paper, we propose a lightweight attention mechanism called fusion attention, which imitates human visual behavior. We process the 3D image of the LAA using a method that involves overview observation followed by detailed observation. In the overview observation stage, the image features are pooled along the three dimensions of length, width, and height. The obtained features from the three dimensions are then separately input into the spatial attention and channel attention modules to learn the regions of interest. In the detailed observation stage, the attention results from the previous stage are fused using element-wise multiplication and combined with the original feature map to enhance feature learning. The fusion attention mechanism was evaluated on a left atrial appendage dataset provided by Liaoning Provincial People's Hospital, resulting in an average Dice coefficient of 0.8855. The results indicate that the fusion attention mechanism achieves better segmentation results on 3D images compared to existing lightweight attention mechanisms.
{"title":"Segmentation of the left atrial appendage based on fusion attention.","authors":"Guodong Zhang, Kaichao Liang, Yanlin Li, Tingyu Liang, Zhaoxuan Gong, Ronghui Ju, Dazhe Zhao, Zhuoning Zhang","doi":"10.1007/s11517-024-03104-0","DOIUrl":"10.1007/s11517-024-03104-0","url":null,"abstract":"<p><p>In clinical practice, the morphology of the left atrial appendage (LAA) plays an important role in the selection of LAA closure devices for LAA closure procedures. The morphology determination is influenced by the segmentation results. The LAA occupies only a small part of the entire 3D medical image, and the segmentation results are more likely to be biased towards the background region, making the segmentation of the LAA challenging. In this paper, we propose a lightweight attention mechanism called fusion attention, which imitates human visual behavior. We process the 3D image of the LAA using a method that involves overview observation followed by detailed observation. In the overview observation stage, the image features are pooled along the three dimensions of length, width, and height. The obtained features from the three dimensions are then separately input into the spatial attention and channel attention modules to learn the regions of interest. In the detailed observation stage, the attention results from the previous stage are fused using element-wise multiplication and combined with the original feature map to enhance feature learning. The fusion attention mechanism was evaluated on a left atrial appendage dataset provided by Liaoning Provincial People's Hospital, resulting in an average Dice coefficient of 0.8855. The results indicate that the fusion attention mechanism achieves better segmentation results on 3D images compared to existing lightweight attention mechanisms.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"2999-3012"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140899456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-25DOI: 10.1007/s11517-024-03130-y
Siyu Liu, Haoran Wang, Shiman Li, Chenxi Zhang
Accurate brain tumor segmentation with multi-modal MRI images is crucial, but missing modalities in clinical practice often reduce accuracy. The aim of this study is to propose a mixture-of-experts and semantic-guided network to tackle the issue of missing modalities in brain tumor segmentation. We introduce a transformer-based encoder with novel mixture-of-experts blocks. In each block, four modality experts aim for modality-specific feature learning. Learnable modality embeddings are employed to alleviate the negative effect of missing modalities. We also introduce a decoder guided by semantic information, designed to pay higher attention to various tumor regions. Finally, we conduct extensive comparison experiments with other models as well as ablation experiments to validate the performance of the proposed model on the BraTS2018 dataset. The proposed model can accurately segment brain tumor sub-regions even with missing modalities. It achieves an average Dice score of 0.81 for the whole tumor, 0.66 for the tumor core, and 0.52 for the enhanced tumor across the 15 modality combinations, achieving top or near-top results in most cases, while also exhibiting a lower computational cost. Our mixture-of-experts and sematic-guided network achieves accurate and reliable brain tumor segmentation results with missing modalities, indicating its significant potential for clinical applications. Our source code is already available at https://github.com/MaggieLSY/MESG-Net .
{"title":"Mixture-of-experts and semantic-guided network for brain tumor segmentation with missing MRI modalities.","authors":"Siyu Liu, Haoran Wang, Shiman Li, Chenxi Zhang","doi":"10.1007/s11517-024-03130-y","DOIUrl":"10.1007/s11517-024-03130-y","url":null,"abstract":"<p><p>Accurate brain tumor segmentation with multi-modal MRI images is crucial, but missing modalities in clinical practice often reduce accuracy. The aim of this study is to propose a mixture-of-experts and semantic-guided network to tackle the issue of missing modalities in brain tumor segmentation. We introduce a transformer-based encoder with novel mixture-of-experts blocks. In each block, four modality experts aim for modality-specific feature learning. Learnable modality embeddings are employed to alleviate the negative effect of missing modalities. We also introduce a decoder guided by semantic information, designed to pay higher attention to various tumor regions. Finally, we conduct extensive comparison experiments with other models as well as ablation experiments to validate the performance of the proposed model on the BraTS2018 dataset. The proposed model can accurately segment brain tumor sub-regions even with missing modalities. It achieves an average Dice score of 0.81 for the whole tumor, 0.66 for the tumor core, and 0.52 for the enhanced tumor across the 15 modality combinations, achieving top or near-top results in most cases, while also exhibiting a lower computational cost. Our mixture-of-experts and sematic-guided network achieves accurate and reliable brain tumor segmentation results with missing modalities, indicating its significant potential for clinical applications. Our source code is already available at https://github.com/MaggieLSY/MESG-Net .</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3179-3191"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141092267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The current diagnosis of diabetic retinopathy is based on fundus images and clinical experience. However, considering the ineffectiveness and non-portability of medical devices, we aimed to develop a diagnostic model for diabetic retinopathy based on glucose series data from the wearable continuous glucose monitoring system. Therefore, this study developed a novel method, i.e., double deep latent autoencoder, for exploring glycemic variability influence from multi-day glucose data for diabetic retinopathy. Specifically, the model proposed in this research could encode continuous glucose sensor data with non-continuous and variable length via the integration of a data reorganization module and a novel encoding module with fragmented-missing-wise objective function. Additionally, the model implements a double deep autoencoder, which integrated convolutional neural network, long short-term memory, to jointly capturing the inter-day and intra-day glucose latent features from glucose series. The effectiveness of the proposed model is evaluated through a cross-validation method to clinical datasets of 765 type 2 diabetes patients. The proposed method achieves the highest accuracy value (0.89), precision value (0.88), and F1 score (0.73). The results suggest that our model can be used to remotely diagnose and screen for diabetic retinopathy by learning potential features of glucose series data collected by wearable continuous glucose monitoring systems.
目前,糖尿病视网膜病变的诊断主要基于眼底图像和临床经验。然而,考虑到医疗设备的低效性和不可携带性,我们的目标是根据可穿戴连续血糖监测系统的血糖序列数据开发糖尿病视网膜病变诊断模型。因此,本研究开发了一种新方法,即双深潜自编码器,用于从多日血糖数据中探索血糖变异对糖尿病视网膜病变的影响。具体来说,本研究提出的模型通过整合数据重组模块和具有碎片-缺失-明智目标函数的新型编码模块,可对非连续且长度可变的连续葡萄糖传感器数据进行编码。此外,该模型还实现了双深度自动编码器,该编码器集成了卷积神经网络和长短期记忆,可联合捕捉葡萄糖序列中的日间和日内葡萄糖潜特征。通过对 765 名 2 型糖尿病患者的临床数据集进行交叉验证,评估了所提模型的有效性。所提出的方法获得了最高的准确度值(0.89)、精确度值(0.88)和 F1 分数(0.73)。结果表明,通过学习可穿戴连续血糖监测系统收集的血糖序列数据的潜在特征,我们的模型可用于远程诊断和筛查糖尿病视网膜病变。
{"title":"DDLA: a double deep latent autoencoder for diabetic retinopathy diagnose based on continuous glucose sensors.","authors":"Rui Tao, Hongru Li, Jingyi Lu, Youhe Huang, Yaxin Wang, Wei Lu, Xiaopeng Shao, Jian Zhou, Xia Yu","doi":"10.1007/s11517-024-03120-0","DOIUrl":"10.1007/s11517-024-03120-0","url":null,"abstract":"<p><p>The current diagnosis of diabetic retinopathy is based on fundus images and clinical experience. However, considering the ineffectiveness and non-portability of medical devices, we aimed to develop a diagnostic model for diabetic retinopathy based on glucose series data from the wearable continuous glucose monitoring system. Therefore, this study developed a novel method, i.e., double deep latent autoencoder, for exploring glycemic variability influence from multi-day glucose data for diabetic retinopathy. Specifically, the model proposed in this research could encode continuous glucose sensor data with non-continuous and variable length via the integration of a data reorganization module and a novel encoding module with fragmented-missing-wise objective function. Additionally, the model implements a double deep autoencoder, which integrated convolutional neural network, long short-term memory, to jointly capturing the inter-day and intra-day glucose latent features from glucose series. The effectiveness of the proposed model is evaluated through a cross-validation method to clinical datasets of 765 type 2 diabetes patients. The proposed method achieves the highest accuracy value (0.89), precision value (0.88), and F1 score (0.73). The results suggest that our model can be used to remotely diagnose and screen for diabetic retinopathy by learning potential features of glucose series data collected by wearable continuous glucose monitoring systems.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3089-3106"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141076646","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-24DOI: 10.1007/s11517-024-03128-6
Swati Sharma, Martin L Buist
The gastrointestinal (GI) peristalsis is an involuntary wave-like contraction of the GI wall that helps to propagate food along the tract. Many GI diseases, e.g., gastroparesis, are known to cause motility disorders in which the physiological contractile patterns of the wall get disrupted. Therefore, to understand the pathophysiology of these diseases, it is necessary to understand the mechanism of GI motility. We present a coupled electromechanical model to describe the mechanism of GI motility and the transduction pathway of cellular electrical activities into mechanical deformation and the generation of intraluminal pressure (IP) waves in the GI tract. The proposed model consolidates a smooth muscle cell (SMC) model, an actin-myosin interaction model, a hyperelastic constitutive model, and a Windkessel model to construct a coupled model that can describe the origin of peristaltic contractions in the intestine. The key input to the model is external electrical stimuli, which are converted into mechanical contractile waves in the wall. The model recreated experimental observations efficiently and was able to establish a relationship between change in luminal volume and pressure with the compliance of the GI wall and the peripheral resistance to bolus flow. The proposed model will help us understand the GI tract's function in physiological and pathophysiological conditions.
{"title":"The origin of intraluminal pressure waves in gastrointestinal tract.","authors":"Swati Sharma, Martin L Buist","doi":"10.1007/s11517-024-03128-6","DOIUrl":"10.1007/s11517-024-03128-6","url":null,"abstract":"<p><p>The gastrointestinal (GI) peristalsis is an involuntary wave-like contraction of the GI wall that helps to propagate food along the tract. Many GI diseases, e.g., gastroparesis, are known to cause motility disorders in which the physiological contractile patterns of the wall get disrupted. Therefore, to understand the pathophysiology of these diseases, it is necessary to understand the mechanism of GI motility. We present a coupled electromechanical model to describe the mechanism of GI motility and the transduction pathway of cellular electrical activities into mechanical deformation and the generation of intraluminal pressure (IP) waves in the GI tract. The proposed model consolidates a smooth muscle cell (SMC) model, an actin-myosin interaction model, a hyperelastic constitutive model, and a Windkessel model to construct a coupled model that can describe the origin of peristaltic contractions in the intestine. The key input to the model is external electrical stimuli, which are converted into mechanical contractile waves in the wall. The model recreated experimental observations efficiently and was able to establish a relationship between change in luminal volume and pressure with the compliance of the GI wall and the peripheral resistance to bolus flow. The proposed model will help us understand the GI tract's function in physiological and pathophysiological conditions.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3151-3161"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141087748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-01Epub Date: 2024-05-18DOI: 10.1007/s11517-024-03122-y
A Velayudham, K Madhan Kumar, Krishna Priya M S
Medical image denoising has been a subject of extensive research, with various techniques employed to enhance image quality and facilitate more accurate diagnostics. The evolution of denoising methods has highlighted impressive results but struggled to strike equilibrium between noise reduction and edge preservation which limits its applicability in various domains. This paper manifests the novel methodology that integrates an adaptive masking strategy, transformer-based U-Net Prior generator, edge enhancement module, and modified non-local block (MNLB) for denoising brain MRI clinical images. The adaptive masking strategy maintains the vital information through dynamic mask generation while the prior generator by capturing hierarchical features regenerates the high-quality prior MRI images. Finally, these images are fed to the edge enhancement module to boost structural information by maintaining crucial edge details, and the MNLB produces the denoised output by deriving non-local contextual information. The comprehensive experimental assessment is performed by employing two datasets namely the brain tumor MRI dataset and Alzheimer's dataset for diverse metrics and compared with conventional denoising approaches. The proposed denoising methodology achieves a PSNR of 40.965 and SSIM of 0.938 on the Alzheimer's dataset and also achieves a PSNR of 40.002 and SSIM of 0.926 on the brain tumor MRI dataset at a noise level of 50% revealing its supremacy in noise minimization. Furthermore, the impact of different masking ratios on denoising performance is analyzed which reveals that the proposed method showed PSNR of 40.965, SSIM of 0.938, MAE of 5.847, and MSE of 3.672 at the masking ratio of 60%. Moreover, the findings pave the way for the advancement of clinical image processing, facilitating precise detection of tumors in clinical MRI images.
{"title":"Enhancing clinical diagnostics: novel denoising methodology for brain MRI with adaptive masking and modified non-local block.","authors":"A Velayudham, K Madhan Kumar, Krishna Priya M S","doi":"10.1007/s11517-024-03122-y","DOIUrl":"10.1007/s11517-024-03122-y","url":null,"abstract":"<p><p>Medical image denoising has been a subject of extensive research, with various techniques employed to enhance image quality and facilitate more accurate diagnostics. The evolution of denoising methods has highlighted impressive results but struggled to strike equilibrium between noise reduction and edge preservation which limits its applicability in various domains. This paper manifests the novel methodology that integrates an adaptive masking strategy, transformer-based U-Net Prior generator, edge enhancement module, and modified non-local block (MNLB) for denoising brain MRI clinical images. The adaptive masking strategy maintains the vital information through dynamic mask generation while the prior generator by capturing hierarchical features regenerates the high-quality prior MRI images. Finally, these images are fed to the edge enhancement module to boost structural information by maintaining crucial edge details, and the MNLB produces the denoised output by deriving non-local contextual information. The comprehensive experimental assessment is performed by employing two datasets namely the brain tumor MRI dataset and Alzheimer's dataset for diverse metrics and compared with conventional denoising approaches. The proposed denoising methodology achieves a PSNR of 40.965 and SSIM of 0.938 on the Alzheimer's dataset and also achieves a PSNR of 40.002 and SSIM of 0.926 on the brain tumor MRI dataset at a noise level of 50% revealing its supremacy in noise minimization. Furthermore, the impact of different masking ratios on denoising performance is analyzed which reveals that the proposed method showed PSNR of 40.965, SSIM of 0.938, MAE of 5.847, and MSE of 3.672 at the masking ratio of 60%. Moreover, the findings pave the way for the advancement of clinical image processing, facilitating precise detection of tumors in clinical MRI images.</p>","PeriodicalId":49840,"journal":{"name":"Medical & Biological Engineering & Computing","volume":" ","pages":"3043-3056"},"PeriodicalIF":2.6,"publicationDate":"2024-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140959489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}