Pub Date : 2024-08-15DOI: 10.1007/s10278-024-01218-3
Mediha Erturk, Muhammet Üsame Öziç, Melek Tassoker
Periodontal disease is a significant global oral health problem. Radiographic staging is critical in determining periodontitis severity and treatment requirements. This study aims to automatically stage periodontal bone loss using a deep learning approach using bite-wing images. A total of 1752 bite-wing images were used for the study. Radiological examinations were classified into 4 groups. Healthy (normal), no bone loss; stage I (mild destruction), bone loss in the coronal third (< 15%); stage II (moderate destruction), bone loss is in the coronal third and from 15 to 33% (15-33%); stage III-IV (severe destruction), bone loss extending from the middle third to the apical third with furcation destruction (> 33%). All images were converted to 512 × 400 dimensions using bilinear interpolation. The data was divided into 80% training validation and 20% testing. The classification module of the YOLOv8 deep learning model was used for the artificial intelligence-based classification of the images. Based on four class results, it was trained using fivefold cross-validation after transfer learning and fine tuning. After the training, 20% of test data, which the system had never seen, were analyzed using the artificial intelligence weights obtained in each cross-validation. Training and test results were calculated with average accuracy, precision, recall, and F1-score performance metrics. Test images were analyzed with Eigen-CAM explainability heat maps. In the classification of bite-wing images as healthy, mild destruction, moderate destruction, and severe destruction, training performance results were 86.100% accuracy, 84.790% precision, 82.350% recall, and 84.411% F1-score, and test performance results were 83.446% accuracy, 81.742% precision, 80.883% recall, and 81.090% F1-score. The deep learning model gave successful results in staging periodontal bone loss in bite-wing images. Classification scores were relatively high for normal (no bone loss) and severe bone loss in bite-wing images, as they are more clearly visible than mild and moderate damage.
{"title":"Deep Convolutional Neural Network for Automated Staging of Periodontal Bone Loss Severity on Bite-wing Radiographs: An Eigen-CAM Explainability Mapping Approach.","authors":"Mediha Erturk, Muhammet Üsame Öziç, Melek Tassoker","doi":"10.1007/s10278-024-01218-3","DOIUrl":"https://doi.org/10.1007/s10278-024-01218-3","url":null,"abstract":"<p><p>Periodontal disease is a significant global oral health problem. Radiographic staging is critical in determining periodontitis severity and treatment requirements. This study aims to automatically stage periodontal bone loss using a deep learning approach using bite-wing images. A total of 1752 bite-wing images were used for the study. Radiological examinations were classified into 4 groups. Healthy (normal), no bone loss; stage I (mild destruction), bone loss in the coronal third (< 15%); stage II (moderate destruction), bone loss is in the coronal third and from 15 to 33% (15-33%); stage III-IV (severe destruction), bone loss extending from the middle third to the apical third with furcation destruction (> 33%). All images were converted to 512 × 400 dimensions using bilinear interpolation. The data was divided into 80% training validation and 20% testing. The classification module of the YOLOv8 deep learning model was used for the artificial intelligence-based classification of the images. Based on four class results, it was trained using fivefold cross-validation after transfer learning and fine tuning. After the training, 20% of test data, which the system had never seen, were analyzed using the artificial intelligence weights obtained in each cross-validation. Training and test results were calculated with average accuracy, precision, recall, and F1-score performance metrics. Test images were analyzed with Eigen-CAM explainability heat maps. In the classification of bite-wing images as healthy, mild destruction, moderate destruction, and severe destruction, training performance results were 86.100% accuracy, 84.790% precision, 82.350% recall, and 84.411% F1-score, and test performance results were 83.446% accuracy, 81.742% precision, 80.883% recall, and 81.090% F1-score. The deep learning model gave successful results in staging periodontal bone loss in bite-wing images. Classification scores were relatively high for normal (no bone loss) and severe bone loss in bite-wing images, as they are more clearly visible than mild and moderate damage.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141989824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-15DOI: 10.1007/s10278-024-01226-3
Rongguo Zhang, Chenhao Pei, Ji Shi, Shaokang Wang
In the field of deep learning for medical image analysis, training models from scratch are often used and sometimes, transfer learning from pretrained parameters on ImageNet models is also adopted. However, there is no universally accepted medical image dataset specifically designed for pretraining models currently. The purpose of this study is to construct such a general dataset and validate its effectiveness on downstream medical imaging tasks, including classification and segmentation. In this work, we first build a medical image dataset by collecting several public medical image datasets (CPMID). And then, some pretrained models used for transfer learning are obtained based on CPMID. Various-complexity Resnet and the Vision Transformer network are used as the backbone architectures. In the tasks of classification and segmentation on three other datasets, we compared the experimental results of training from scratch, from the pretrained parameters on ImageNet, and from the pretrained parameters on CPMID. Accuracy, the area under the receiver operating characteristic curve, and class activation map are used as metrics for classification performance. Intersection over Union as the metric is for segmentation evaluation. Utilizing the pretrained parameters on the constructed dataset CPMID, we achieved the best classification accuracy, weighted accuracy, and ROC-AUC values on three validation datasets. Notably, the average classification accuracy outperformed ImageNet-based results by 4.30%, 8.86%, and 3.85% respectively. Furthermore, we achieved the optimal balanced outcome of performance and efficiency in both classification and segmentation tasks. The pretrained parameters on the proposed dataset CPMID are very effective for common tasks in medical image analysis such as classification and segmentation.
{"title":"Construction and Validation of a General Medical Image Dataset for Pretraining.","authors":"Rongguo Zhang, Chenhao Pei, Ji Shi, Shaokang Wang","doi":"10.1007/s10278-024-01226-3","DOIUrl":"https://doi.org/10.1007/s10278-024-01226-3","url":null,"abstract":"<p><p>In the field of deep learning for medical image analysis, training models from scratch are often used and sometimes, transfer learning from pretrained parameters on ImageNet models is also adopted. However, there is no universally accepted medical image dataset specifically designed for pretraining models currently. The purpose of this study is to construct such a general dataset and validate its effectiveness on downstream medical imaging tasks, including classification and segmentation. In this work, we first build a medical image dataset by collecting several public medical image datasets (CPMID). And then, some pretrained models used for transfer learning are obtained based on CPMID. Various-complexity Resnet and the Vision Transformer network are used as the backbone architectures. In the tasks of classification and segmentation on three other datasets, we compared the experimental results of training from scratch, from the pretrained parameters on ImageNet, and from the pretrained parameters on CPMID. Accuracy, the area under the receiver operating characteristic curve, and class activation map are used as metrics for classification performance. Intersection over Union as the metric is for segmentation evaluation. Utilizing the pretrained parameters on the constructed dataset CPMID, we achieved the best classification accuracy, weighted accuracy, and ROC-AUC values on three validation datasets. Notably, the average classification accuracy outperformed ImageNet-based results by 4.30%, 8.86%, and 3.85% respectively. Furthermore, we achieved the optimal balanced outcome of performance and efficiency in both classification and segmentation tasks. The pretrained parameters on the proposed dataset CPMID are very effective for common tasks in medical image analysis such as classification and segmentation.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141989823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-15DOI: 10.1007/s10278-024-01227-2
Kyung Hwa Lee, Jungwook Lee, Gwang Hyeon Choi, Jihye Yun, Jiseon Kang, Jonggi Choi, Kang Mo Kim, Namkug Kim
The objective of this study was to develop and evaluate a model for predicting post-treatment survival in hepatocellular carcinoma (HCC) patients using their CT images and clinical information, including various treatment information. We collected pre-treatment contrast-enhanced CT images and clinical information including patient-related factors, initial treatment options, and survival status from 692 patients. The patient cohort was divided into a training cohort (n = 507), a testing cohort (n = 146), and an external CT cohort (n = 39), which included patients who underwent CT scans at other institutions. After model training using fivefold cross-validation, model validation was performed on both the testing cohort and the external CT cohort. Our cascaded model employed a 3D convolutional neural network (CNN) to extract features from CT images and derive final survival probabilities. These probabilities were obtained by concatenating previously predicted probabilities for each interval with the patient-related factors and treatment options. We utilized two consecutive fully connected layers for this process, resulting in a number of final outputs corresponding to the number of time intervals, with values representing conditional survival probabilities for each interval. Performance was assessed using the concordance index (C-index), the mean cumulative/dynamic area under the receiver operating characteristics curve (mC/D AUC), and the mean Brier score (mBS), calculated every 3 months. Through an ablation study, we found that using DenseNet-121 as the backbone network and setting the prediction interval to 6 months optimized the model's performance. The integration of multimodal data resulted in superior predictive capabilities compared to models using only CT images or clinical information (C index 0.824 [95% CI 0.822-0.826], mC/D AUC 0.893 [95% CI 0.891-0.895], and mBS 0.121 [95% CI 0.120-0.123] for internal test cohort; C index 0.750 [95% CI 0.747-0.753], mC/D AUC 0.819 [95% CI 0.816-0.823], and mBS 0.159 [95% CI 0.158-0.161] for external CT cohort, respectively). Our CNN-based discrete-time survival prediction model with CT images and clinical information demonstrated promising results in predicting post-treatment survival of patients with HCC.
本研究的目的是利用肝细胞癌(HCC)患者的 CT 图像和临床信息(包括各种治疗信息),开发并评估一个预测患者治疗后生存期的模型。我们收集了 692 名患者的治疗前对比增强 CT 图像和临床信息,包括患者相关因素、初始治疗方案和生存状况。患者队列分为训练队列(507 人)、测试队列(146 人)和外部 CT 队列(39 人),外部 CT 队列包括在其他机构接受 CT 扫描的患者。使用五倍交叉验证进行模型训练后,在测试队列和外部 CT 队列中进行模型验证。我们的级联模型采用三维卷积神经网络(CNN)从 CT 图像中提取特征,并得出最终的生存概率。这些概率是通过将先前预测的每个区间的概率与患者相关因素和治疗方案连接起来而获得的。在这一过程中,我们使用了两个连续的全连接层,最终输出的数量与时间间隔的数量相对应,其值代表了每个时间间隔的条件生存概率。使用一致性指数(C-index)、接收器操作特征曲线下的平均累积/动态面积(mC/D AUC)以及每 3 个月计算一次的平均布赖尔评分(mBS)来评估性能。通过一项消融研究,我们发现使用 DenseNet-121 作为骨干网络并将预测时间间隔设定为 6 个月可以优化模型的性能。与仅使用 CT 图像或临床信息的模型相比,多模态数据的整合带来了更优越的预测能力(C 指数 0.824 [95% CI 0.822-0.826], mC/D AUC 0.893 [95% CI 0.891-0.895]和 mBS 0.121 [95% CI 0.120-0.123];外部 CT 队列的 C 指数分别为 0.750 [95% CI 0.747-0.753]、mC/D AUC 0.819 [95% CI 0.816-0.823] 和 mBS 0.159 [95% CI 0.158-0.161])。我们的基于CNN的离散时间生存预测模型结合了CT图像和临床信息,在预测HCC患者的治疗后生存率方面取得了良好的效果。
{"title":"Deep Learning-Based Prediction of Post-treatment Survival in Hepatocellular Carcinoma Patients Using Pre-treatment CT Images and Clinical Data.","authors":"Kyung Hwa Lee, Jungwook Lee, Gwang Hyeon Choi, Jihye Yun, Jiseon Kang, Jonggi Choi, Kang Mo Kim, Namkug Kim","doi":"10.1007/s10278-024-01227-2","DOIUrl":"https://doi.org/10.1007/s10278-024-01227-2","url":null,"abstract":"<p><p>The objective of this study was to develop and evaluate a model for predicting post-treatment survival in hepatocellular carcinoma (HCC) patients using their CT images and clinical information, including various treatment information. We collected pre-treatment contrast-enhanced CT images and clinical information including patient-related factors, initial treatment options, and survival status from 692 patients. The patient cohort was divided into a training cohort (n = 507), a testing cohort (n = 146), and an external CT cohort (n = 39), which included patients who underwent CT scans at other institutions. After model training using fivefold cross-validation, model validation was performed on both the testing cohort and the external CT cohort. Our cascaded model employed a 3D convolutional neural network (CNN) to extract features from CT images and derive final survival probabilities. These probabilities were obtained by concatenating previously predicted probabilities for each interval with the patient-related factors and treatment options. We utilized two consecutive fully connected layers for this process, resulting in a number of final outputs corresponding to the number of time intervals, with values representing conditional survival probabilities for each interval. Performance was assessed using the concordance index (C-index), the mean cumulative/dynamic area under the receiver operating characteristics curve (mC/D AUC), and the mean Brier score (mBS), calculated every 3 months. Through an ablation study, we found that using DenseNet-121 as the backbone network and setting the prediction interval to 6 months optimized the model's performance. The integration of multimodal data resulted in superior predictive capabilities compared to models using only CT images or clinical information (C index 0.824 [95% CI 0.822-0.826], mC/D AUC 0.893 [95% CI 0.891-0.895], and mBS 0.121 [95% CI 0.120-0.123] for internal test cohort; C index 0.750 [95% CI 0.747-0.753], mC/D AUC 0.819 [95% CI 0.816-0.823], and mBS 0.159 [95% CI 0.158-0.161] for external CT cohort, respectively). Our CNN-based discrete-time survival prediction model with CT images and clinical information demonstrated promising results in predicting post-treatment survival of patients with HCC.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141989825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-13DOI: 10.1007/s10278-024-01201-y
Rupali Vyas, Deepak Rao Khadatkar
Pneumonia is a severe health concern, particularly for vulnerable groups, needing early and correct classification for optimal treatment. This study addresses the use of deep learning combined with machine learning classifiers (DLxMLCs) for pneumonia classification from chest X-ray (CXR) images. We deployed modified VGG19, ResNet50V2, and DenseNet121 models for feature extraction, followed by five machine learning classifiers (logistic regression, support vector machine, decision tree, random forest, artificial neural network). The approach we suggested displayed remarkable accuracy, with VGG19 and DenseNet121 models obtaining 99.98% accuracy when combined with random forest or decision tree classifiers. ResNet50V2 achieved 99.25% accuracy with random forest. These results illustrate the advantages of merging deep learning models with machine learning classifiers in boosting the speedy and accurate identification of pneumonia. The study underlines the potential of DLxMLC systems in enhancing diagnostic accuracy and efficiency. By integrating these models into clinical practice, healthcare practitioners could greatly boost patient care and results. Future research should focus on refining these models and exploring their application to other medical imaging tasks, as well as including explainability methodologies to better understand their decision-making processes and build trust in their clinical use. This technique promises promising breakthroughs in medical imaging and patient management.
{"title":"Ensemble of Deep Learning Architectures with Machine Learning for Pneumonia Classification Using Chest X-rays.","authors":"Rupali Vyas, Deepak Rao Khadatkar","doi":"10.1007/s10278-024-01201-y","DOIUrl":"https://doi.org/10.1007/s10278-024-01201-y","url":null,"abstract":"<p><p>Pneumonia is a severe health concern, particularly for vulnerable groups, needing early and correct classification for optimal treatment. This study addresses the use of deep learning combined with machine learning classifiers (DLxMLCs) for pneumonia classification from chest X-ray (CXR) images. We deployed modified VGG19, ResNet50V2, and DenseNet121 models for feature extraction, followed by five machine learning classifiers (logistic regression, support vector machine, decision tree, random forest, artificial neural network). The approach we suggested displayed remarkable accuracy, with VGG19 and DenseNet121 models obtaining 99.98% accuracy when combined with random forest or decision tree classifiers. ResNet50V2 achieved 99.25% accuracy with random forest. These results illustrate the advantages of merging deep learning models with machine learning classifiers in boosting the speedy and accurate identification of pneumonia. The study underlines the potential of DLxMLC systems in enhancing diagnostic accuracy and efficiency. By integrating these models into clinical practice, healthcare practitioners could greatly boost patient care and results. Future research should focus on refining these models and exploring their application to other medical imaging tasks, as well as including explainability methodologies to better understand their decision-making processes and build trust in their clinical use. This technique promises promising breakthroughs in medical imaging and patient management.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141977551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-13DOI: 10.1007/s10278-024-00994-2
Peter Kamel, Adway Kanhere, Pranav Kulkarni, Mazhar Khalid, Rachel Steger, Uttam Bodanapally, Dheeraj Gandhi, Vishwa Parekh, Paul H Yi
Segmentation of infarcts is clinically important in ischemic stroke management and prognostication. It is unclear what role the combination of DWI, ADC, and FLAIR MRI sequences provide for deep learning in infarct segmentation. Recent technologies in model self-configuration have promised greater performance and generalizability through automated optimization. We assessed the utility of DWI, ADC, and FLAIR sequences on ischemic stroke segmentation, compared self-configuring nnU-Net models to conventional U-Net models without manual optimization, and evaluated the generalizability of results on an external clinical dataset. 3D self-configuring nnU-Net models and standard 3D U-Net models with MONAI were trained on 200 infarcts using DWI, ADC, and FLAIR sequences separately and in all combinations. Segmentation results were compared between models using paired t-test comparison on a hold-out test set of 50 cases. The highest performing model was externally validated on a clinical dataset of 50 MRIs. nnU-Net with DWI sequences attained a Dice score of 0.810 ± 0.155. There was no statistically significant difference when DWI sequences were supplemented with ADC and FLAIR images (Dice score of 0.813 ± 0.150; p = 0.15). nnU-Net models significantly outperformed standard U-Net models for all sequence combinations (p < 0.001). On the external dataset, Dice scores measured 0.704 ± 0.199 for positive cases with false positives with intracranial hemorrhage. Highly optimized neural networks such as nnU-Net provide excellent stroke segmentation even when only provided DWI images, without significant improvement from other sequences. This differs from-and significantly outperforms-standard U-Net architectures. Results translated well to the external clinical environment and provide the groundwork for optimized acute stroke segmentation on MRI.
{"title":"Optimizing Acute Stroke Segmentation on MRI Using Deep Learning: Self-Configuring Neural Networks Provide High Performance Using Only DWI Sequences.","authors":"Peter Kamel, Adway Kanhere, Pranav Kulkarni, Mazhar Khalid, Rachel Steger, Uttam Bodanapally, Dheeraj Gandhi, Vishwa Parekh, Paul H Yi","doi":"10.1007/s10278-024-00994-2","DOIUrl":"https://doi.org/10.1007/s10278-024-00994-2","url":null,"abstract":"<p><p>Segmentation of infarcts is clinically important in ischemic stroke management and prognostication. It is unclear what role the combination of DWI, ADC, and FLAIR MRI sequences provide for deep learning in infarct segmentation. Recent technologies in model self-configuration have promised greater performance and generalizability through automated optimization. We assessed the utility of DWI, ADC, and FLAIR sequences on ischemic stroke segmentation, compared self-configuring nnU-Net models to conventional U-Net models without manual optimization, and evaluated the generalizability of results on an external clinical dataset. 3D self-configuring nnU-Net models and standard 3D U-Net models with MONAI were trained on 200 infarcts using DWI, ADC, and FLAIR sequences separately and in all combinations. Segmentation results were compared between models using paired t-test comparison on a hold-out test set of 50 cases. The highest performing model was externally validated on a clinical dataset of 50 MRIs. nnU-Net with DWI sequences attained a Dice score of 0.810 ± 0.155. There was no statistically significant difference when DWI sequences were supplemented with ADC and FLAIR images (Dice score of 0.813 ± 0.150; p = 0.15). nnU-Net models significantly outperformed standard U-Net models for all sequence combinations (p < 0.001). On the external dataset, Dice scores measured 0.704 ± 0.199 for positive cases with false positives with intracranial hemorrhage. Highly optimized neural networks such as nnU-Net provide excellent stroke segmentation even when only provided DWI images, without significant improvement from other sequences. This differs from-and significantly outperforms-standard U-Net architectures. Results translated well to the external clinical environment and provide the groundwork for optimized acute stroke segmentation on MRI.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141977552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The diagnosis and treatment of pulmonary hypertension have changed dramatically through the re-defined diagnostic criteria and advanced drug development in the past decade. The application of Artificial Intelligence for the detection of elevated pulmonary arterial pressure (ePAP) was reported recently. Artificial Intelligence (AI) has demonstrated the capability to identify ePAP and its association with hospitalization due to heart failure when analyzing chest X-rays (CXR). An AI model based on electrocardiograms (ECG) has shown promise in not only detecting ePAP but also in predicting future risks related to cardiovascular mortality. We aimed to develop an AI model integrating ECG and CXR to detect ePAP and evaluate their performance. We developed a deep-learning model (DLM) using paired ECG and CXR to detect ePAP (systolic pulmonary artery pressure > 50 mmHg in transthoracic echocardiography). This model was further validated in a community hospital. Additionally, our DLM was evaluated for its ability to predict future occurrences of left ventricular dysfunction (LVD, ejection fraction < 35%) and cardiovascular mortality. The AUCs for detecting ePAP were as follows: 0.8261 with ECG (sensitivity 76.6%, specificity 74.5%), 0.8525 with CXR (sensitivity 82.8%, specificity 72.7%), and 0.8644 with a combination of both (sensitivity 78.6%, specificity 79.2%) in the internal dataset. In the external validation dataset, the AUCs for ePAP detection were 0.8348 with ECG, 0.8605 with CXR, and 0.8734 with the combination. Furthermore, using the combination of ECGs and CXR, the negative predictive value (NPV) was 98% in the internal dataset and 98.1% in the external dataset. Patients with ePAP detected by the DLM using combination had a higher risk of new-onset LVD with a hazard ratio (HR) of 4.51 (95% CI: 3.54-5.76) in the internal dataset and cardiovascular mortality with a HR of 6.08 (95% CI: 4.66-7.95). Similar results were seen in the external validation dataset. The DLM, integrating ECG and CXR, effectively detected ePAP with a strong NPV and forecasted future risks of developing LVD and cardiovascular mortality. This model has the potential to expedite the early identification of pulmonary hypertension in patients, prompting further evaluation through echocardiography and, when necessary, right heart catheterization (RHC), potentially resulting in enhanced cardiovascular outcomes.
过去十年间,通过重新定义诊断标准和先进的药物开发,肺动脉高压的诊断和治疗发生了巨大变化。最近有报道称,人工智能可用于检测肺动脉压升高(ePAP)。人工智能(AI)已证明有能力在分析胸部 X 光片(CXR)时识别 ePAP 及其与心衰住院的关联。基于心电图(ECG)的人工智能模型不仅能检测出电子心率,还能预测与心血管死亡率相关的未来风险。我们的目标是开发一种整合心电图和 X 光造影的人工智能模型来检测电子心率,并评估其性能。我们开发了一种深度学习模型(DLM),利用成对的心电图和心血管造影来检测 ePAP(经胸超声心动图显示肺动脉收缩压大于 50 mmHg)。该模型在一家社区医院得到了进一步验证。此外,还评估了我们的 DLM 预测未来发生左心室功能障碍(LVD,射血分数
{"title":"A Deep-Learning-Enabled Electrocardiogram and Chest X-Ray for Detecting Pulmonary Arterial Hypertension.","authors":"Pang-Yen Liu, Shi-Chue Hsing, Dung-Jang Tsai, Chin Lin, Chin-Sheng Lin, Chih-Hung Wang, Wen-Hui Fang","doi":"10.1007/s10278-024-01225-4","DOIUrl":"https://doi.org/10.1007/s10278-024-01225-4","url":null,"abstract":"<p><p>The diagnosis and treatment of pulmonary hypertension have changed dramatically through the re-defined diagnostic criteria and advanced drug development in the past decade. The application of Artificial Intelligence for the detection of elevated pulmonary arterial pressure (ePAP) was reported recently. Artificial Intelligence (AI) has demonstrated the capability to identify ePAP and its association with hospitalization due to heart failure when analyzing chest X-rays (CXR). An AI model based on electrocardiograms (ECG) has shown promise in not only detecting ePAP but also in predicting future risks related to cardiovascular mortality. We aimed to develop an AI model integrating ECG and CXR to detect ePAP and evaluate their performance. We developed a deep-learning model (DLM) using paired ECG and CXR to detect ePAP (systolic pulmonary artery pressure > 50 mmHg in transthoracic echocardiography). This model was further validated in a community hospital. Additionally, our DLM was evaluated for its ability to predict future occurrences of left ventricular dysfunction (LVD, ejection fraction < 35%) and cardiovascular mortality. The AUCs for detecting ePAP were as follows: 0.8261 with ECG (sensitivity 76.6%, specificity 74.5%), 0.8525 with CXR (sensitivity 82.8%, specificity 72.7%), and 0.8644 with a combination of both (sensitivity 78.6%, specificity 79.2%) in the internal dataset. In the external validation dataset, the AUCs for ePAP detection were 0.8348 with ECG, 0.8605 with CXR, and 0.8734 with the combination. Furthermore, using the combination of ECGs and CXR, the negative predictive value (NPV) was 98% in the internal dataset and 98.1% in the external dataset. Patients with ePAP detected by the DLM using combination had a higher risk of new-onset LVD with a hazard ratio (HR) of 4.51 (95% CI: 3.54-5.76) in the internal dataset and cardiovascular mortality with a HR of 6.08 (95% CI: 4.66-7.95). Similar results were seen in the external validation dataset. The DLM, integrating ECG and CXR, effectively detected ePAP with a strong NPV and forecasted future risks of developing LVD and cardiovascular mortality. This model has the potential to expedite the early identification of pulmonary hypertension in patients, prompting further evaluation through echocardiography and, when necessary, right heart catheterization (RHC), potentially resulting in enhanced cardiovascular outcomes.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141972525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To evaluate the usefulness of low-keV multiphasic computed tomography (CT) with deep learning image reconstruction (DLIR) in improving the delineation of pancreatic ductal adenocarcinoma (PDAC) compared to conventional hybrid iterative reconstruction (HIR). Thirty-five patients with PDAC who underwent multiphasic CT were retrospectively evaluated. Raw data were reconstructed with two energy levels (40 keV and 70 keV) of virtual monochromatic imaging (VMI) using HIR (ASiR-V50%) and DLIR (TrueFidelity-H). Contrast-to-noise ratio (CNRtumor) was calculated from the CT values within regions of interest in tumor and normal pancreas in the pancreatic parenchymal phase images. Lesion conspicuity of PDAC in pancreatic parenchymal phase on 40-keV HIR, 40-keV DLIR, and 70-keV DLIR images was qualitatively rated on a 5-point scale, using 70-keV HIR images as reference (score 1 = poor; score 3 = equivalent to reference; score 5 = excellent) by two radiologists. CNRtumor of 40-keV DLIR images (median 10.4, interquartile range (IQR) 7.8-14.9) was significantly higher than that of the other VMIs (40 keV HIR, median 6.2, IQR 4.4-8.5, P < 0.0001; 70-keV DLIR, median 6.3, IQR 5.1-9.9, P = 0.0002; 70-keV HIR, median 4.2, IQR 3.1-6.1, P < 0.0001). CNRtumor of 40-keV DLIR images were significantly better than those of the 40-keV HIR and 70-keV HIR images by 72 ± 22% and 211 ± 340%, respectively. Lesion conspicuity scores on 40-keV DLIR images (observer 1, 4.5 ± 0.7; observer 2, 3.4 ± 0.5) were significantly higher than on 40-keV HIR (observer 1, 3.3 ± 0.9, P < 0.0001; observer 2, 3.1 ± 0.4, P = 0.013). DLIR is a promising reconstruction method to improve PDAC delineation in 40-keV VMI at the pancreatic parenchymal phase compared to conventional HIR.
{"title":"The Usefulness of Low-Kiloelectron Volt Virtual Monochromatic Contrast-Enhanced Computed Tomography with Deep Learning Image Reconstruction Technique in Improving the Delineation of Pancreatic Ductal Adenocarcinoma.","authors":"Yasutaka Ichikawa, Yoshinori Kanii, Akio Yamazaki, Mai Kobayashi, Kensuke Domae, Motonori Nagata, Hajime Sakuma","doi":"10.1007/s10278-024-01214-7","DOIUrl":"https://doi.org/10.1007/s10278-024-01214-7","url":null,"abstract":"<p><p>To evaluate the usefulness of low-keV multiphasic computed tomography (CT) with deep learning image reconstruction (DLIR) in improving the delineation of pancreatic ductal adenocarcinoma (PDAC) compared to conventional hybrid iterative reconstruction (HIR). Thirty-five patients with PDAC who underwent multiphasic CT were retrospectively evaluated. Raw data were reconstructed with two energy levels (40 keV and 70 keV) of virtual monochromatic imaging (VMI) using HIR (ASiR-V50%) and DLIR (TrueFidelity-H). Contrast-to-noise ratio (CNR<sub>tumor</sub>) was calculated from the CT values within regions of interest in tumor and normal pancreas in the pancreatic parenchymal phase images. Lesion conspicuity of PDAC in pancreatic parenchymal phase on 40-keV HIR, 40-keV DLIR, and 70-keV DLIR images was qualitatively rated on a 5-point scale, using 70-keV HIR images as reference (score 1 = poor; score 3 = equivalent to reference; score 5 = excellent) by two radiologists. CNR<sub>tumor</sub> of 40-keV DLIR images (median 10.4, interquartile range (IQR) 7.8-14.9) was significantly higher than that of the other VMIs (40 keV HIR, median 6.2, IQR 4.4-8.5, P < 0.0001; 70-keV DLIR, median 6.3, IQR 5.1-9.9, P = 0.0002; 70-keV HIR, median 4.2, IQR 3.1-6.1, P < 0.0001). CNR<sub>tumor</sub> of 40-keV DLIR images were significantly better than those of the 40-keV HIR and 70-keV HIR images by 72 ± 22% and 211 ± 340%, respectively. Lesion conspicuity scores on 40-keV DLIR images (observer 1, 4.5 ± 0.7; observer 2, 3.4 ± 0.5) were significantly higher than on 40-keV HIR (observer 1, 3.3 ± 0.9, P < 0.0001; observer 2, 3.1 ± 0.4, P = 0.013). DLIR is a promising reconstruction method to improve PDAC delineation in 40-keV VMI at the pancreatic parenchymal phase compared to conventional HIR.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141972526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-12DOI: 10.1007/s10278-024-01215-6
Yuqin Min, Jing Li, Shouqiang Jia, Yuehua Li, Shengdong Nie
Time-of-flight magnetic resonance angiography (TOF-MRA) is a non-contrast technique used to visualize neurovascular. However, manual reconstruction of the volume render (VR) by radiologists is time-consuming and labor-intensive. Deep learning-based (DL-based) vessel segmentation technology may provide intelligent automation workflow. To evaluate the image quality of DL vessel segmentation for automatically acquiring intracranial arteries in TOF-MRA. A total of 394 TOF-MRA scans were selected, which included cerebral vascular health, aneurysms, or stenoses. Both our proposed method and two state-of-the-art DL methods are evaluated on external datasets for generalization ability. For qualitative assessment, two experienced clinical radiologists evaluated the image quality of cerebrovascular diagnostic and visualization (scoring 0-5 as unacceptable to excellent) obtained by manual VR reconstruction or automatic convolutional neural network (CNN) segmentation. The proposed CNN outperforms the other two DL-based methods in clinical scoring on external datasets, and its visualization was evaluated by readers as having the appearance of the radiologists' manual reconstructions. Scoring of proposed CNN and VR of intracranial arteries demonstrated good to excellent agreement with no significant differences (median, 5.0 and 5.0, P ≥ 12) at healthy-type scans. All proposed CNN image quality were considered to have adequate diagnostic quality (median scores > 2). Quantitative analysis demonstrated a superior dice similarity coefficient of cerebrovascular overlap (training sets and validation sets; 0.947 and 0.927). Automatic cerebrovascular segmentation using DL is feasible and the image quality in terms of vessel integrity, collateral circulation and lesion morphology is comparable to expert manual VR without significant differences.
{"title":"Automated Cerebrovascular Segmentation and Visualization of Intracranial Time-of-Flight Magnetic Resonance Angiography Based on Deep Learning.","authors":"Yuqin Min, Jing Li, Shouqiang Jia, Yuehua Li, Shengdong Nie","doi":"10.1007/s10278-024-01215-6","DOIUrl":"https://doi.org/10.1007/s10278-024-01215-6","url":null,"abstract":"<p><p>Time-of-flight magnetic resonance angiography (TOF-MRA) is a non-contrast technique used to visualize neurovascular. However, manual reconstruction of the volume render (VR) by radiologists is time-consuming and labor-intensive. Deep learning-based (DL-based) vessel segmentation technology may provide intelligent automation workflow. To evaluate the image quality of DL vessel segmentation for automatically acquiring intracranial arteries in TOF-MRA. A total of 394 TOF-MRA scans were selected, which included cerebral vascular health, aneurysms, or stenoses. Both our proposed method and two state-of-the-art DL methods are evaluated on external datasets for generalization ability. For qualitative assessment, two experienced clinical radiologists evaluated the image quality of cerebrovascular diagnostic and visualization (scoring 0-5 as unacceptable to excellent) obtained by manual VR reconstruction or automatic convolutional neural network (CNN) segmentation. The proposed CNN outperforms the other two DL-based methods in clinical scoring on external datasets, and its visualization was evaluated by readers as having the appearance of the radiologists' manual reconstructions. Scoring of proposed CNN and VR of intracranial arteries demonstrated good to excellent agreement with no significant differences (median, 5.0 and 5.0, P ≥ 12) at healthy-type scans. All proposed CNN image quality were considered to have adequate diagnostic quality (median scores > 2). Quantitative analysis demonstrated a superior dice similarity coefficient of cerebrovascular overlap (training sets and validation sets; 0.947 and 0.927). Automatic cerebrovascular segmentation using DL is feasible and the image quality in terms of vessel integrity, collateral circulation and lesion morphology is comparable to expert manual VR without significant differences.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141918551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-09DOI: 10.1007/s10278-024-01202-x
Fatemeh Ghezloo, Oliver H Chang, Stevan R Knezevich, Kristin C Shaw, Kia Gianni Thigpen, Lisa M Reisch, Linda G Shapiro, Joann G Elmore
Deep learning techniques offer improvements in computer-aided diagnosis systems. However, acquiring image domain annotations is challenging due to the knowledge and commitment required of expert pathologists. Pathologists often identify regions in whole slide images with diagnostic relevance rather than examining the entire slide, with a positive correlation between the time spent on these critical image regions and diagnostic accuracy. In this paper, a heatmap is generated to represent pathologists' viewing patterns during diagnosis and used to guide a deep learning architecture during training. The proposed system outperforms traditional approaches based on color and texture image characteristics, integrating pathologists' domain expertise to enhance region of interest detection without needing individual case annotations. Evaluating our best model, a U-Net model with a pre-trained ResNet-18 encoder, on a skin biopsy whole slide image dataset for melanoma diagnosis, shows its potential in detecting regions of interest, surpassing conventional methods with an increase of 20%, 11%, 22%, and 12% in precision, recall, F1-score, and Intersection over Union, respectively. In a clinical evaluation, three dermatopathologists agreed on the model's effectiveness in replicating pathologists' diagnostic viewing behavior and accurately identifying critical regions. Finally, our study demonstrates that incorporating heatmaps as supplementary signals can enhance the performance of computer-aided diagnosis systems. Without the availability of eye tracking data, identifying precise focus areas is challenging, but our approach shows promise in assisting pathologists in improving diagnostic accuracy and efficiency, streamlining annotation processes, and aiding the training of new pathologists.
{"title":"Robust ROI Detection in Whole Slide Images Guided by Pathologists' Viewing Patterns.","authors":"Fatemeh Ghezloo, Oliver H Chang, Stevan R Knezevich, Kristin C Shaw, Kia Gianni Thigpen, Lisa M Reisch, Linda G Shapiro, Joann G Elmore","doi":"10.1007/s10278-024-01202-x","DOIUrl":"https://doi.org/10.1007/s10278-024-01202-x","url":null,"abstract":"<p><p>Deep learning techniques offer improvements in computer-aided diagnosis systems. However, acquiring image domain annotations is challenging due to the knowledge and commitment required of expert pathologists. Pathologists often identify regions in whole slide images with diagnostic relevance rather than examining the entire slide, with a positive correlation between the time spent on these critical image regions and diagnostic accuracy. In this paper, a heatmap is generated to represent pathologists' viewing patterns during diagnosis and used to guide a deep learning architecture during training. The proposed system outperforms traditional approaches based on color and texture image characteristics, integrating pathologists' domain expertise to enhance region of interest detection without needing individual case annotations. Evaluating our best model, a U-Net model with a pre-trained ResNet-18 encoder, on a skin biopsy whole slide image dataset for melanoma diagnosis, shows its potential in detecting regions of interest, surpassing conventional methods with an increase of 20%, 11%, 22%, and 12% in precision, recall, F1-score, and Intersection over Union, respectively. In a clinical evaluation, three dermatopathologists agreed on the model's effectiveness in replicating pathologists' diagnostic viewing behavior and accurately identifying critical regions. Finally, our study demonstrates that incorporating heatmaps as supplementary signals can enhance the performance of computer-aided diagnosis systems. Without the availability of eye tracking data, identifying precise focus areas is challenging, but our approach shows promise in assisting pathologists in improving diagnostic accuracy and efficiency, streamlining annotation processes, and aiding the training of new pathologists.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141915024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-08-09DOI: 10.1007/s10278-024-01075-0
Yul Hee Lee, Ji-Su Shim, Young Jae Kim, Ji Soo Jeon, Sung-Yoon Kang, Sang Pyo Lee, Sang Min Lee, Kwang Gi Kim
The skin prick test (SPT) is a key tool for identifying sensitized allergens associated with immunoglobulin E-mediated allergic diseases such as asthma, allergic rhinitis, atopic dermatitis, urticaria, angioedema, and anaphylaxis. However, the SPT is labor-intensive and time-consuming due to the necessity of measuring the sizes of the erythema and wheals induced by allergens on the skin. In this study, we used an image preprocessing method and a deep learning model to segment wheals and erythema in SPT images captured by a smartphone camera. Subsequently, we assessed the deep learning model's performance by comparing the results with ground-truth data. Using contrast-limited adaptive histogram equalization (CLAHE), an image preprocessing technique designed to enhance image contrast, we augmented the chromatic contrast in 46 SPT images from 33 participants. We established a deep learning model for wheal and erythema segmentation using 144 and 150 training datasets, respectively. The wheal segmentation model achieved an accuracy of 0.9985, a sensitivity of 0.5621, a specificity of 0.9995, and a Dice similarity coefficient of 0.7079, whereas the erythema segmentation model achieved an accuracy of 0.9660, a sensitivity of 0.5787, a specificity of 0.97977, and a Dice similarity coefficient of 0.6636. The use of image preprocessing and deep learning technology in SPT is expected to have a significant positive impact on medical practice by ensuring the accurate segmentation of wheals and erythema, producing consistent evaluation results, and simplifying diagnostic processes.
{"title":"Allergy Wheal and Erythema Segmentation Using Attention U-Net.","authors":"Yul Hee Lee, Ji-Su Shim, Young Jae Kim, Ji Soo Jeon, Sung-Yoon Kang, Sang Pyo Lee, Sang Min Lee, Kwang Gi Kim","doi":"10.1007/s10278-024-01075-0","DOIUrl":"https://doi.org/10.1007/s10278-024-01075-0","url":null,"abstract":"<p><p>The skin prick test (SPT) is a key tool for identifying sensitized allergens associated with immunoglobulin E-mediated allergic diseases such as asthma, allergic rhinitis, atopic dermatitis, urticaria, angioedema, and anaphylaxis. However, the SPT is labor-intensive and time-consuming due to the necessity of measuring the sizes of the erythema and wheals induced by allergens on the skin. In this study, we used an image preprocessing method and a deep learning model to segment wheals and erythema in SPT images captured by a smartphone camera. Subsequently, we assessed the deep learning model's performance by comparing the results with ground-truth data. Using contrast-limited adaptive histogram equalization (CLAHE), an image preprocessing technique designed to enhance image contrast, we augmented the chromatic contrast in 46 SPT images from 33 participants. We established a deep learning model for wheal and erythema segmentation using 144 and 150 training datasets, respectively. The wheal segmentation model achieved an accuracy of 0.9985, a sensitivity of 0.5621, a specificity of 0.9995, and a Dice similarity coefficient of 0.7079, whereas the erythema segmentation model achieved an accuracy of 0.9660, a sensitivity of 0.5787, a specificity of 0.97977, and a Dice similarity coefficient of 0.6636. The use of image preprocessing and deep learning technology in SPT is expected to have a significant positive impact on medical practice by ensuring the accurate segmentation of wheals and erythema, producing consistent evaluation results, and simplifying diagnostic processes.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141908779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}