Accurate delineation of the clinical target volume (CTV) is a crucial prerequisite for safe and effective radiotherapy characterized. This study addresses the integration of magnetic resonance (MR) images to aid in target delineation on computed tomography (CT) images. However, obtaining MR images directly can be challenging. Therefore, we employ AI-based image generation techniques to “intelligentially generate” MR images from CT images to improve CTV delineation based on CT images. To generate high-quality MR images, we propose an attention-guided single-loop image generation model. The model can yield higher-quality images by introducing an attention mechanism in feature extraction and enhancing the loss function. Based on the generated MR images, we propose a CTV segmentation model fusing multi-scale features through image fusion and a hollow space pyramid module to enhance segmentation accuracy. The image generation model used in this study improves the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) from 14.87 and 0.58 to 16.72 and 0.67, respectively, and improves the feature distribution distance and learning-perception image similarity from 180.86 and 0.28 to 110.98 and 0.22, achieving higher quality image generation. The proposed segmentation method demonstrates high accuracy, compared with the FCN method, the intersection over union ratio and the Dice coefficient are improved from 0.8360 and 0.8998 to 0.9043 and 0.9473, respectively. Hausdorff distance and mean surface distance decreased from 5.5573 mm and 2.3269 mm to 4.7204 mm and 0.9397 mm, respectively, achieving clinically acceptable segmentation accuracy. Our method might reduce physicians’ manual workload and accelerate the diagnosis and treatment process while decreasing inter-observer variability in identifying anatomical structures.
{"title":"Reliable Delineation of Clinical Target Volumes for Cervical Cancer Radiotherapy on CT/MR Dual-Modality Images","authors":"Ying Sun, Yuening Wang, Kexin Gan, Yuxin Wang, Ying Chen, Yun Ge, Jie Yuan, Hanzi Xu","doi":"10.1007/s10278-023-00951-5","DOIUrl":"https://doi.org/10.1007/s10278-023-00951-5","url":null,"abstract":"<p>Accurate delineation of the clinical target volume (CTV) is a crucial prerequisite for safe and effective radiotherapy characterized. This study addresses the integration of magnetic resonance (MR) images to aid in target delineation on computed tomography (CT) images. However, obtaining MR images directly can be challenging. Therefore, we employ AI-based image generation techniques to “intelligentially generate” MR images from CT images to improve CTV delineation based on CT images. To generate high-quality MR images, we propose an attention-guided single-loop image generation model. The model can yield higher-quality images by introducing an attention mechanism in feature extraction and enhancing the loss function. Based on the generated MR images, we propose a CTV segmentation model fusing multi-scale features through image fusion and a hollow space pyramid module to enhance segmentation accuracy. The image generation model used in this study improves the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) from 14.87 and 0.58 to 16.72 and 0.67, respectively, and improves the feature distribution distance and learning-perception image similarity from 180.86 and 0.28 to 110.98 and 0.22, achieving higher quality image generation. The proposed segmentation method demonstrates high accuracy, compared with the FCN method, the intersection over union ratio and the Dice coefficient are improved from 0.8360 and 0.8998 to 0.9043 and 0.9473, respectively. Hausdorff distance and mean surface distance decreased from 5.5573 mm and 2.3269 mm to 4.7204 mm and 0.9397 mm, respectively, achieving clinically acceptable segmentation accuracy. Our method might reduce physicians’ manual workload and accelerate the diagnosis and treatment process while decreasing inter-observer variability in identifying anatomical structures.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"1 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00939-1
Yun Wang, Deng Lyu, Lei Hu, Junhong Wu, Shaofeng Duan, Taohu Zhou, Wenting Tu, Yi Xiao, Li Fan, Shiyuan Liu
The study aims to investigate the value of intratumoral and peritumoral radiomics and clinical-radiological features for predicting spread through air spaces (STAS) in patients with clinical stage IA non-small cell lung cancer (NSCLC). A total of 336 NSCLC patients from our hospital were randomly divided into the training cohort (n = 236) and the internal validation cohort (n = 100) at a ratio of 7:3, and 69 patients from the other two external hospitals were collected as the external validation cohort. Univariate and multivariate analyses were used to select clinical-radiological features and construct a clinical model. The GTV, PTV5, PTV10, PTV15, PTV20, GPTV5, GPTV10, GPTV15, and GPTV20 models were constructed based on intratumoral and peritumoral (5 mm, 10 mm, 15 mm, 20 mm) radiomics features. Additionally, the radscore of the optimal radiomics model and clinical-radiological predictors were used to construct a combined model and plot a nomogram. Lastly, the ROC curve and AUC value were used to evaluate the diagnostic performance of the model. Tumor density type (OR = 6.738) and distal ribbon sign (OR = 5.141) were independent risk factors for the occurrence of STAS. The GPTV10 model outperformed the other radiomics models, and its AUC values were 0.887, 0.876, and 0.868 in the three cohorts. The AUC values of the combined model constructed based on GPTV10 radscore and clinical-radiological predictors were 0.901, 0.875, and 0.878. DeLong test results revealed that the combined model was superior to the clinical model in the three cohorts. The nomogram based on GPTV10 radscore and clinical-radiological features exhibited high predictive efficiency for STAS status in NSCLC.
{"title":"CT-Based Intratumoral and Peritumoral Radiomics Nomograms for the Preoperative Prediction of Spread Through Air Spaces in Clinical Stage IA Non-small Cell Lung Cancer","authors":"Yun Wang, Deng Lyu, Lei Hu, Junhong Wu, Shaofeng Duan, Taohu Zhou, Wenting Tu, Yi Xiao, Li Fan, Shiyuan Liu","doi":"10.1007/s10278-023-00939-1","DOIUrl":"https://doi.org/10.1007/s10278-023-00939-1","url":null,"abstract":"<p>The study aims to investigate the value of intratumoral and peritumoral radiomics and clinical-radiological features for predicting spread through air spaces (STAS) in patients with clinical stage IA non-small cell lung cancer (NSCLC). A total of 336 NSCLC patients from our hospital were randomly divided into the training cohort (<i>n</i> = 236) and the internal validation cohort (<i>n</i> = 100) at a ratio of 7:3, and 69 patients from the other two external hospitals were collected as the external validation cohort. Univariate and multivariate analyses were used to select clinical-radiological features and construct a clinical model. The GTV, PTV5, PTV10, PTV15, PTV20, GPTV5, GPTV10, GPTV15, and GPTV20 models were constructed based on intratumoral and peritumoral (5 mm, 10 mm, 15 mm, 20 mm) radiomics features. Additionally, the radscore of the optimal radiomics model and clinical-radiological predictors were used to construct a combined model and plot a nomogram. Lastly, the ROC curve and AUC value were used to evaluate the diagnostic performance of the model. Tumor density type (OR = 6.738) and distal ribbon sign (OR = 5.141) were independent risk factors for the occurrence of STAS. The GPTV10 model outperformed the other radiomics models, and its AUC values were 0.887, 0.876, and 0.868 in the three cohorts. The AUC values of the combined model constructed based on GPTV10 radscore and clinical-radiological predictors were 0.901, 0.875, and 0.878. DeLong test results revealed that the combined model was superior to the clinical model in the three cohorts. The nomogram based on GPTV10 radscore and clinical-radiological features exhibited high predictive efficiency for STAS status in NSCLC.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"24 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Prompt and correct detection of pulmonary tuberculosis (PTB) is critical in preventing its spread. We aimed to develop a deep learning–based algorithm for detecting PTB on chest X-ray (CXRs) in the emergency department. This retrospective study included 3498 CXRs acquired from the National Taiwan University Hospital (NTUH). The images were chronologically split into a training dataset, NTUH-1519 (images acquired during the years 2015 to 2019; n = 2144), and a testing dataset, NTUH-20 (images acquired during the year 2020; n = 1354). Public databases, including the NIH ChestX-ray14 dataset (model training; 112,120 images), Montgomery County (model testing; 138 images), and Shenzhen (model testing; 662 images), were also used in model development. EfficientNetV2 was the basic architecture of the algorithm. Images from ChestX-ray14 were employed for pseudo-labelling to perform semi-supervised learning. The algorithm demonstrated excellent performance in detecting PTB (area under the receiver operating characteristic curve [AUC] 0.878, 95% confidence interval [CI] 0.854–0.900) in NTUH-20. The algorithm showed significantly better performance in posterior-anterior (PA) CXR (AUC 0.940, 95% CI 0.912–0.965, p-value < 0.001) compared with anterior–posterior (AUC 0.782, 95% CI 0.644–0.897) or portable anterior–posterior (AUC 0.869, 95% CI 0.814–0.918) CXR. The algorithm accurately detected cases of bacteriologically confirmed PTB (AUC 0.854, 95% CI 0.823–0.883). Finally, the algorithm tested favourably in Montgomery County (AUC 0.838, 95% CI 0.765–0.904) and Shenzhen (AUC 0.806, 95% CI 0.771–0.839). A deep learning–based algorithm could detect PTB on CXR with excellent performance, which may help shorten the interval between detection and airborne isolation for patients with PTB.
及时、正确地检测肺结核(PTB)对于防止其传播至关重要。我们旨在开发一种基于深度学习的算法,用于检测急诊科胸部 X 光片(CXR)上的肺结核。这项回顾性研究包括从台湾大学医院(台大医院)获取的 3498 张 CXR。这些图像按时间顺序分为训练数据集 NTUH-1519(2015 年至 2019 年采集的图像;n = 2144)和测试数据集 NTUH-20(2020 年采集的图像;n = 1354)。在模型开发过程中还使用了公共数据库,包括美国国立卫生研究院ChestX-ray14数据集(模型训练;112,120张图像)、蒙哥马利县(模型测试;138张图像)和深圳(模型测试;662张图像)。EfficientNetV2 是该算法的基本架构。来自 ChestX-ray14 的图像被用于伪标签,以执行半监督学习。在 NTUH-20 中,该算法在检测 PTB 方面表现出色(接收者操作特征曲线下面积 [AUC] 0.878,95% 置信区间 [CI] 0.854-0.900)。与前后位(AUC 0.782,95% CI 0.644-0.897)或便携式前后位(AUC 0.869,95% CI 0.814-0.918)CXR 相比,该算法在后前位(PA)CXR 中的表现明显更好(AUC 0.940,95% CI 0.912-0.965,p 值为 0.001)。该算法能准确检测出细菌学确诊的 PTB 病例(AUC 0.854,95% CI 0.823-0.883)。最后,该算法在蒙哥马利县(AUC 0.838,95% CI 0.765-0.904)和深圳(AUC 0.806,95% CI 0.771-0.839)的测试结果良好。基于深度学习的算法可以在CXR上检测出PTB,并且表现出色,这可能有助于缩短PTB患者从检测到空气隔离的时间间隔。
{"title":"Deep Learning–based Diagnosis of Pulmonary Tuberculosis on Chest X-ray in the Emergency Department: A Retrospective Study","authors":"Chih-Hung Wang, Weishan Chang, Meng-Rui Lee, Joyce Tay, Cheng-Yi Wu, Meng-Che Wu, Holger R. Roth, Dong Yang, Can Zhao, Weichung Wang, Chien-Hua Huang","doi":"10.1007/s10278-023-00952-4","DOIUrl":"https://doi.org/10.1007/s10278-023-00952-4","url":null,"abstract":"<p>Prompt and correct detection of pulmonary tuberculosis (PTB) is critical in preventing its spread. We aimed to develop a deep learning–based algorithm for detecting PTB on chest X-ray (CXRs) in the emergency department. This retrospective study included 3498 CXRs acquired from the National Taiwan University Hospital (NTUH). The images were chronologically split into a training dataset, NTUH-1519 (images acquired during the years 2015 to 2019; <i>n</i> = 2144), and a testing dataset, NTUH-20 (images acquired during the year 2020; <i>n</i> = 1354). Public databases, including the NIH ChestX-ray14 dataset (model training; 112,120 images), Montgomery County (model testing; 138 images), and Shenzhen (model testing; 662 images), were also used in model development. EfficientNetV2 was the basic architecture of the algorithm. Images from ChestX-ray14 were employed for pseudo-labelling to perform semi-supervised learning. The algorithm demonstrated excellent performance in detecting PTB (area under the receiver operating characteristic curve [AUC] 0.878, 95% confidence interval [CI] 0.854–0.900) in NTUH-20. The algorithm showed significantly better performance in posterior-anterior (PA) CXR (AUC 0.940, 95% CI 0.912–0.965, <i>p-value</i> < 0.001) compared with anterior–posterior (AUC 0.782, 95% CI 0.644–0.897) or portable anterior–posterior (AUC 0.869, 95% CI 0.814–0.918) CXR. The algorithm accurately detected cases of bacteriologically confirmed PTB (AUC 0.854, 95% CI 0.823–0.883). Finally, the algorithm tested favourably in Montgomery County (AUC 0.838, 95% CI 0.765–0.904) and Shenzhen (AUC 0.806, 95% CI 0.771–0.839). A deep learning–based algorithm could detect PTB on CXR with excellent performance, which may help shorten the interval between detection and airborne isolation for patients with PTB.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"39 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139423453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00916-8
Sheikh Burhan ul haque, Aasim Zafar
In the realm of medical diagnostics, the utilization of deep learning techniques, notably in the context of radiology images, has emerged as a transformative force. The significance of artificial intelligence (AI), specifically machine learning (ML) and deep learning (DL), lies in their capacity to rapidly and accurately diagnose diseases from radiology images. This capability has been particularly vital during the COVID-19 pandemic, where rapid and precise diagnosis played a pivotal role in managing the spread of the virus. DL models, trained on vast datasets of radiology images, have showcased remarkable proficiency in distinguishing between normal and COVID-19-affected cases, offering a ray of hope amidst the crisis. However, as with any technological advancement, vulnerabilities emerge. Deep learning-based diagnostic models, although proficient, are not immune to adversarial attacks. These attacks, characterized by carefully crafted perturbations to input data, can potentially disrupt the models’ decision-making processes. In the medical context, such vulnerabilities could have dire consequences, leading to misdiagnoses and compromised patient care. To address this, we propose a two-phase defense framework that combines advanced adversarial learning and adversarial image filtering techniques. We use a modified adversarial learning algorithm to enhance the model’s resilience against adversarial examples during the training phase. During the inference phase, we apply JPEG compression to mitigate perturbations that cause misclassification. We evaluate our approach on three models based on ResNet-50, VGG-16, and Inception-V3. These models perform exceptionally in classifying radiology images (X-ray and CT) of lung regions into normal, pneumonia, and COVID-19 pneumonia categories. We then assess the vulnerability of these models to three targeted adversarial attacks: fast gradient sign method (FGSM), projected gradient descent (PGD), and basic iterative method (BIM). The results show a significant drop in model performance after the attacks. However, our defense framework greatly improves the models’ resistance to adversarial attacks, maintaining high accuracy on adversarial examples. Importantly, our framework ensures the reliability of the models in diagnosing COVID-19 from clean images.
{"title":"Robust Medical Diagnosis: A Novel Two-Phase Deep Learning Framework for Adversarial Proof Disease Detection in Radiology Images","authors":"Sheikh Burhan ul haque, Aasim Zafar","doi":"10.1007/s10278-023-00916-8","DOIUrl":"https://doi.org/10.1007/s10278-023-00916-8","url":null,"abstract":"<p>In the realm of medical diagnostics, the utilization of deep learning techniques, notably in the context of radiology images, has emerged as a transformative force. The significance of artificial intelligence (AI), specifically machine learning (ML) and deep learning (DL), lies in their capacity to rapidly and accurately diagnose diseases from radiology images. This capability has been particularly vital during the COVID-19 pandemic, where rapid and precise diagnosis played a pivotal role in managing the spread of the virus. DL models, trained on vast datasets of radiology images, have showcased remarkable proficiency in distinguishing between normal and COVID-19-affected cases, offering a ray of hope amidst the crisis. However, as with any technological advancement, vulnerabilities emerge. Deep learning-based diagnostic models, although proficient, are not immune to adversarial attacks. These attacks, characterized by carefully crafted perturbations to input data, can potentially disrupt the models’ decision-making processes. In the medical context, such vulnerabilities could have dire consequences, leading to misdiagnoses and compromised patient care. To address this, we propose a two-phase defense framework that combines advanced adversarial learning and adversarial image filtering techniques. We use a modified adversarial learning algorithm to enhance the model’s resilience against adversarial examples during the training phase. During the inference phase, we apply JPEG compression to mitigate perturbations that cause misclassification. We evaluate our approach on three models based on ResNet-50, VGG-16, and Inception-V3. These models perform exceptionally in classifying radiology images (X-ray and CT) of lung regions into normal, pneumonia, and COVID-19 pneumonia categories. We then assess the vulnerability of these models to three targeted adversarial attacks: fast gradient sign method (FGSM), projected gradient descent (PGD), and basic iterative method (BIM). The results show a significant drop in model performance after the attacks. However, our defense framework greatly improves the models’ resistance to adversarial attacks, maintaining high accuracy on adversarial examples. Importantly, our framework ensures the reliability of the models in diagnosing COVID-19 from clean images.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"41 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139420755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
To develop a fully automatic urinary stone detection system (kidney, ureter, and bladder) and to test it in a real clinical environment. The local institutional review board approved this retrospective single-center study that used non-enhanced abdominopelvic CT scans from patients admitted urology (uPatients) and emergency (ePatients). The uPatients were randomly divided into training and validation sets in a ratio of 3:1. We designed a cascade urinary stone map location-feature pyramid networks (USm-FPNs) and innovatively proposed a ureter distance heatmap method to estimate the ureter position on non-enhanced CT to further reduce the false positives. The performances of the system were compared using the free-response receiver operating characteristic curve and the precision-recall curve. This study included 811 uPatients and 356 ePatients. At stone level, the cascade detector USm-FPNs has the mean of false positives per scan (mFP) 1.88 with the sensitivity 0.977 in validation set, and mFP was further reduced to 1.18 with the sensitivity 0.977 after combining the ureter distance heatmap. At patient level, the sensitivity and precision were as high as 0.995 and 0.990 in validation set, respectively. In a real clinical set of ePatients (27.5% of patients contain stones), the mFP was 1.31 with as high as sensitivity 0.977, and the diagnostic time reduced by > 20% with the system help. A fully automatic detection system for entire urinary stones on non-enhanced CT scans was proposed and reduces obviously the burden on junior radiologists without compromising sensitivity in real emergency data.
{"title":"Automatic Urinary Stone Detection System for Abdominal Non-Enhanced CT Images Reduces the Burden on Radiologists","authors":"Zhaoyu Xing, Zuhui Zhu, Zhenxing Jiang, Jingshi Zhao, Qin Chen, Wei Xing, Liang Pan, Yan Zeng, Aie Liu, Jiule Ding","doi":"10.1007/s10278-023-00946-2","DOIUrl":"https://doi.org/10.1007/s10278-023-00946-2","url":null,"abstract":"<p>To develop a fully automatic urinary stone detection system (kidney, ureter, and bladder) and to test it in a real clinical environment. The local institutional review board approved this retrospective single-center study that used non-enhanced abdominopelvic CT scans from patients admitted urology (uPatients) and emergency (ePatients). The uPatients were randomly divided into training and validation sets in a ratio of 3:1. We designed a cascade urinary stone map location-feature pyramid networks (USm-FPNs) and innovatively proposed a ureter distance heatmap method to estimate the ureter position on non-enhanced CT to further reduce the false positives. The performances of the system were compared using the free-response receiver operating characteristic curve and the precision-recall curve. This study included 811 uPatients and 356 ePatients. At stone level, the cascade detector USm-FPNs has the mean of false positives per scan (mFP) 1.88 with the sensitivity 0.977 in validation set, and mFP was further reduced to 1.18 with the sensitivity 0.977 after combining the ureter distance heatmap. At patient level, the sensitivity and precision were as high as 0.995 and 0.990 in validation set, respectively. In a real clinical set of ePatients (27.5% of patients contain stones), the mFP was 1.31 with as high as sensitivity 0.977, and the diagnostic time reduced by > 20% with the system help. A fully automatic detection system for entire urinary stones on non-enhanced CT scans was proposed and reduces obviously the burden on junior radiologists without compromising sensitivity in real emergency data.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"69 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00943-5
Abstract
Deep neural networks have demonstrated promising performance in screening mammography with recent studies reporting performance at or above the level of trained radiologists on internal datasets. However, it remains unclear whether the performance of these trained models is robust and replicates across external datasets. In this study, we evaluate four state-of-the-art publicly available models using four publicly available mammography datasets (CBIS-DDSM, INbreast, CMMD, OMI-DB). Where test data was available, published results were replicated. The best-performing model, which achieved an area under the ROC curve (AUC) of 0.88 on internal data from NYU, achieved here an AUC of 0.9 on the external CMMD dataset (N = 826 exams). On the larger OMI-DB dataset (N = 11,440 exams), it achieved an AUC of 0.84 but did not match the performance of individual radiologists (at a specificity of 0.92, the sensitivity was 0.97 for the radiologist and 0.53 for the network for a 1-year follow-up). The network showed higher performance for in situ cancers, as opposed to invasive cancers. Among invasive cancers, it was relatively weaker at identifying asymmetries and was relatively stronger at identifying masses. The three other trained models that we evaluated all performed poorly on external datasets. Independent validation of trained models is an essential step to ensure safe and reliable use. Future progress in AI for mammography may depend on a concerted effort to make larger datasets publicly available that span multiple clinical sites.
{"title":"Robustness of Deep Networks for Mammography: Replication Across Public Datasets","authors":"","doi":"10.1007/s10278-023-00943-5","DOIUrl":"https://doi.org/10.1007/s10278-023-00943-5","url":null,"abstract":"<h3>Abstract</h3> <p>Deep neural networks have demonstrated promising performance in screening mammography with recent studies reporting performance at or above the level of trained radiologists on internal datasets. However, it remains unclear whether the performance of these trained models is robust and replicates across external datasets. In this study, we evaluate four state-of-the-art publicly available models using four publicly available mammography datasets (CBIS-DDSM, INbreast, CMMD, OMI-DB). Where test data was available, published results were replicated. The best-performing model, which achieved an area under the ROC curve (AUC) of 0.88 on internal data from NYU, achieved here an AUC of 0.9 on the external CMMD dataset (<em>N</em> = 826 exams). On the larger OMI-DB dataset (<em>N</em> = 11,440 exams), it achieved an AUC of 0.84 but did not match the performance of individual radiologists (at a specificity of 0.92, the sensitivity was 0.97 for the radiologist and 0.53 for the network for a 1-year follow-up). The network showed higher performance for in situ cancers, as opposed to invasive cancers. Among invasive cancers, it was relatively weaker at identifying asymmetries and was relatively stronger at identifying masses. The three other trained models that we evaluated all performed poorly on external datasets. Independent validation of trained models is an essential step to ensure safe and reliable use. Future progress in AI for mammography may depend on a concerted effort to make larger datasets publicly available that span multiple clinical sites.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"2 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00956-0
Xinhua Dai, Anjie Liu, Junhong Liu, Mengjun Zhan, Yuanyuan Liu, Wenchi Ke, Lei Shi, Xinyu Huang, Hu Chen, Zhenhua Deng, Fei Fan
Adult age estimation is one of the most challenging problems in forensic science and physical anthropology. In this study, we aimed to develop and evaluate machine learning (ML) methods based on the modified Gustafson’s criteria for dental age estimation. In this retrospective study, a total of 851 orthopantomograms were collected from patients aged 15 to 40 years old. The secondary dentin formation (SE), periodontal recession (PE), and attrition (AT) of four mandibular premolars were analyzed according to the modified Gustafson’s criteria. Ten ML models were generated and compared for age estimation. The partial least squares regressor outperformed other models in males with a mean absolute error (MAE) of 4.151 years. The support vector regressor (MAE = 3.806 years) showed good performance in females. The accuracy of ML models is better than the single-tooth model provided in the previous studies (MAE = 4.747 years in males and MAE = 4.957 years in females). The Shapley additive explanations method was used to reveal the importance of the 12 features in ML models and found that AT and PE are the most influential in age estimation. The findings suggest that the modified Gustafson method can be effectively employed for adult age estimation in the southwest Chinese population. Furthermore, this study highlights the potential of machine learning models to assist experts in achieving accurate and interpretable age estimation.
成人年龄估计是法医学和体质人类学中最具挑战性的问题之一。在这项研究中,我们旨在开发和评估基于修改后的 Gustafson 牙齿年龄估计标准的机器学习(ML)方法。在这项回顾性研究中,共收集了 851 张 15 至 40 岁患者的正畸照片。根据修改后的 Gustafson 标准分析了四颗下颌前磨牙的继发性牙本质形成(SE)、牙周退缩(PE)和损耗(AT)。在年龄估计方面,共生成并比较了 10 个 ML 模型。偏最小二乘回归模型在男性中的表现优于其他模型,平均绝对误差(MAE)为 4.151 岁。支持向量回归模型(MAE = 3.806 岁)在女性中表现良好。ML 模型的准确性优于以往研究中提供的单齿模型(男性 MAE = 4.747 年,女性 MAE = 4.957 年)。利用 Shapley 加法解释法揭示了 12 个特征在 ML 模型中的重要性,发现 AT 和 PE 对年龄估计的影响最大。研究结果表明,改进的 Gustafson 方法可以有效地用于中国西南地区人群的成人年龄估计。此外,本研究还强调了机器学习模型在协助专家实现准确、可解释的年龄估计方面的潜力。
{"title":"Machine Learning Supported the Modified Gustafson’s Criteria for Dental Age Estimation in Southwest China","authors":"Xinhua Dai, Anjie Liu, Junhong Liu, Mengjun Zhan, Yuanyuan Liu, Wenchi Ke, Lei Shi, Xinyu Huang, Hu Chen, Zhenhua Deng, Fei Fan","doi":"10.1007/s10278-023-00956-0","DOIUrl":"https://doi.org/10.1007/s10278-023-00956-0","url":null,"abstract":"<p>Adult age estimation is one of the most challenging problems in forensic science and physical anthropology. In this study, we aimed to develop and evaluate machine learning (ML) methods based on the modified Gustafson’s criteria for dental age estimation. In this retrospective study, a total of 851 orthopantomograms were collected from patients aged 15 to 40 years old. The secondary dentin formation (SE), periodontal recession (PE), and attrition (AT) of four mandibular premolars were analyzed according to the modified Gustafson’s criteria. Ten ML models were generated and compared for age estimation. The partial least squares regressor outperformed other models in males with a mean absolute error (MAE) of 4.151 years. The support vector regressor (MAE = 3.806 years) showed good performance in females. The accuracy of ML models is better than the single-tooth model provided in the previous studies (MAE = 4.747 years in males and MAE = 4.957 years in females). The Shapley additive explanations method was used to reveal the importance of the 12 features in ML models and found that AT and PE are the most influential in age estimation. The findings suggest that the modified Gustafson method can be effectively employed for adult age estimation in the southwest Chinese population. Furthermore, this study highlights the potential of machine learning models to assist experts in achieving accurate and interpretable age estimation.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"118 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00904-y
Ling Ma, Gen Li, Xingyu Feng, Qiliang Fan, Lizhi Liu
Lung cancer is the leading cause of cancer death. Since lung cancer appears as nodules in the early stage, detecting the pulmonary nodules in an early phase could enhance the treatment efficiency and improve the survival rate of patients. The development of computer-aided analysis technology has made it possible to automatically detect lung nodules in Computed Tomography (CT) screening. In this paper, we propose a novel detection network, TiCNet. It is attempted to embed a transformer module in the 3D Convolutional Neural Network (CNN) for pulmonary nodule detection on CT images. First, we integrate the transformer and CNN in an end-to-end structure to capture both the short- and long-range dependency to provide rich information on the characteristics of nodules. Second, we design the attention block and multi-scale skip pathways for improving the detection of small nodules. Last, we develop a two-head detector to guarantee high sensitivity and specificity. Experimental results on the LUNA16 dataset and PN9 dataset showed that our proposed TiCNet achieved superior performance compared with existing lung nodule detection methods. Moreover, the effectiveness of each module has been proven. The proposed TiCNet model is an effective tool for pulmonary nodule detection. Validation revealed that this model exhibited excellent performance, suggesting its potential usefulness to support lung cancer screening.
{"title":"TiCNet: Transformer in Convolutional Neural Network for Pulmonary Nodule Detection on CT Images","authors":"Ling Ma, Gen Li, Xingyu Feng, Qiliang Fan, Lizhi Liu","doi":"10.1007/s10278-023-00904-y","DOIUrl":"https://doi.org/10.1007/s10278-023-00904-y","url":null,"abstract":"<p>Lung cancer is the leading cause of cancer death. Since lung cancer appears as nodules in the early stage, detecting the pulmonary nodules in an early phase could enhance the treatment efficiency and improve the survival rate of patients. The development of computer-aided analysis technology has made it possible to automatically detect lung nodules in Computed Tomography (CT) screening. In this paper, we propose a novel detection network, TiCNet. It is attempted to embed a transformer module in the 3D Convolutional Neural Network (CNN) for pulmonary nodule detection on CT images. First, we integrate the transformer and CNN in an end-to-end structure to capture both the short- and long-range dependency to provide rich information on the characteristics of nodules. Second, we design the attention block and multi-scale skip pathways for improving the detection of small nodules. Last, we develop a two-head detector to guarantee high sensitivity and specificity. Experimental results on the LUNA16 dataset and PN9 dataset showed that our proposed TiCNet achieved superior performance compared with existing lung nodule detection methods. Moreover, the effectiveness of each module has been proven. The proposed TiCNet model is an effective tool for pulmonary nodule detection. Validation revealed that this model exhibited excellent performance, suggesting its potential usefulness to support lung cancer screening.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"85 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00937-3
Chongfeng Duan, Dapeng Hao, Jiufa Cui, Gang Wang, Wenjian Xu, Nan Li, Xuejun Liu
The objective of this study was to predict Ki-67 proliferation index of meningioma by using a nomogram based on clinical, radiomics, and deep transfer learning (DTL) features. A total of 318 cases were enrolled in the study. The clinical, radiomics, and DTL features were selected to construct models. The calculation of radiomics and DTL score was completed by using selected features and correlation coefficient. The deep transfer learning radiomics (DTLR) nomogram was constructed by selected clinical features, radiomics score, and DTL score. The area under the receiver operator characteristic curve (AUC) was calculated. The models were compared by Delong test of AUCs and decision curve analysis (DCA). The features of sex, size, and peritumoral edema were selected to construct clinical model. Seven radiomics features and 15 DTL features were selected. The AUCs of clinical, radiomics, DTL model, and DTLR nomogram were 0.746, 0.75, 0.717, and 0.779 respectively. DTLR nomogram had the highest AUC of 0.779 (95% CI 0.6643–0.8943) with an accuracy rate of 0.734, a sensitivity value of 0.719, and a specificity value of 0.75 in test set. There was no significant difference in AUCs among four models in Delong test. The DTLR nomogram had a larger net benefit than other models across all the threshold probability. The DTLR nomogram had a satisfactory performance in Ki-67 prediction and could be a new evaluation method of meningioma which would be useful in the clinical decision-making.
{"title":"An MRI-Based Deep Transfer Learning Radiomics Nomogram to Predict Ki-67 Proliferation Index of Meningioma","authors":"Chongfeng Duan, Dapeng Hao, Jiufa Cui, Gang Wang, Wenjian Xu, Nan Li, Xuejun Liu","doi":"10.1007/s10278-023-00937-3","DOIUrl":"https://doi.org/10.1007/s10278-023-00937-3","url":null,"abstract":"<p>The objective of this study was to predict Ki-67 proliferation index of meningioma by using a nomogram based on clinical, radiomics, and deep transfer learning (DTL) features. A total of 318 cases were enrolled in the study. The clinical, radiomics, and DTL features were selected to construct models. The calculation of radiomics and DTL score was completed by using selected features and correlation coefficient. The deep transfer learning radiomics (DTLR) nomogram was constructed by selected clinical features, radiomics score, and DTL score. The area under the receiver operator characteristic curve (AUC) was calculated. The models were compared by Delong test of AUCs and decision curve analysis (DCA). The features of sex, size, and peritumoral edema were selected to construct clinical model. Seven radiomics features and 15 DTL features were selected. The AUCs of clinical, radiomics, DTL model, and DTLR nomogram were 0.746, 0.75, 0.717, and 0.779 respectively. DTLR nomogram had the highest AUC of 0.779 (95% CI 0.6643–0.8943) with an accuracy rate of 0.734, a sensitivity value of 0.719, and a specificity value of 0.75 in test set. There was no significant difference in AUCs among four models in Delong test. The DTLR nomogram had a larger net benefit than other models across all the threshold probability. The DTLR nomogram had a satisfactory performance in Ki-67 prediction and could be a new evaluation method of meningioma which would be useful in the clinical decision-making.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"40 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00965-z
Florent L. Besson, Sylvain Faure
Kinetic modeling represents the ultimate foundations of PET quantitative imaging, a unique opportunity to better characterize the diseases or prevent the reduction of drugs development. Primarily designed for research, parametric imaging based on PET kinetic modeling may become a reality in future clinical practice, enhanced by the technical abilities of the latest generation of commercially available PET systems. In the era of precision medicine, such paradigm shift should be promoted, regardless of the PET system. In order to anticipate and stimulate this emerging clinical paradigm shift, we developed a constructor-independent software package, called PET KinetiX, allowing a faster and easier computation of parametric images from any 4D PET DICOM series, at the whole field of view level. The PET KinetiX package is currently a plug-in for Osirix DICOM viewer. The package provides a suite of five PET kinetic models: Patlak, Logan, 1-tissue compartment model, 2-tissue compartment model, and first pass blood flow. After uploading the 4D-PET DICOM series into Osirix, the image processing requires very few steps: the choice of the kinetic model and the definition of an input function. After a 2-min process, the PET parametric and error maps of the chosen model are automatically estimated voxel-wise and written in DICOM format. The software benefits from the graphical user interface of Osirix, making it user-friendly. Compared to PMOD-PKIN (version 4.4) on twelve 18F-FDG PET dynamic datasets, PET KinetiX provided an absolute bias of 0.1% (0.05–0.25) and 5.8% (3.3–12.3) for KiPatlak and Ki2TCM, respectively. Several clinical research illustrative cases acquired on different hybrid PET systems (standard or extended axial fields of view, PET/CT, and PET/MRI), with different acquisition schemes (single-bed single-pass or multi-bed multipass), are also provided. PET KinetiX is a very fast and efficient independent research software that helps molecular imaging users easily and quickly produce 3D PET parametric images from any reconstructed 4D-PET data acquired on standard or large PET systems.
动力学建模代表了 PET 定量成像的最终基础,是更好地描述疾病特征或防止药物开发减少的独特机会。基于 PET 动力学建模的参数成像主要是为研究设计的,在最新一代商用 PET 系统技术能力的提升下,它可能会在未来的临床实践中成为现实。在精准医疗时代,无论使用哪种 PET 系统,都应促进这种模式的转变。为了预测和促进这一新兴临床范式的转变,我们开发了一个独立于构造器的软件包,名为 PET KinetiX,可以更快、更方便地计算来自任何 4D PET DICOM 系列的全视场参数图像。PET KinetiX 软件包目前是 Osirix DICOM 查看器的插件。该软件包提供一套五种 PET 动力学模型:Patlak、Logan、1-组织间隙模型、2-组织间隙模型和首过血流。将 4D-PET DICOM 系列上传到 Osirix 后,图像处理只需几个步骤:选择动力学模型和定义输入函数。经过 2 分钟的处理后,所选模型的 PET 参数图和误差图将自动按体素估算,并以 DICOM 格式写入。该软件得益于 Osirix 的图形用户界面,用户使用起来非常方便。在 12 个 18F-FDG PET 动态数据集上,PET KinetiX 与 PMOD-PKIN(4.4 版)相比,KiPatlak 和 Ki2TCM 的绝对偏差分别为 0.1%(0.05-0.25)和 5.8%(3.3-12.3)。此外,还提供了在不同的混合 PET 系统(标准或扩展轴向视场、PET/CT 和 PET/MRI)、不同的采集方案(单床单通或多床多通)上采集的几个临床研究示例。PET KinetiX 是一款非常快速高效的独立研究软件,可帮助分子成像用户从标准或大型 PET 系统获取的任何重建 4D-PET 数据中轻松快速地生成三维 PET 参数图像。
{"title":"PET KinetiX—A Software Solution for PET Parametric Imaging at the Whole Field of View Level","authors":"Florent L. Besson, Sylvain Faure","doi":"10.1007/s10278-023-00965-z","DOIUrl":"https://doi.org/10.1007/s10278-023-00965-z","url":null,"abstract":"<p>Kinetic modeling represents the ultimate foundations of PET quantitative imaging, a unique opportunity to better characterize the diseases or prevent the reduction of drugs development. Primarily designed for research, parametric imaging based on PET kinetic modeling may become a reality in future clinical practice, enhanced by the technical abilities of the latest generation of commercially available PET systems. In the era of precision medicine, such paradigm shift should be promoted, regardless of the PET system. In order to anticipate and stimulate this emerging clinical paradigm shift, we developed a constructor-independent software package, called PET KinetiX, allowing a faster and easier computation of parametric images from any 4D PET DICOM series, at the whole field of view level. The PET KinetiX package is currently a plug-in for Osirix DICOM viewer. The package provides a suite of five PET kinetic models: Patlak, Logan, 1-tissue compartment model, 2-tissue compartment model, and first pass blood flow. After uploading the 4D-PET DICOM series into Osirix, the image processing requires very few steps: the choice of the kinetic model and the definition of an input function. After a 2-min process, the PET parametric and error maps of the chosen model are automatically estimated voxel-wise and written in DICOM format. The software benefits from the graphical user interface of Osirix, making it user-friendly. Compared to PMOD-PKIN (version 4.4) on twelve <sup>18</sup>F-FDG PET dynamic datasets, PET KinetiX provided an absolute bias of 0.1% (0.05–0.25) and 5.8% (3.3–12.3) for Ki<sub>Patlak</sub> and Ki<sub>2TCM</sub>, respectively. Several clinical research illustrative cases acquired on different hybrid PET systems (standard or extended axial fields of view, PET/CT, and PET/MRI), with different acquisition schemes (single-bed single-pass or multi-bed multipass), are also provided. PET KinetiX is a very fast and efficient independent research software that helps molecular imaging users easily and quickly produce 3D PET parametric images from any reconstructed 4D-PET data acquired on standard or large PET systems.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"62 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}