Pub Date : 2024-01-11DOI: 10.1007/s10278-023-00903-z
Jia Chen, Lei Liu, Ziying He, Danke Su, Chanzhen Liu
To explore the value of CT-based radiomics model in the differential diagnosis of benign ovarian tumors (BeOTs), borderline ovarian tumors (BOTs), and early malignant ovarian tumors (eMOTs). The retrospective research was conducted with pathologically confirmed 258 ovarian tumor patients from January 2014 to February 2021. The patients were randomly allocated to a training cohort (n = 198) and a test cohort (n = 60). By providing a three-dimensional (3D) characterization of the volume of interest (VOI) at the maximum level of images, 4238 radiomic features were extracted from the VOI per patient. The Wilcoxon–Mann–Whitney (WMW) test, least absolute shrinkage and selection operator (LASSO), and support vector machine (SVM) were employed to select the radiomic features. Five machine learning (ML) algorithms were applied to construct three-class diagnostic models. Leave-one-out cross-validation (LOOCV) was implemented to evaluate the performance of the radiomics models. The test cohort was used to verify the generalization ability of the radiomics models. The receiver-operating characteristic (ROC) was used to evaluate diagnostic performance of radiomics model. Global and discrimination performance of five models was evaluated by average area under the ROC curve (AUC). The average ROC indicated that random forest (RF) diagnostic model in training cohort demonstrated the best diagnostic performance (micro/macro average AUC, 0.98/0.99), which was then confirmed with by LOOCV (micro/macro average AUC, 0.89/0.88) and external validation (test cohort) (micro/macro average AUC, 0.81/0.79). Our proposed CT-based radiomics diagnostic models may effectively assist in preoperatively differentiating BeOTs, BOTs, and eMOTs.
{"title":"CT-Based Radiomics and Machine Learning for Differentiating Benign, Borderline, and Early-Stage Malignant Ovarian Tumors","authors":"Jia Chen, Lei Liu, Ziying He, Danke Su, Chanzhen Liu","doi":"10.1007/s10278-023-00903-z","DOIUrl":"https://doi.org/10.1007/s10278-023-00903-z","url":null,"abstract":"<p>To explore the value of CT-based radiomics model in the differential diagnosis of benign ovarian tumors (BeOTs), borderline ovarian tumors (BOTs), and early malignant ovarian tumors (eMOTs). The retrospective research was conducted with pathologically confirmed 258 ovarian tumor patients from January 2014 to February 2021. The patients were randomly allocated to a training cohort (<i>n</i> = 198) and a test cohort (<i>n</i> = 60). By providing a three-dimensional (3D) characterization of the volume of interest (VOI) at the maximum level of images, 4238 radiomic features were extracted from the VOI per patient. The Wilcoxon–Mann–Whitney (WMW) test, least absolute shrinkage and selection operator (LASSO), and support vector machine (SVM) were employed to select the radiomic features. Five machine learning (ML) algorithms were applied to construct three-class diagnostic models. Leave-one-out cross-validation (LOOCV) was implemented to evaluate the performance of the radiomics models. The test cohort was used to verify the generalization ability of the radiomics models. The receiver-operating characteristic (ROC) was used to evaluate diagnostic performance of radiomics model. Global and discrimination performance of five models was evaluated by average area under the ROC curve (AUC). The average ROC indicated that random forest (RF) diagnostic model in training cohort demonstrated the best diagnostic performance (micro/macro average AUC, 0.98/0.99), which was then confirmed with by LOOCV (micro/macro average AUC, 0.89/0.88) and external validation (test cohort) (micro/macro average AUC, 0.81/0.79). Our proposed CT-based radiomics diagnostic models may effectively assist in preoperatively differentiating BeOTs, BOTs, and eMOTs.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"5 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139463286","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pretreatment patient-specific quality assurance (prePSQA) is conducted to confirm the accuracy of the radiotherapy dose delivered. However, the process of prePSQA measurement is time consuming and exacerbates the workload for medical physicists. The purpose of this work is to propose a novel deep learning (DL) network to improve the accuracy and efficiency of prePSQA. A modified invertible and variable augmented network was developed to predict the three-dimensional (3D) measurement-guided dose (MDose) distribution of 300 cancer patients who underwent volumetric modulated arc therapy (VMAT) between 2018 and 2021, in which 240 cases were randomly selected for training, and 60 for testing. For simplicity, the present approach was termed as “IVPSQA.” The input data include CT images, radiotherapy dose exported from the treatment planning system, and MDose distribution extracted from the verification system. Adam algorithm was used for first-order gradient-based optimization of stochastic objective functions. The IVPSQA model obtained high-quality 3D prePSQA dose distribution maps in head and neck, chest, and abdomen cases, and outperformed the existing U-Net-based prediction approaches in terms of dose difference maps and horizontal profiles comparison. Moreover, quantitative evaluation metrics including SSIM, MSE, and MAE demonstrated that the proposed approach achieved a good agreement with ground truth and yield promising gains over other advanced methods. This study presented the first work on predicting 3D prePSQA dose distribution by using the IVPSQA model. The proposed method could be taken as a clinical guidance tool and help medical physicists to reduce the measurement work of prePSQA.
进行治疗前患者特定质量保证(prePSQA)是为了确认放疗剂量的准确性。然而,预PSQA测量过程耗时,加重了医学物理学家的工作量。这项工作的目的是提出一种新型深度学习(DL)网络,以提高预PSQA的准确性和效率。研究人员开发了一种改进的可逆和可变增强网络,用于预测 2018 年至 2021 年间接受容积调制弧治疗(VMAT)的 300 例癌症患者的三维(3D)测量引导剂量(MDose)分布,其中随机选取 240 例进行训练,60 例进行测试。为简单起见,本方法被称为 "IVPSQA"。输入数据包括 CT 图像、治疗计划系统导出的放疗剂量和验证系统提取的 MDose 分布。亚当算法用于随机目标函数的一阶梯度优化。IVPSQA 模型在头颈部、胸部和腹部病例中获得了高质量的 3D prePSQA 剂量分布图,在剂量差图和水平剖面比较方面优于现有的基于 U-Net 的预测方法。此外,包括 SSIM、MSE 和 MAE 在内的定量评估指标表明,所提出的方法与地面实况具有很好的一致性,与其他先进方法相比具有很好的收益。这项研究首次提出了利用 IVPSQA 模型预测 3D prePSQA 剂量分布的方法。所提出的方法可作为临床指导工具,帮助医学物理学家减少预PSQA的测量工作。
{"title":"Invertible and Variable Augmented Network for Pretreatment Patient-Specific Quality Assurance Dose Prediction","authors":"Zhongsheng Zou, Changfei Gong, Lingpeng Zeng, Yu Guan, Bin Huang, Xiuwen Yu, Qiegen Liu, Minghui Zhang","doi":"10.1007/s10278-023-00930-w","DOIUrl":"https://doi.org/10.1007/s10278-023-00930-w","url":null,"abstract":"<p>Pretreatment patient-specific quality assurance (prePSQA) is conducted to confirm the accuracy of the radiotherapy dose delivered. However, the process of prePSQA measurement is time consuming and exacerbates the workload for medical physicists. The purpose of this work is to propose a novel deep learning (DL) network to improve the accuracy and efficiency of prePSQA. A modified invertible and variable augmented network was developed to predict the three-dimensional (3D) measurement-guided dose (MDose) distribution of 300 cancer patients who underwent volumetric modulated arc therapy (VMAT) between 2018 and 2021, in which 240 cases were randomly selected for training, and 60 for testing. For simplicity, the present approach was termed as “IVPSQA.” The input data include CT images, radiotherapy dose exported from the treatment planning system, and MDose distribution extracted from the verification system. Adam algorithm was used for first-order gradient-based optimization of stochastic objective functions. The IVPSQA model obtained high-quality 3D prePSQA dose distribution maps in head and neck, chest, and abdomen cases, and outperformed the existing U-Net-based prediction approaches in terms of dose difference maps and horizontal profiles comparison. Moreover, quantitative evaluation metrics including SSIM, MSE, and MAE demonstrated that the proposed approach achieved a good agreement with ground truth and yield promising gains over other advanced methods. This study presented the first work on predicting 3D prePSQA dose distribution by using the IVPSQA model. The proposed method could be taken as a clinical guidance tool and help medical physicists to reduce the measurement work of prePSQA.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"7 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Deep learning (DL) has recently attracted attention for data processing in positron emission tomography (PET). Attenuation correction (AC) without computed tomography (CT) data is one of the interests. Here, we present, to our knowledge, the first attempt to generate an attenuation map of the human head via Sim2Real DL-based tissue composition estimation from model training using only the simulated PET dataset. The DL model accepts a two-dimensional non-attenuation-corrected PET image as input and outputs a four-channel tissue-composition map of soft tissue, bone, cavity, and background. Then, an attenuation map is generated by a linear combination of the tissue composition maps and, finally, used as input for scatter+random estimation and as an initial estimate for attenuation map reconstruction by the maximum likelihood attenuation correction factor (MLACF), i.e., the DL estimate is refined by the MLACF. Preliminary results using clinical brain PET data showed that the proposed DL model tended to estimate anatomical details inaccurately, especially in the neck-side slices. However, it succeeded in estimating overall anatomical structures, and the PET quantitative accuracy with DL-based AC was comparable to that with CT-based AC. Thus, the proposed DL-based approach combined with the MLACF is also a promising CT-less AC approach.
最近,深度学习(DL)在正电子发射断层扫描(PET)数据处理方面引起了关注。不使用计算机断层扫描(CT)数据进行衰减校正(AC)是其中一个关注点。据我们所知,我们首次尝试通过 Sim2Real DL 基于模型训练的组织成分估计,仅使用模拟 PET 数据集生成人体头部的衰减图。DL 模型接受二维非衰减校正 PET 图像作为输入,并输出软组织、骨骼、空腔和背景的四通道组织构成图。然后,通过组织成分图的线性组合生成衰减图,最后作为散射+随机估计的输入,并作为最大似然衰减校正因子(MLACF)重建衰减图的初始估计值,即 DL 估计值由 MLACF 精炼。使用临床脑 PET 数据得出的初步结果显示,所提出的 DL 模型对解剖细节的估计往往不准确,尤其是在颈侧切片中。但是,它成功地估计了整体解剖结构,基于 DL 的 AC 的 PET 定量准确性与基于 CT 的 AC 相当。因此,建议的基于 DL 的方法与 MLACF 相结合也是一种很有前景的无 CT AC 方法。
{"title":"Generating PET Attenuation Maps via Sim2Real Deep Learning–Based Tissue Composition Estimation Combined with MLACF","authors":"Tetsuya Kobayashi, Yui Shigeki, Yoshiyuki Yamakawa, Yumi Tsutsumida, Tetsuro Mizuta, Kohei Hanaoka, Shota Watanabe, Daisuke Morimoto‑Ishikawa, Takahiro Yamada, Hayato Kaida, Kazunari Ishii","doi":"10.1007/s10278-023-00902-0","DOIUrl":"https://doi.org/10.1007/s10278-023-00902-0","url":null,"abstract":"<p>Deep learning (DL) has recently attracted attention for data processing in positron emission tomography (PET). Attenuation correction (AC) without computed tomography (CT) data is one of the interests. Here, we present, to our knowledge, the first attempt to generate an attenuation map of the human head via Sim2Real DL-based tissue composition estimation from model training using only the simulated PET dataset. The DL model accepts a two-dimensional non-attenuation-corrected PET image as input and outputs a four-channel tissue-composition map of soft tissue, bone, cavity, and background. Then, an attenuation map is generated by a linear combination of the tissue composition maps and, finally, used as input for scatter+random estimation and as an initial estimate for attenuation map reconstruction by the maximum likelihood attenuation correction factor (MLACF), i.e., the DL estimate is refined by the MLACF. Preliminary results using clinical brain PET data showed that the proposed DL model tended to estimate anatomical details inaccurately, especially in the neck-side slices. However, it succeeded in estimating overall anatomical structures, and the PET quantitative accuracy with DL-based AC was comparable to that with CT-based AC. Thus, the proposed DL-based approach combined with the MLACF is also a promising CT-less AC approach.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"322 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Deep stromal invasion is an important pathological factor associated with the treatments and prognosis of cervical cancer patients. Accurate determination of deep stromal invasion before radical hysterectomy (RH) is of great value for early clinical treatment decision-making and improving the prognosis of these patients. Machine learning is gradually applied in the construction of clinical models to improve the accuracy of clinical diagnosis or prediction, but whether machine learning can improve the preoperative diagnosis accuracy of deep stromal invasion in patients with cervical cancer was still unclear. This cross-sectional study was to construct three preoperative diagnostic models for deep stromal invasion in patients with early cervical cancer based on clinical, radiomics, and clinical combined radiomics data using the machine learning method. We enrolled 229 patients with early cervical cancer receiving RH combined with pelvic lymph node dissection (PLND). The least absolute shrinkage and selection operator (LASSO) and the fivefold cross-validation were applied to screen out radiomics features. Univariate and multivariate logistic regression analyses were applied to identify clinical predictors. All subjects were divided into the training set (n = 160) and testing set (n = 69) at a ratio of 7:3. Three light gradient boosting machine (LightGBM) models were constructed in the training set and verified in the testing set. The radiomics features were statistically different between deep stromal invasion < 1/3 group and deep stromal invasion ≥ 1/3 group. In the training set, the area under the curve (AUC) of the prediction model based on radiomics features was 0.951 (95% confidence interval (CI) 0.922–0.980), the AUC of the prediction model based on clinical predictors was 0.769 (95% CI 0.703–0.835), and the AUC of the prediction model based on radiomics features and clinical predictors was 0.969 (95% CI 0.947–0.990). The AUC of the prediction model based on radiomics features and clinical predictors was 0.914 (95% CI 0.848–0.980) in the testing set. The prediction model for deep stromal invasion in patients with early cervical cancer based on clinical and radiomics data exhibited good predictive performance with an AUC of 0.969, which might help the clinicians early identify patients with high risk of deep stromal invasion and provide timely interventions.
深层间质侵犯是与宫颈癌患者的治疗和预后相关的重要病理因素。在根治性子宫切除术(RH)前对深部间质浸润进行准确判断,对早期临床治疗决策和改善这些患者的预后具有重要价值。机器学习逐渐被应用于临床模型的构建,以提高临床诊断或预测的准确性,但机器学习能否提高宫颈癌患者术前深部间质侵犯的诊断准确性仍不清楚。本横断面研究旨在基于临床、放射组学和临床联合放射组学数据,利用机器学习方法构建三种早期宫颈癌患者深层间质浸润的术前诊断模型。我们招募了229名接受RH联合盆腔淋巴结清扫术(PLND)的早期宫颈癌患者。应用最小绝对收缩和选择算子(LASSO)和五倍交叉验证筛选出放射组学特征。应用单变量和多变量逻辑回归分析确定临床预测因素。所有受试者按 7:3 的比例分为训练集(n = 160)和测试集(n = 69)。在训练集中构建了三个光梯度增强机(LightGBM)模型,并在测试集中进行了验证。深基质侵犯< 1/3组和深基质侵犯≥1/3组的放射组学特征有统计学差异。在训练集中,基于放射组学特征的预测模型的曲线下面积(AUC)为 0.951(95% 置信区间(CI)0.922-0.980),基于临床预测因子的预测模型的曲线下面积(AUC)为 0.769(95% CI 0.703-0.835),基于放射组学特征和临床预测因子的预测模型的曲线下面积(AUC)为 0.969(95% CI 0.947-0.990)。在测试集中,基于放射组学特征和临床预测因子的预测模型的AUC为0.914(95% CI 0.848-0.980)。基于临床和放射组学数据的早期宫颈癌患者深部间质侵犯预测模型显示出良好的预测性能,AUC为0.969,这可能有助于临床医生早期识别深部间质侵犯高风险患者并提供及时干预。
{"title":"Machine Learning-Based Multiparametric Magnetic Resonance Imaging Radiomics Model for Preoperative Predicting the Deep Stromal Invasion in Patients with Early Cervical Cancer","authors":"Haowen Yan, Gaoting Huang, Zhihe Yang, Yirong Chen, Zhiming Xiang","doi":"10.1007/s10278-023-00906-w","DOIUrl":"https://doi.org/10.1007/s10278-023-00906-w","url":null,"abstract":"<p>Deep stromal invasion is an important pathological factor associated with the treatments and prognosis of cervical cancer patients. Accurate determination of deep stromal invasion before radical hysterectomy (RH) is of great value for early clinical treatment decision-making and improving the prognosis of these patients. Machine learning is gradually applied in the construction of clinical models to improve the accuracy of clinical diagnosis or prediction, but whether machine learning can improve the preoperative diagnosis accuracy of deep stromal invasion in patients with cervical cancer was still unclear. This cross-sectional study was to construct three preoperative diagnostic models for deep stromal invasion in patients with early cervical cancer based on clinical, radiomics, and clinical combined radiomics data using the machine learning method. We enrolled 229 patients with early cervical cancer receiving RH combined with pelvic lymph node dissection (PLND). The least absolute shrinkage and selection operator (LASSO) and the fivefold cross-validation were applied to screen out radiomics features. Univariate and multivariate logistic regression analyses were applied to identify clinical predictors. All subjects were divided into the training set (<i>n</i> = 160) and testing set (<i>n</i> = 69) at a ratio of 7:3. Three light gradient boosting machine (LightGBM) models were constructed in the training set and verified in the testing set. The radiomics features were statistically different between deep stromal invasion < 1/3 group and deep stromal invasion ≥ 1/3 group. In the training set, the area under the curve (AUC) of the prediction model based on radiomics features was 0.951 (95% confidence interval (CI) 0.922–0.980), the AUC of the prediction model based on clinical predictors was 0.769 (95% CI 0.703–0.835), and the AUC of the prediction model based on radiomics features and clinical predictors was 0.969 (95% CI 0.947–0.990). The AUC of the prediction model based on radiomics features and clinical predictors was 0.914 (95% CI 0.848–0.980) in the testing set. The prediction model for deep stromal invasion in patients with early cervical cancer based on clinical and radiomics data exhibited good predictive performance with an AUC of 0.969, which might help the clinicians early identify patients with high risk of deep stromal invasion and provide timely interventions.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"7 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421484","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00915-9
Liangliang Zhang, Ming Fan, Lihua Li
Pharmacokinetic (PK) parameters, revealing changes in the tumor microenvironment, are related to the pathological information of breast cancer. Tracer kinetic models (e.g., Tofts-Kety model) with a nonlinear least square solver are commonly used to estimate PK parameters. However, the method is sensitive to noise in images. To relieve the effects of noise, a deconvolution (DEC) method, which was validated on synthetic concentration–time series, was proposed to accurately calculate PK parameters from breast dynamic contrast-enhanced magnetic resonance imaging. A time-to-peak-based tumor partitioning method was used to divide the whole tumor into three tumor subregions with different kinetic patterns. Radiomic features were calculated from the tumor subregion and whole tumor-based PK parameter maps. The optimal features determined by the fivefold cross-validation method were used to build random forest classifiers to predict molecular subtypes, Ki-67, and tumor grade. The diagnostic performance evaluated by the area under the receiver operating characteristic curve (AUC) was compared between the subregion and whole tumor-based PK parameters. The results showed that the DEC method obtained more accurate PK parameters than the Tofts method. Moreover, the results showed that the subregion-based Ktrans (best AUCs = 0.8319, 0.7032, 0.7132, 0.7490, 0.8074, and 0.6950) achieved a better diagnostic performance than the whole tumor-based Ktrans (AUCs = 0.8222, 0.6970, 0.6511, 0.7109, 0.7620, and 0.5894) for molecular subtypes, Ki-67, and tumor grade. These findings indicate that DEC-based Ktrans in the subregion has the potential to accurately predict molecular subtypes, Ki-67, and tumor grade.
{"title":"Deconvolution-Based Pharmacokinetic Analysis to Improve the Prediction of Pathological Information of Breast Cancer","authors":"Liangliang Zhang, Ming Fan, Lihua Li","doi":"10.1007/s10278-023-00915-9","DOIUrl":"https://doi.org/10.1007/s10278-023-00915-9","url":null,"abstract":"<p>Pharmacokinetic (PK) parameters, revealing changes in the tumor microenvironment, are related to the pathological information of breast cancer. Tracer kinetic models (e.g., Tofts-Kety model) with a nonlinear least square solver are commonly used to estimate PK parameters. However, the method is sensitive to noise in images. To relieve the effects of noise, a deconvolution (DEC) method, which was validated on synthetic concentration–time series, was proposed to accurately calculate PK parameters from breast dynamic contrast-enhanced magnetic resonance imaging. A time-to-peak-based tumor partitioning method was used to divide the whole tumor into three tumor subregions with different kinetic patterns. Radiomic features were calculated from the tumor subregion and whole tumor-based PK parameter maps. The optimal features determined by the fivefold cross-validation method were used to build random forest classifiers to predict molecular subtypes, Ki-67, and tumor grade. The diagnostic performance evaluated by the area under the receiver operating characteristic curve (AUC) was compared between the subregion and whole tumor-based PK parameters. The results showed that the DEC method obtained more accurate PK parameters than the Tofts method. Moreover, the results showed that the subregion-based <i>K</i><sup>trans</sup> (best AUCs = 0.8319, 0.7032, 0.7132, 0.7490, 0.8074, and 0.6950) achieved a better diagnostic performance than the whole tumor-based <i>K</i><sup>trans</sup> (AUCs = 0.8222, 0.6970, 0.6511, 0.7109, 0.7620, and 0.5894) for molecular subtypes, Ki-67, and tumor grade. These findings indicate that DEC-based <i>K</i><sup>trans</sup> in the subregion has the potential to accurately predict molecular subtypes, Ki-67, and tumor grade.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"269 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00948-0
Jinwon Kim, Hyebin Lee, Sung Suk Oh, Jinhee Jang, Hyunyeol Lee
Knowledge of input blood to the brain, which is represented as total cerebral blood flow (tCBF), is important in evaluating brain health. Phase-contrast (PC) magnetic resonance imaging (MRI) enables blood velocity mapping, allowing for noninvasive measurements of tCBF. In the procedure, manual selection of brain-feeding arteries is an essential step, but is time-consuming and often subjective. Thus, the purpose of this work was to develop and validate a deep learning (DL)-based technique for automated tCBF quantifications. To enhance the DL segmentation performance on arterial blood vessels, in the preprocessing step magnitude and phase images of PC MRI were multiplied several times. Thereafter, a U-Net was trained on 218 images for three-class segmentation. Network performance was evaluated in terms of the Dice coefficient and the intersection-over-union (IoU) on 40 test images, and additionally, on externally acquired 20 datasets. Finally, tCBF was calculated from the DL-predicted vessel segmentation maps, and its accuracy was statistically assessed with the correlation of determination (R2), the intraclass correlation coefficient (ICC), paired t-tests, and Bland-Altman analysis, in comparison to manually derived values. Overall, the DL segmentation network provided accurate labeling of arterial blood vessels for both internal (Dice=0.92, IoU=0.86) and external (Dice=0.90, IoU=0.82) tests. Furthermore, statistical analyses for tCBF estimates revealed good agreement between automated versus manual quantifications in both internal (R2=0.85, ICC=0.91, p=0.52) and external (R2=0.88, ICC=0.93, p=0.88) test groups. The results suggest feasibility of a simple and automated protocol for quantifying tCBF from neck PC MRI and deep learning.
{"title":"Automated Quantification of Total Cerebral Blood Flow from Phase-Contrast MRI and Deep Learning","authors":"Jinwon Kim, Hyebin Lee, Sung Suk Oh, Jinhee Jang, Hyunyeol Lee","doi":"10.1007/s10278-023-00948-0","DOIUrl":"https://doi.org/10.1007/s10278-023-00948-0","url":null,"abstract":"<p>Knowledge of input blood to the brain, which is represented as total cerebral blood flow (tCBF), is important in evaluating brain health. Phase-contrast (PC) magnetic resonance imaging (MRI) enables blood velocity mapping, allowing for noninvasive measurements of tCBF. In the procedure, manual selection of brain-feeding arteries is an essential step, but is time-consuming and often subjective. Thus, the purpose of this work was to develop and validate a deep learning (DL)-based technique for automated tCBF quantifications. To enhance the DL segmentation performance on arterial blood vessels, in the preprocessing step magnitude and phase images of PC MRI were multiplied several times. Thereafter, a U-Net was trained on 218 images for three-class segmentation. Network performance was evaluated in terms of the Dice coefficient and the intersection-over-union (IoU) on 40 test images, and additionally, on externally acquired 20 datasets. Finally, tCBF was calculated from the DL-predicted vessel segmentation maps, and its accuracy was statistically assessed with the correlation of determination (R<sup>2</sup>), the intraclass correlation coefficient (ICC), paired t-tests, and Bland-Altman analysis, in comparison to manually derived values. Overall, the DL segmentation network provided accurate labeling of arterial blood vessels for both internal (Dice=0.92, IoU=0.86) and external (Dice=0.90, IoU=0.82) tests. Furthermore, statistical analyses for tCBF estimates revealed good agreement between automated versus manual quantifications in both internal (R<sup>2</sup>=0.85, ICC=0.91, p=0.52) and external (R<sup>2</sup>=0.88, ICC=0.93, p=0.88) test groups. The results suggest feasibility of a simple and automated protocol for quantifying tCBF from neck PC MRI and deep learning.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"62 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421625","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00947-1
Abstract
In the field of medicine, rapidly and accurately segmenting organs in medical images is a crucial application of computer technology. This paper introduces a feature map module, Strength Attention Area Signed Distance Map (SAA-SDM), based on the principal component analysis (PCA) principle. The module is designed to accelerate neural networks’ convergence speed in rapidly achieving high precision. SAA-SDM provides the neural network with confidence information regarding the target and background, similar to the signed distance map (SDM), thereby enhancing the network’s understanding of semantic information related to the target. Furthermore, this paper presents a training scheme tailored for the module, aiming to achieve finer segmentation and improved generalization performance. Validation of our approach is carried out using TRUS and chest X-ray datasets. Experimental results demonstrate that our method significantly enhances neural networks’ convergence speed and precision. For instance, the convergence speed of UNet and UNET + + is improved by more than 30%. Moreover, Segformer achieves an increase of over 6% and 3% in mIoU (mean Intersection over Union) on two test datasets without requiring pre-trained parameters. Our approach reduces the time and resource costs associated with training neural networks for organ segmentation tasks while effectively guiding the network to achieve meaningful learning even without pre-trained parameters.
{"title":"SAA-SDM: Neural Networks Faster Learned to Segment Organ Images","authors":"","doi":"10.1007/s10278-023-00947-1","DOIUrl":"https://doi.org/10.1007/s10278-023-00947-1","url":null,"abstract":"<h3>Abstract</h3> <p>In the field of medicine, rapidly and accurately segmenting organs in medical images is a crucial application of computer technology. This paper introduces a feature map module, Strength Attention Area Signed Distance Map (SAA-SDM), based on the principal component analysis (PCA) principle. The module is designed to accelerate neural networks’ convergence speed in rapidly achieving high precision. SAA-SDM provides the neural network with confidence information regarding the target and background, similar to the signed distance map (SDM), thereby enhancing the network’s understanding of semantic information related to the target. Furthermore, this paper presents a training scheme tailored for the module, aiming to achieve finer segmentation and improved generalization performance. Validation of our approach is carried out using TRUS and chest X-ray datasets. Experimental results demonstrate that our method significantly enhances neural networks’ convergence speed and precision. For instance, the convergence speed of UNet and UNET + + is improved by more than 30%. Moreover, Segformer achieves an increase of over 6% and 3% in mIoU (mean Intersection over Union) on two test datasets without requiring pre-trained parameters. Our approach reduces the time and resource costs associated with training neural networks for organ segmentation tasks while effectively guiding the network to achieve meaningful learning even without pre-trained parameters. </p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"82 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We aimed to develop machine learning (ML)-based algorithms to assist physicians in ultrasound-guided localization of cricoid cartilage (CC) and thyroid cartilage (TC) in cricothyroidotomy. Adult female volunteers were prospectively recruited from two hospitals between September and December, 2020. Ultrasonographic images were collected via a modified longitudinal technique. You Only Look Once (YOLOv5s), Faster Regions with Convolutional Neural Network features (Faster R-CNN), and Single Shot Detector (SSD) were selected as the model architectures. A total of 488 women (mean age: 36.0 years) participated in the study, contributing to a total of 292,053 frames of ultrasonographic images. The derived ML-based algorithms demonstrated excellent discriminative performance for the presence of CC (area under the receiver operating characteristic curve [AUC]: YOLOv5s, 0.989, 95% confidence interval [CI]: 0.982–0.994; Faster R-CNN, 0.986, 95% CI: 0.980–0.991; SSD, 0.968, 95% CI: 0.956–0.977) and TC (AUC: YOLOv5s, 0.989, 95% CI: 0.977–0.997; Faster R-CNN, 0.981, 95% CI: 0.965–0.991; SSD, 0.982, 95% CI: 0.973–0.990). Furthermore, in the frames where the model could correctly indicate the presence of CC or TC, it also accurately localized CC (intersection-over-union: YOLOv5s, 0.753, 95% CI: 0.739–0.765; Faster R-CNN, 0.720, 95% CI: 0.709–0.732; SSD, 0.739, 95% CI: 0.726–0.751) or TC (intersection-over-union: YOLOv5s, 0.739, 95% CI: 0.722–0.755; Faster R-CNN, 0.709, 95% CI: 0.687–0.730; SSD, 0.713, 95% CI: 0.695–0.730). The ML-based algorithms could identify anatomical landmarks for cricothyroidotomy in adult females with favorable discriminative and localization performance. Further studies are warranted to transfer this algorithm to hand-held portable ultrasound devices for clinical use.
{"title":"Application of Machine Learning to Ultrasonography in Identifying Anatomical Landmarks for Cricothyroidotomy Among Female Adults: A Multi-center Prospective Observational Study","authors":"Chih-Hung Wang, Jia-Da Li, Cheng-Yi Wu, Yu-Chen Wu, Joyce Tay, Meng-Che Wu, Ching-Hang Hsu, Yi-Kuan Liu, Chu-Song Chen, Chien-Hua Huang","doi":"10.1007/s10278-023-00929-3","DOIUrl":"https://doi.org/10.1007/s10278-023-00929-3","url":null,"abstract":"<p>We aimed to develop machine learning (ML)-based algorithms to assist physicians in ultrasound-guided localization of cricoid cartilage (CC) and thyroid cartilage (TC) in cricothyroidotomy. Adult female volunteers were prospectively recruited from two hospitals between September and December, 2020. Ultrasonographic images were collected via a modified longitudinal technique. You Only Look Once (YOLOv5s), Faster Regions with Convolutional Neural Network features (Faster R-CNN), and Single Shot Detector (SSD) were selected as the model architectures. A total of 488 women (mean age: 36.0 years) participated in the study, contributing to a total of 292,053 frames of ultrasonographic images. The derived ML-based algorithms demonstrated excellent discriminative performance for the presence of CC (area under the receiver operating characteristic curve [AUC]: YOLOv5s, 0.989, 95% confidence interval [CI]: 0.982–0.994; Faster R-CNN, 0.986, 95% CI: 0.980–0.991; SSD, 0.968, 95% CI: 0.956–0.977) and TC (AUC: YOLOv5s, 0.989, 95% CI: 0.977–0.997; Faster R-CNN, 0.981, 95% CI: 0.965–0.991; SSD, 0.982, 95% CI: 0.973–0.990). Furthermore, in the frames where the model could correctly indicate the presence of CC or TC, it also accurately localized CC (intersection-over-union: YOLOv5s, 0.753, 95% CI: 0.739–0.765; Faster R-CNN, 0.720, 95% CI: 0.709–0.732; SSD, 0.739, 95% CI: 0.726–0.751) or TC (intersection-over-union: YOLOv5s, 0.739, 95% CI: 0.722–0.755; Faster R-CNN, 0.709, 95% CI: 0.687–0.730; SSD, 0.713, 95% CI: 0.695–0.730). The ML-based algorithms could identify anatomical landmarks for cricothyroidotomy in adult females with favorable discriminative and localization performance. Further studies are warranted to transfer this algorithm to hand-held portable ultrasound devices for clinical use.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"8 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00933-7
Venkata Sainath Gupta Thadikemalla, Niels K. Focke, Sudhakar Tummala
This paper presents a fully automated pipeline using a sparse convolutional autoencoder for quality control (QC) of affine registrations in large-scale T1-weighted (T1w) and T2-weighted (T2w) magnetic resonance imaging (MRI) studies. Here, a customized 3D convolutional encoder-decoder (autoencoder) framework is proposed and the network is trained in a fully unsupervised manner. For cross-validating the proposed model, we used 1000 correctly aligned MRI images of the human connectome project young adult (HCP-YA) dataset. We proposed that the quality of the registration is proportional to the reconstruction error of the autoencoder. Further, to make this method applicable to unseen datasets, we have proposed dataset-specific optimal threshold calculation (using the reconstruction error) from ROC analysis that requires a subset of the correctly aligned and artificially generated misalignments specific to that dataset. The calculated optimum threshold is used for testing the quality of remaining affine registrations from the corresponding datasets. The proposed framework was tested on four unseen datasets from autism brain imaging data exchange (ABIDE I, 215 subjects), information eXtraction from images (IXI, 577 subjects), Open Access Series of Imaging Studies (OASIS4, 646 subjects), and “Food and Brain” study (77 subjects). The framework has achieved excellent performance for T1w and T2w affine registrations with an accuracy of 100% for HCP-YA. Further, we evaluated the generality of the model on four unseen datasets and obtained accuracies of 81.81% for ABIDE I (only T1w), 93.45% (T1w) and 81.75% (T2w) for OASIS4, and 92.59% for “Food and Brain” study (only T1w) and in the range 88–97% for IXI (for both T1w and T2w and stratified concerning scanner vendor and magnetic field strengths). Moreover, the real failures from “Food and Brain” and OASIS4 datasets were detected with sensitivities of 100% and 80% for T1w and T2w, respectively. In addition, AUCs of > 0.88 in all scenarios were obtained during threshold calculation on the four test sets.
{"title":"A 3D Sparse Autoencoder for Fully Automated Quality Control of Affine Registrations in Big Data Brain MRI Studies","authors":"Venkata Sainath Gupta Thadikemalla, Niels K. Focke, Sudhakar Tummala","doi":"10.1007/s10278-023-00933-7","DOIUrl":"https://doi.org/10.1007/s10278-023-00933-7","url":null,"abstract":"<p>This paper presents a fully automated pipeline using a sparse convolutional autoencoder for quality control (QC) of affine registrations in large-scale T1-weighted (T1w) and T2-weighted (T2w) magnetic resonance imaging (MRI) studies. Here, a customized 3D convolutional encoder-decoder (autoencoder) framework is proposed and the network is trained in a fully unsupervised manner. For cross-validating the proposed model, we used 1000 correctly aligned MRI images of the human connectome project young adult (HCP-YA) dataset. We proposed that the quality of the registration is proportional to the reconstruction error of the autoencoder. Further, to make this method applicable to unseen datasets, we have proposed dataset-specific optimal threshold calculation (using the reconstruction error) from ROC analysis that requires a subset of the correctly aligned and artificially generated misalignments specific to that dataset. The calculated optimum threshold is used for testing the quality of remaining affine registrations from the corresponding datasets. The proposed framework was tested on four unseen datasets from autism brain imaging data exchange (ABIDE I, 215 subjects), information eXtraction from images (IXI, 577 subjects), Open Access Series of Imaging Studies (OASIS4, 646 subjects), and “Food and Brain” study (77 subjects). The framework has achieved excellent performance for T1w and T2w affine registrations with an accuracy of 100% for HCP-YA. Further, we evaluated the generality of the model on four unseen datasets and obtained accuracies of 81.81% for ABIDE I (only T1w), 93.45% (T1w) and 81.75% (T2w) for OASIS4, and 92.59% for “Food and Brain” study (only T1w) and in the range 88–97% for IXI (for both T1w and T2w and stratified concerning scanner vendor and magnetic field strengths). Moreover, the real failures from “Food and Brain” and OASIS4 datasets were detected with sensitivities of 100% and 80% for T1w and T2w, respectively. In addition, AUCs of > 0.88 in all scenarios were obtained during threshold calculation on the four test sets.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"298 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-01-10DOI: 10.1007/s10278-023-00911-z
Abinaya K., Sivakumar B.
Cervical cancer is a significant health problem worldwide, and early detection and treatment are critical to improving patient outcomes. To address this challenge, a deep learning (DL)-based cervical classification system is proposed using 3D convolutional neural network and Vision Transformer (ViT) module. The proposed model leverages the capability of 3D CNN to extract spatiotemporal features from cervical images and employs the ViT model to capture and learn complex feature representations. The model consists of an input layer that receives cervical images, followed by a 3D convolution block, which extracts features from the images. The feature maps generated are down-sampled using max-pooling block to eliminate redundant information and preserve important features. Four Vision Transformer models are employed to extract efficient feature maps of different levels of abstraction. The output of each Vision Transformer model is an efficient set of feature maps that captures spatiotemporal information at a specific level of abstraction. The feature maps generated by the Vision Transformer models are then supplied into the 3D feature pyramid network (FPN) module for feature concatenation. The 3D squeeze-and-excitation (SE) block is employed to obtain efficient feature maps that recalibrate the feature responses of the network based on the interdependencies between different feature maps, thereby improving the discriminative power of the model. At last, dimension minimization of feature maps is executed using 3D average pooling layer. Its output is then fed into a kernel extreme learning machine (KELM) for classification into one of the five classes. The KELM uses radial basis kernel function (RBF) for mapping features in high-dimensional feature space and classifying the input samples. The superiority of the proposed model is known using simulation results, achieving an accuracy of 98.6%, demonstrating its potential as an effective tool for cervical cancer classification. Also, it can be used as a diagnostic supportive tool to assist medical experts in accurately identifying cervical cancer in patients.
宫颈癌是全球范围内的重大健康问题,早期检测和治疗对改善患者预后至关重要。为了应对这一挑战,我们利用三维卷积神经网络和视觉转换器(ViT)模块,提出了一种基于深度学习(DL)的宫颈癌分类系统。该模型利用三维卷积神经网络的能力从宫颈图像中提取时空特征,并采用 ViT 模型捕捉和学习复杂的特征表征。该模型由接收宫颈图像的输入层和从图像中提取特征的三维卷积块组成。生成的特征图使用最大池化块进行下采样,以消除冗余信息并保留重要特征。四种视觉变换器模型用于提取不同抽象程度的高效特征图。每个视觉转换器模型的输出都是一组高效的特征图,可捕捉特定抽象层次的时空信息。然后,由视觉转换器模型生成的特征图将被输入三维特征金字塔网络(FPN)模块进行特征串联。三维挤压激励(SE)模块用于获取高效的特征图,根据不同特征图之间的相互依存关系重新校准网络的特征响应,从而提高模型的判别能力。最后,利用三维平均池化层对特征图进行维度最小化处理。然后,将其输出输入内核极端学习机(KELM),将其分类为五个类别之一。KELM 使用径向基核函数(RBF)在高维特征空间中映射特征,并对输入样本进行分类。模拟结果表明,该模型的准确率高达 98.6%,显示了其作为宫颈癌分类有效工具的潜力。此外,它还可用作辅助诊断工具,协助医学专家准确识别宫颈癌患者。
{"title":"A Deep Learning-Based Approach for Cervical Cancer Classification Using 3D CNN and Vision Transformer","authors":"Abinaya K., Sivakumar B.","doi":"10.1007/s10278-023-00911-z","DOIUrl":"https://doi.org/10.1007/s10278-023-00911-z","url":null,"abstract":"<p>Cervical cancer is a significant health problem worldwide, and early detection and treatment are critical to improving patient outcomes. To address this challenge, a deep learning (DL)-based cervical classification system is proposed using 3D convolutional neural network and Vision Transformer (ViT) module. The proposed model leverages the capability of 3D CNN to extract spatiotemporal features from cervical images and employs the ViT model to capture and learn complex feature representations. The model consists of an input layer that receives cervical images, followed by a 3D convolution block, which extracts features from the images. The feature maps generated are down-sampled using max-pooling block to eliminate redundant information and preserve important features. Four Vision Transformer models are employed to extract efficient feature maps of different levels of abstraction. The output of each Vision Transformer model is an efficient set of feature maps that captures spatiotemporal information at a specific level of abstraction. The feature maps generated by the Vision Transformer models are then supplied into the 3D feature pyramid network (FPN) module for feature concatenation. The 3D squeeze-and-excitation (SE) block is employed to obtain efficient feature maps that recalibrate the feature responses of the network based on the interdependencies between different feature maps, thereby improving the discriminative power of the model. At last, dimension minimization of feature maps is executed using 3D average pooling layer. Its output is then fed into a kernel extreme learning machine (KELM) for classification into one of the five classes. The KELM uses radial basis kernel function (RBF) for mapping features in high-dimensional feature space and classifying the input samples. The superiority of the proposed model is known using simulation results, achieving an accuracy of 98.6%, demonstrating its potential as an effective tool for cervical cancer classification. Also, it can be used as a diagnostic supportive tool to assist medical experts in accurately identifying cervical cancer in patients.</p>","PeriodicalId":50214,"journal":{"name":"Journal of Digital Imaging","volume":"21 1","pages":""},"PeriodicalIF":4.4,"publicationDate":"2024-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139421833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}