Pub Date : 2024-07-01Epub Date: 2024-07-09DOI: 10.1117/1.JMI.11.4.044501
Daniel D Liang, David D Liang, Marc J Pomeroy, Yongfeng Gao, Licheng R Kuo, Lihong C Li
Purpose: Medical imaging-based machine learning (ML) for computer-aided diagnosis of in vivo lesions consists of two basic components or modules of (i) feature extraction from non-invasively acquired medical images and (ii) feature classification for prediction of malignancy of lesions detected or localized in the medical images. This study investigates their individual performances for diagnosis of low-dose computed tomography (CT) screening-detected lesions of pulmonary nodules and colorectal polyps.
Approach: Three feature extraction methods were investigated. One uses the mathematical descriptor of gray-level co-occurrence image texture measure to extract the Haralick image texture features (HFs). One uses the convolutional neural network (CNN) architecture to extract deep learning (DL) image abstractive features (DFs). The third one uses the interactions between lesion tissues and X-ray energy of CT to extract tissue-energy specific characteristic features (TFs). All the above three categories of extracted features were classified by the random forest (RF) classifier with comparison to the DL-CNN method, which reads the images, extracts the DFs, and classifies the DFs in an end-to-end manner. The ML diagnosis of lesions or prediction of lesion malignancy was measured by the area under the receiver operating characteristic curve (AUC). Three lesion image datasets were used. The lesions' tissue pathological reports were used as the learning labels.
Results: Experiments on the three datasets produced AUC values of 0.724 to 0.878 for the HFs, 0.652 to 0.965 for the DFs, and 0.985 to 0.996 for the TFs, compared to the DL-CNN of 0.694 to 0.964. These experimental outcomes indicate that the RF classifier performed comparably to the DL-CNN classification module and the extraction of tissue-energy specific characteristic features dramatically improved AUC value.
Conclusions: The feature extraction module is more important than the feature classification module. Extraction of tissue-energy specific characteristic features is more important than extraction of image abstractive and characteristic features.
{"title":"Examining feature extraction and classification modules in machine learning for diagnosis of low-dose computed tomographic screening-detected <i>in vivo</i> lesions.","authors":"Daniel D Liang, David D Liang, Marc J Pomeroy, Yongfeng Gao, Licheng R Kuo, Lihong C Li","doi":"10.1117/1.JMI.11.4.044501","DOIUrl":"10.1117/1.JMI.11.4.044501","url":null,"abstract":"<p><strong>Purpose: </strong>Medical imaging-based machine learning (ML) for computer-aided diagnosis of <i>in vivo</i> lesions consists of two basic components or modules of (i) feature extraction from non-invasively acquired medical images and (ii) feature classification for prediction of malignancy of lesions detected or localized in the medical images. This study investigates their individual performances for diagnosis of low-dose computed tomography (CT) screening-detected lesions of pulmonary nodules and colorectal polyps.</p><p><strong>Approach: </strong>Three feature extraction methods were investigated. One uses the mathematical descriptor of gray-level co-occurrence image texture measure to extract the Haralick image texture features (HFs). One uses the convolutional neural network (CNN) architecture to extract deep learning (DL) image abstractive features (DFs). The third one uses the interactions between lesion tissues and X-ray energy of CT to extract tissue-energy specific characteristic features (TFs). All the above three categories of extracted features were classified by the random forest (RF) classifier with comparison to the DL-CNN method, which reads the images, extracts the DFs, and classifies the DFs in an end-to-end manner. The ML diagnosis of lesions or prediction of lesion malignancy was measured by the area under the receiver operating characteristic curve (AUC). Three lesion image datasets were used. The lesions' tissue pathological reports were used as the learning labels.</p><p><strong>Results: </strong>Experiments on the three datasets produced AUC values of 0.724 to 0.878 for the HFs, 0.652 to 0.965 for the DFs, and 0.985 to 0.996 for the TFs, compared to the DL-CNN of 0.694 to 0.964. These experimental outcomes indicate that the RF classifier performed comparably to the DL-CNN classification module and the extraction of tissue-energy specific characteristic features dramatically improved AUC value.</p><p><strong>Conclusions: </strong>The feature extraction module is more important than the feature classification module. Extraction of tissue-energy specific characteristic features is more important than extraction of image abstractive and characteristic features.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"044501"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11234229/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141591735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Purpose: Tetralogy of Fallot (TOF) is a congenital heart disease, and patients undergo surgical repair early in their lives. The evaluation of TOF patients is continuous through their adulthood. The use of cardiac magnetic resonance imaging (CMR) is vital for the evaluation of TOF patients. We aim to correlate advanced MRI sequences [parametric longitudinal relaxation time (T1), extracellular volume (ECV) mapping] with cardiac functionality to provide biomarkers for the evaluation of these patients.
Methods: A complete CMR examination with the same imaging protocol was conducted in a total of 11 TOF patients and a control group of 25 healthy individuals. A Modified Look-Locker Inversion recovery (MOLLI) sequence was included to acquire the global T1 myocardial relaxation times of the left ventricular (LV) pre and post-contrast administration. Appropriate software (Circle cmr42) was used for the CMR analysis and the calculation of native, post-contrast T1, and ECV maps. A regression analysis was conducted for the correlation between global LV T1 values and right ventricular (RV) functional indices.
Results: Statistically significant results were obtained for RV cardiac index [RV_CI= -32.765 + 0.029 × T1 native; ], RV end diastolic volume [RV_EDV/BSA = -1023.872 + 0.902 × T1 native; ], and RV end systolic volume [RV_ESV/BSA = -536.704 + 0.472 × T1 native; ].
Conclusions: We further support the diagnostic importance of T1 mapping as a structural imaging tool in CMR. In addition to the well-known affected RV function in TOF patients, the LV structure is also impaired as there is a strong correlation between LV T1 mapping and RV function, evoking that the heart operates as an entity.
{"title":"Left ventricular structural integrity on tetralogy of Fallot patients: approach using longitudinal relaxation time mapping.","authors":"Giorgos Broumpoulis, Efstratios Karavasilis, Niki Lama, Ioannis Papadopoulos, Panagiotis Zachos, Sotiria Apostolopoulou, Nikolaos Kelekis","doi":"10.1117/1.JMI.11.4.044004","DOIUrl":"10.1117/1.JMI.11.4.044004","url":null,"abstract":"<p><strong>Purpose: </strong>Tetralogy of Fallot (TOF) is a congenital heart disease, and patients undergo surgical repair early in their lives. The evaluation of TOF patients is continuous through their adulthood. The use of cardiac magnetic resonance imaging (CMR) is vital for the evaluation of TOF patients. We aim to correlate advanced MRI sequences [parametric longitudinal relaxation time (T1), extracellular volume (ECV) mapping] with cardiac functionality to provide biomarkers for the evaluation of these patients.</p><p><strong>Methods: </strong>A complete CMR examination with the same imaging protocol was conducted in a total of 11 TOF patients and a control group of 25 healthy individuals. A Modified Look-Locker Inversion recovery (MOLLI) sequence was included to acquire the global T1 myocardial relaxation times of the left ventricular (LV) pre and post-contrast administration. Appropriate software (Circle cmr42) was used for the CMR analysis and the calculation of native, post-contrast T1, and ECV maps. A regression analysis was conducted for the correlation between global LV T1 values and right ventricular (RV) functional indices.</p><p><strong>Results: </strong>Statistically significant results were obtained for RV cardiac index [RV_CI= -32.765 + 0.029 × T1 native; <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.003</mn></mrow> </math> ], RV end diastolic volume [RV_EDV/BSA = -1023.872 + 0.902 × T1 native; <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.001</mn></mrow> </math> ], and RV end systolic volume [RV_ESV/BSA = -536.704 + 0.472 × T1 native; <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.011</mn></mrow> </math> ].</p><p><strong>Conclusions: </strong>We further support the diagnostic importance of T1 mapping as a structural imaging tool in CMR. In addition to the well-known affected RV function in TOF patients, the LV structure is also impaired as there is a strong correlation between LV T1 mapping and RV function, evoking that the heart operates as an entity.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"044004"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11293558/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141890473","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-08-13DOI: 10.1117/1.JMI.11.4.045503
Joshua D Herman, Rachel E Roca, Alexandra G O'Neill, Marcus L Wong, Sajan Goud Lingala, Angel R Pineda
Purpose: Recent research explores using neural networks to reconstruct undersampled magnetic resonance imaging. Because of the complexity of the artifacts in the reconstructed images, there is a need to develop task-based approaches to image quality. We compared conventional global quantitative metrics to evaluate image quality in undersampled images generated by a neural network with human observer performance in a detection task. The purpose is to study which acceleration (2×, 3×, 4×, 5×) would be chosen with the conventional metrics and compare it to the acceleration chosen by human observer performance.
Approach: We used common global metrics for evaluating image quality: the normalized root mean squared error (NRMSE) and structural similarity (SSIM). These metrics are compared with a measure of image quality that incorporates a subtle signal for a specific task to allow for image quality assessment that locally evaluates the effect of undersampling on a signal. We used a U-Net to reconstruct under-sampled images with 2×, 3×, 4×, and 5× one-dimensional undersampling rates. Cross-validation was performed for a 500- and a 4000-image training set with both SSIM and MSE losses. A two-alternative forced choice (2-AFC) observer study was carried out for detecting a subtle signal (small blurred disk) from images with the 4000-image training set.
Results: We found that for both loss functions, the human observer performance on the 2-AFC studies led to a choice of a 2× undersampling, but the SSIM and NRMSE led to a choice of a 3× undersampling.
Conclusions: For this detection task using a subtle small signal at the edge of detectability, SSIM and NRMSE led to an overestimate of the achievable undersampling using a U-Net before a steep loss of image quality between 2×, 3×, 4×, 5× undersampling rates when compared to the performance of human observers in the detection task.
{"title":"Task-based assessment for neural networks: evaluating undersampled MRI reconstructions based on human observer signal detection.","authors":"Joshua D Herman, Rachel E Roca, Alexandra G O'Neill, Marcus L Wong, Sajan Goud Lingala, Angel R Pineda","doi":"10.1117/1.JMI.11.4.045503","DOIUrl":"10.1117/1.JMI.11.4.045503","url":null,"abstract":"<p><strong>Purpose: </strong>Recent research explores using neural networks to reconstruct undersampled magnetic resonance imaging. Because of the complexity of the artifacts in the reconstructed images, there is a need to develop task-based approaches to image quality. We compared conventional global quantitative metrics to evaluate image quality in undersampled images generated by a neural network with human observer performance in a detection task. The purpose is to study which acceleration (2×, 3×, 4×, 5×) would be chosen with the conventional metrics and compare it to the acceleration chosen by human observer performance.</p><p><strong>Approach: </strong>We used common global metrics for evaluating image quality: the normalized root mean squared error (NRMSE) and structural similarity (SSIM). These metrics are compared with a measure of image quality that incorporates a subtle signal for a specific task to allow for image quality assessment that locally evaluates the effect of undersampling on a signal. We used a U-Net to reconstruct under-sampled images with 2×, 3×, 4×, and 5× one-dimensional undersampling rates. Cross-validation was performed for a 500- and a 4000-image training set with both SSIM and MSE losses. A two-alternative forced choice (2-AFC) observer study was carried out for detecting a subtle signal (small blurred disk) from images with the 4000-image training set.</p><p><strong>Results: </strong>We found that for both loss functions, the human observer performance on the 2-AFC studies led to a choice of a 2× undersampling, but the SSIM and NRMSE led to a choice of a 3× undersampling.</p><p><strong>Conclusions: </strong>For this detection task using a subtle small signal at the edge of detectability, SSIM and NRMSE led to an overestimate of the achievable undersampling using a U-Net before a steep loss of image quality between 2×, 3×, 4×, 5× undersampling rates when compared to the performance of human observers in the detection task.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"045503"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11321363/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141983636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-08-07DOI: 10.1117/1.JMI.11.4.044507
Mohammad Mehdi Farhangi, Michael Maynord, Cornelia Fermüller, Yiannis Aloimonos, Berkman Sahiner, Nicholas Petrick
Purpose: Synthetic datasets hold the potential to offer cost-effective alternatives to clinical data, ensuring privacy protections and potentially addressing biases in clinical data. We present a method leveraging such datasets to train a machine learning algorithm applied as part of a computer-aided detection (CADe) system.
Approach: Our proposed approach utilizes clinically acquired computed tomography (CT) scans of a physical anthropomorphic phantom into which manufactured lesions were inserted to train a machine learning algorithm. We treated the training database obtained from the anthropomorphic phantom as a simplified representation of clinical data and increased the variability in this dataset using a set of randomized and parameterized augmentations. Furthermore, to mitigate the inherent differences between phantom and clinical datasets, we investigated adding unlabeled clinical data into the training pipeline.
Results: We apply our proposed method to the false positive reduction stage of a lung nodule CADe system in CT scans, in which regions of interest containing potential lesions are classified as nodule or non-nodule regions. Experimental results demonstrate the effectiveness of the proposed method; the system trained on labeled data from physical phantom scans and unlabeled clinical data achieves a sensitivity of 90% at eight false positives per scan. Furthermore, the experimental results demonstrate the benefit of the physical phantom in which the performance in terms of competitive performance metric increased by 6% when a training set consisting of 50 clinical CT scans was enlarged by the scans obtained from the physical phantom.
Conclusions: The scalability of synthetic datasets can lead to improved CADe performance, particularly in scenarios in which the size of the labeled clinical data is limited or subject to inherent bias. Our proposed approach demonstrates an effective utilization of synthetic datasets for training machine learning algorithms.
{"title":"Exploring synthetic datasets for computer-aided detection: a case study using phantom scan data for enhanced lung nodule false positive reduction.","authors":"Mohammad Mehdi Farhangi, Michael Maynord, Cornelia Fermüller, Yiannis Aloimonos, Berkman Sahiner, Nicholas Petrick","doi":"10.1117/1.JMI.11.4.044507","DOIUrl":"10.1117/1.JMI.11.4.044507","url":null,"abstract":"<p><strong>Purpose: </strong>Synthetic datasets hold the potential to offer cost-effective alternatives to clinical data, ensuring privacy protections and potentially addressing biases in clinical data. We present a method leveraging such datasets to train a machine learning algorithm applied as part of a computer-aided detection (CADe) system.</p><p><strong>Approach: </strong>Our proposed approach utilizes clinically acquired computed tomography (CT) scans of a physical anthropomorphic phantom into which manufactured lesions were inserted to train a machine learning algorithm. We treated the training database obtained from the anthropomorphic phantom as a simplified representation of clinical data and increased the variability in this dataset using a set of randomized and parameterized augmentations. Furthermore, to mitigate the inherent differences between phantom and clinical datasets, we investigated adding unlabeled clinical data into the training pipeline.</p><p><strong>Results: </strong>We apply our proposed method to the false positive reduction stage of a lung nodule CADe system in CT scans, in which regions of interest containing potential lesions are classified as nodule or non-nodule regions. Experimental results demonstrate the effectiveness of the proposed method; the system trained on labeled data from physical phantom scans and unlabeled clinical data achieves a sensitivity of 90% at eight false positives per scan. Furthermore, the experimental results demonstrate the benefit of the physical phantom in which the performance in terms of competitive performance metric increased by 6% when a training set consisting of 50 clinical CT scans was enlarged by the scans obtained from the physical phantom.</p><p><strong>Conclusions: </strong>The scalability of synthetic datasets can lead to improved CADe performance, particularly in scenarios in which the size of the labeled clinical data is limited or subject to inherent bias. Our proposed approach demonstrates an effective utilization of synthetic datasets for training machine learning algorithms.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"044507"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11304989/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141907942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-08-10DOI: 10.1117/1.JMI.11.4.045502
Jelena M Mihailovic, Yoshihisa Kanaji, Daniel Miller, Malcolm R Bell, Kenneth A Fetterly
Purpose: Spatio-temporal variability in clinical fluoroscopy and cine angiography images combined with nonlinear image processing prevents the application of traditional image quality measurements in the cardiac catheterization laboratory. We aimed to develop and validate methods to measure human observer impressions of the image quality.
Approach: Multi-frame images of the thorax of a euthanized pig were acquired to provide an anatomical background. The detector dose was varied from 6 to 200 nGy (increments 2×), and 0.6 and 1.0 mm focal spots were used. Two coronary stents with/without 0.5 mm separation and a synthetic right coronary artery (RCA) with hemispherical defects were embedded into the background images as test objects. The quantitative observer ( ) performance was measured using a two-alternating forced-choice test of whether stents were separated and by a count of visible right coronary artery defects. Qualitative impressions of noise, spatial resolution, and overall image quality were measured using a visual analog scale (VAS). A paired -test and multinomial logistic regression model were used to identify statistically significant factors affecting the observer's impression image quality.
Results: The proportion of correct detection of stent separation and the number of reported right coronary artery defects changed significantly with detector dose increment in the 6 to 100 nGy ( ). Although a trend favored the 0.6 versus 1.0 mm focal spot for these quantitative assessments, this was insignificant. Visual analog scale measurements changed significantly with detector dose increments in the range of 24 to 100 nGy and focal spot size ( ). The application of multinomial logistic regression analysis to observer VAS scores demonstrated sensitivity matching of the paired -test applied to quantitative observer performance measurements.
Conclusions: Both quantitative and qualitative measurements of observer impression of the image quality were sensitive to image quality changes associated with changing the detector dose and focal spot size. These findings encourage future work that uses qualitative image quality measurements to assess clinical fluoroscopy and angiography image quality.
{"title":"Comparison of human observer impression of X-ray fluoroscopy and angiography image quality with technical changes to image quality.","authors":"Jelena M Mihailovic, Yoshihisa Kanaji, Daniel Miller, Malcolm R Bell, Kenneth A Fetterly","doi":"10.1117/1.JMI.11.4.045502","DOIUrl":"10.1117/1.JMI.11.4.045502","url":null,"abstract":"<p><strong>Purpose: </strong>Spatio-temporal variability in clinical fluoroscopy and cine angiography images combined with nonlinear image processing prevents the application of traditional image quality measurements in the cardiac catheterization laboratory. We aimed to develop and validate methods to measure human observer impressions of the image quality.</p><p><strong>Approach: </strong>Multi-frame images of the thorax of a euthanized pig were acquired to provide an anatomical background. The detector dose was varied from 6 to 200 nGy (increments 2×), and 0.6 and 1.0 mm focal spots were used. Two coronary stents with/without 0.5 mm separation and a synthetic right coronary artery (RCA) with hemispherical defects were embedded into the background images as test objects. The quantitative observer ( <math><mrow><mi>n</mi> <mo>=</mo> <mn>17</mn></mrow> </math> ) performance was measured using a two-alternating forced-choice test of whether stents were separated and by a count of visible right coronary artery defects. Qualitative impressions of noise, spatial resolution, and overall image quality were measured using a visual analog scale (VAS). A paired <math><mrow><mi>t</mi></mrow> </math> -test and multinomial logistic regression model were used to identify statistically significant factors affecting the observer's impression image quality.</p><p><strong>Results: </strong>The proportion of correct detection of stent separation and the number of reported right coronary artery defects changed significantly with detector dose increment in the 6 to 100 nGy ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.05</mn></mrow> </math> ). Although a trend favored the 0.6 versus 1.0 mm focal spot for these quantitative assessments, this was insignificant. Visual analog scale measurements changed significantly with detector dose increments in the range of 24 to 100 nGy and focal spot size ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.05</mn></mrow> </math> ). The application of multinomial logistic regression analysis to observer VAS scores demonstrated sensitivity matching of the paired <math><mrow><mi>t</mi></mrow> </math> -test applied to quantitative observer performance measurements.</p><p><strong>Conclusions: </strong>Both quantitative and qualitative measurements of observer impression of the image quality were sensitive to image quality changes associated with changing the detector dose and focal spot size. These findings encourage future work that uses qualitative image quality measurements to assess clinical fluoroscopy and angiography image quality.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"045502"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11316400/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141917779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-08-22DOI: 10.1117/1.JMI.11.4.043503
A Kyle Jones, Moiz Ahmad, Shaan M Raza, Stephen R Chen, Jeffrey H Siewerdsen
Purpose: We aim to compare the imaging performance of a cone-beam CT (CBCT) imaging system with noncircular scan protocols (sine-on-sphere) to a conventional circular orbit.
Approach: A biplane C-arm system (ARTIS Icono; Siemens Healthineers) capable of circular and noncircular CBCT acquisition was used, with the latter orbit (sine-on-sphere, "Sine Spin") executing a sinusoidal motion with tilt amplitude over the half-scan orbit. A test phantom was used for the characterization of image uniformity, noise, noise-power spectrum (NPS), spatial resolution [modulation transfer function (MTF) in axial and oblique directions], and cone-beam artifacts. Findings were interpreted using an anthropomorphic head phantom with respect to pertinent tasks in skull base neurosurgery.
Results: The noncircular scan protocol exhibited several advantages associated with improved 3D sampling-evident in the NPS as filling of the null cone about the spatial frequency axis and reduction of cone-beam artifacts. The region of support at the longitudinal extrema was reduced from 16 to at a radial distance of 6.5 cm. Circular and noncircular orbits exhibited nearly identical image uniformity and quantum noise, demonstrating cupping of and overall noise of . Although both the radially averaged axial MTF ( ) and 45 deg oblique MTF ( ) were lower for the noncircular orbit compared with the circular orbit at the default full reconstruction field of view (FOV), there was no difference in spatial resolution for the medium reconstruction FOV (smaller voxel size). Differences in the perceptual image quality for the anthropomorphic phantom reinforced the objective, quantitative findings, including reduced beam-hardening and cone-beam artifacts about structures of interest in the skull base.
Conclusions: Image quality differences between circular and noncircular CBCT orbits were quantitatively evaluated on a clinical system in the context of neurosurgery. The primary performance advantage for the noncircular orbit was the improved sampling and elimination of cone-beam artifacts.
目的:我们旨在比较锥束 CT(CBCT)成像系统的非圆形扫描方案(正弦-球面)与传统圆形轨道的成像性能:方法:使用可进行圆形和非圆形 CBCT 采集的双平面 C 臂系统(ARTIS Icono; Siemens Healthineers),后一种轨道(球面正弦,"正弦旋转")在半扫描轨道上执行倾斜幅度为 ± 10 度的正弦运动。测试模型用于鉴定图像均匀性、噪声、噪声功率谱(NPS)、空间分辨率[轴向和斜向的调制传递函数(MTF)]和锥形光束伪影。使用一个拟人头部模型,结合颅底神经外科的相关任务对研究结果进行解释:非圆形扫描方案在改进三维采样方面具有多项优势--在 NPS 中表现为围绕 f z 空间频率轴填充空锥体和减少锥束伪影。在 6.5 厘米的径向距离上,纵向极值的支撑区域从 16 厘米减少到 12 厘米。圆形和非圆形轨道显示出几乎相同的图像均匀性和量子噪声,显示出- 16.7%的杯突和 ∼ 27 HU 的总体噪声。虽然在默认的全重建视场(FOV)下,非圆形轨道的径向平均轴向 MTF ( f x , y ) 和 45 度斜向 MTF ( f x , y , z ) 比圆形轨道低 20 %,但在中等重建视场(较小的体素尺寸)下,空间分辨率没有差异。拟人化模型的感知图像质量差异加强了客观的定量研究结果,包括减少了颅底相关结构的光束硬化和锥形束伪影:在神经外科的临床系统上对圆形和非圆形 CBCT 眼眶的图像质量差异进行了定量评估。非圆形眼眶的主要性能优势在于改进了取样和消除了锥束伪影。
{"title":"Cone-beam CT with a noncircular (sine-on-sphere) orbit: imaging performance of a clinical system for image-guided interventions.","authors":"A Kyle Jones, Moiz Ahmad, Shaan M Raza, Stephen R Chen, Jeffrey H Siewerdsen","doi":"10.1117/1.JMI.11.4.043503","DOIUrl":"10.1117/1.JMI.11.4.043503","url":null,"abstract":"<p><strong>Purpose: </strong>We aim to compare the imaging performance of a cone-beam CT (CBCT) imaging system with noncircular scan protocols (sine-on-sphere) to a conventional circular orbit.</p><p><strong>Approach: </strong>A biplane C-arm system (ARTIS Icono; Siemens Healthineers) capable of circular and noncircular CBCT acquisition was used, with the latter orbit (sine-on-sphere, \"Sine Spin\") executing a sinusoidal motion with <math><mrow><mo>±</mo> <mn>10</mn> <mtext> </mtext> <mi>deg</mi></mrow> </math> tilt amplitude over the half-scan orbit. A test phantom was used for the characterization of image uniformity, noise, noise-power spectrum (NPS), spatial resolution [modulation transfer function (MTF) in axial and oblique directions], and cone-beam artifacts. Findings were interpreted using an anthropomorphic head phantom with respect to pertinent tasks in skull base neurosurgery.</p><p><strong>Results: </strong>The noncircular scan protocol exhibited several advantages associated with improved 3D sampling-evident in the NPS as filling of the null cone about the <math> <mrow><msub><mi>f</mi> <mi>z</mi></msub> </mrow> </math> spatial frequency axis and reduction of cone-beam artifacts. The region of support at the longitudinal extrema was reduced from 16 to <math><mrow><mo>∼</mo> <mn>12</mn> <mtext> </mtext> <mi>cm</mi></mrow> </math> at a radial distance of 6.5 cm. Circular and noncircular orbits exhibited nearly identical image uniformity and quantum noise, demonstrating cupping of <math><mrow><mo>-</mo> <mn>16.7</mn> <mo>%</mo></mrow> </math> and overall noise of <math><mrow><mo>∼</mo> <mn>27</mn> <mtext> </mtext> <mi>HU</mi></mrow> </math> . Although both the radially averaged axial MTF ( <math> <mrow><msub><mi>f</mi> <mrow><mi>x</mi> <mo>,</mo> <mi>y</mi></mrow> </msub> </mrow> </math> ) and 45 deg oblique MTF ( <math> <mrow><msub><mi>f</mi> <mrow><mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>,</mo> <mi>z</mi></mrow> </msub> </mrow> </math> ) were <math><mrow><mo>∼</mo> <mn>20</mn> <mo>%</mo></mrow> </math> lower for the noncircular orbit compared with the circular orbit at the default full reconstruction field of view (FOV), there was no difference in spatial resolution for the medium reconstruction FOV (smaller voxel size). Differences in the perceptual image quality for the anthropomorphic phantom reinforced the objective, quantitative findings, including reduced beam-hardening and cone-beam artifacts about structures of interest in the skull base.</p><p><strong>Conclusions: </strong>Image quality differences between circular and noncircular CBCT orbits were quantitatively evaluated on a clinical system in the context of neurosurgery. The primary performance advantage for the noncircular orbit was the improved sampling and elimination of cone-beam artifacts.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"043503"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11342057/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142056921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Purpose: Our study investigates the potential benefits of incorporating prior anatomical knowledge into a deep learning (DL) method designed for the automated segmentation of lung lobes in chest CT scans.
Approach: We introduce an automated DL-based approach that leverages anatomical information from the lung's vascular system to guide and enhance the segmentation process. This involves utilizing a lung vessel connectivity (LVC) map, which encodes relevant lung vessel anatomical data. Our study explores the performance of three different neural network architectures within the nnU-Net framework: a standalone U-Net, a multitasking U-Net, and a cascade U-Net.
Results: Experimental findings suggest that the inclusion of LVC information in the DL model can lead to improved segmentation accuracy, particularly, in the challenging boundary regions of expiration chest CT volumes. Furthermore, our study demonstrates the potential for LVC to enhance the model's generalization capabilities. Finally, the method's robustness is evaluated through the segmentation of lung lobes in 10 cases of COVID-19, demonstrating its applicability in the presence of pulmonary diseases.
Conclusions: Incorporating prior anatomical information, such as LVC, into the DL model shows promise for enhancing segmentation performance, particularly in the boundary regions. However, the extent of this improvement has limitations, prompting further exploration of its practical applicability.
{"title":"Lung vessel connectivity map as anatomical prior knowledge for deep learning-based lung lobe segmentation.","authors":"Simone Bendazzoli, Emelie Bäcklin, Örjan Smedby, Birgitta Janerot-Sjoberg, Bryan Connolly, Chunliang Wang","doi":"10.1117/1.JMI.11.4.044001","DOIUrl":"10.1117/1.JMI.11.4.044001","url":null,"abstract":"<p><strong>Purpose: </strong>Our study investigates the potential benefits of incorporating prior anatomical knowledge into a deep learning (DL) method designed for the automated segmentation of lung lobes in chest CT scans.</p><p><strong>Approach: </strong>We introduce an automated DL-based approach that leverages anatomical information from the lung's vascular system to guide and enhance the segmentation process. This involves utilizing a lung vessel connectivity (LVC) map, which encodes relevant lung vessel anatomical data. Our study explores the performance of three different neural network architectures within the nnU-Net framework: a standalone U-Net, a multitasking U-Net, and a cascade U-Net.</p><p><strong>Results: </strong>Experimental findings suggest that the inclusion of LVC information in the DL model can lead to improved segmentation accuracy, particularly, in the challenging boundary regions of expiration chest CT volumes. Furthermore, our study demonstrates the potential for LVC to enhance the model's generalization capabilities. Finally, the method's robustness is evaluated through the segmentation of lung lobes in 10 cases of COVID-19, demonstrating its applicability in the presence of pulmonary diseases.</p><p><strong>Conclusions: </strong>Incorporating prior anatomical information, such as LVC, into the DL model shows promise for enhancing segmentation performance, particularly in the boundary regions. However, the extent of this improvement has limitations, prompting further exploration of its practical applicability.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"044001"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11231955/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141581239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-08-06DOI: 10.1117/1.JMI.11.4.044505
Heather M Whitney, Roni Yoeli-Bik, Jacques S Abramowicz, Li Lan, Hui Li, Ryan E Longman, Ernst Lengyel, Maryellen L Giger
Purpose: Segmentation of ovarian/adnexal masses from surrounding tissue on ultrasound images is a challenging task. The separation of masses into different components may also be important for radiomic feature extraction. Our study aimed to develop an artificial intelligence-based automatic segmentation method for transvaginal ultrasound images that (1) outlines the exterior boundary of adnexal masses and (2) separates internal components.
Approach: A retrospective ultrasound imaging database of adnexal masses was reviewed for exclusion criteria at the patient, mass, and image levels, with one image per mass. The resulting 54 adnexal masses (36 benign/18 malignant) from 53 patients were separated by patient into training (26 benign/12 malignant) and independent test (10 benign/6 malignant) sets. U-net segmentation performance on test images compared to expert detailed outlines was measured using the Dice similarity coefficient (DSC) and the ratio of the Hausdorff distance to the effective diameter of the outline ( ) for each mass. Subsequently, in discovery mode, a two-level fuzzy c-means (FCM) unsupervised clustering approach was used to separate the pixels within masses belonging to hypoechoic or hyperechoic components.
Results: The DSC (median [95% confidence interval]) was 0.91 [0.78, 0.96], and was 0.04 [0.01, 0.12], indicating strong agreement with expert outlines. Clinical review of the internal separation of masses into echogenic components demonstrated a strong association with mass characteristics.
Conclusion: A combined U-net and FCM algorithm for automatic segmentation of adnexal masses and their internal components achieved excellent results compared with expert outlines and review, supporting future radiomic feature-based classification of the masses by components.
目的:在超声图像上将卵巢/附件肿块从周围组织中分离出来是一项具有挑战性的任务。将肿块分离成不同的组成部分对于放射学特征提取也很重要。我们的研究旨在开发一种基于人工智能的经阴道超声图像自动分割方法,该方法可(1)勾勒出附件肿块的外部边界,(2)分离内部成分:方法:对附件包块的回顾性超声成像数据库进行审查,以确定患者、包块和图像层面的排除标准,每个包块一张图像。将 53 名患者的 54 个附件肿块(36 个良性/18 个恶性)按患者分为训练集(26 个良性/12 个恶性)和独立测试集(10 个良性/6 个恶性)。使用戴斯相似系数(DSC)和豪斯多夫距离与每个肿块轮廓的有效直径之比(R HD - D)来衡量 U 网在测试图像上与专家详细轮廓相比的分割性能。随后,在发现模式下,使用两级模糊均值(FCM)无监督聚类方法将肿块内属于低回声或高回声成分的像素分开:DSC(中位数[95%置信区间])为 0.91 [0.78,0.96],R HD - D 为 0.04 [0.01,0.12],表明与专家轮廓非常一致。对肿块内部回声成分的临床分析表明,肿块内部回声成分与肿块特征密切相关:结论:U-net 和 FCM 算法相结合用于附件肿块及其内部成分的自动分割,与专家轮廓和复查结果相比取得了极佳的效果,支持未来基于放射学特征的肿块成分分类。
{"title":"AI-based automated segmentation for ovarian/adnexal masses and their internal components on ultrasound imaging.","authors":"Heather M Whitney, Roni Yoeli-Bik, Jacques S Abramowicz, Li Lan, Hui Li, Ryan E Longman, Ernst Lengyel, Maryellen L Giger","doi":"10.1117/1.JMI.11.4.044505","DOIUrl":"10.1117/1.JMI.11.4.044505","url":null,"abstract":"<p><strong>Purpose: </strong>Segmentation of ovarian/adnexal masses from surrounding tissue on ultrasound images is a challenging task. The separation of masses into different components may also be important for radiomic feature extraction. Our study aimed to develop an artificial intelligence-based automatic segmentation method for transvaginal ultrasound images that (1) outlines the exterior boundary of adnexal masses and (2) separates internal components.</p><p><strong>Approach: </strong>A retrospective ultrasound imaging database of adnexal masses was reviewed for exclusion criteria at the patient, mass, and image levels, with one image per mass. The resulting 54 adnexal masses (36 benign/18 malignant) from 53 patients were separated by patient into training (26 benign/12 malignant) and independent test (10 benign/6 malignant) sets. U-net segmentation performance on test images compared to expert detailed outlines was measured using the Dice similarity coefficient (DSC) and the ratio of the Hausdorff distance to the effective diameter of the outline ( <math> <mrow> <msub><mrow><mi>R</mi></mrow> <mrow><mi>HD</mi> <mtext>-</mtext> <mi>D</mi></mrow> </msub> </mrow> </math> ) for each mass. Subsequently, in discovery mode, a two-level fuzzy c-means (FCM) unsupervised clustering approach was used to separate the pixels within masses belonging to hypoechoic or hyperechoic components.</p><p><strong>Results: </strong>The DSC (median [95% confidence interval]) was 0.91 [0.78, 0.96], and <math> <mrow> <msub><mrow><mi>R</mi></mrow> <mrow><mi>HD</mi> <mtext>-</mtext> <mi>D</mi></mrow> </msub> </mrow> </math> was 0.04 [0.01, 0.12], indicating strong agreement with expert outlines. Clinical review of the internal separation of masses into echogenic components demonstrated a strong association with mass characteristics.</p><p><strong>Conclusion: </strong>A combined U-net and FCM algorithm for automatic segmentation of adnexal masses and their internal components achieved excellent results compared with expert outlines and review, supporting future radiomic feature-based classification of the masses by components.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"044505"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11301525/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141903209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-01Epub Date: 2024-08-24DOI: 10.1117/1.JMI.11.4.044008
Chenyu Gao, Shunxing Bao, Michael E Kim, Nancy R Newlin, Praitayini Kanakaraj, Tianyuan Yao, Gaurav Rudravaram, Yuankai Huo, Daniel Moyer, Kurt Schilling, Walter A Kukull, Arthur W Toga, Derek B Archer, Timothy J Hohman, Bennett A Landman, Zhiyuan Li
<p><strong>Purpose: </strong>In brain diffusion magnetic resonance imaging (dMRI), the volumetric and bundle analyses of whole-brain tissue microstructure and connectivity can be severely impeded by an incomplete field of view (FOV). We aim to develop a method for imputing the missing slices directly from existing dMRI scans with an incomplete FOV. We hypothesize that the imputed image with a complete FOV can improve whole-brain tractography for corrupted data with an incomplete FOV. Therefore, our approach provides a desirable alternative to discarding the valuable brain dMRI data, enabling subsequent tractography analyses that would otherwise be challenging or unattainable with corrupted data.</p><p><strong>Approach: </strong>We propose a framework based on a deep generative model that estimates the absent brain regions in dMRI scans with an incomplete FOV. The model is capable of learning both the diffusion characteristics in diffusion-weighted images (DWIs) and the anatomical features evident in the corresponding structural images for efficiently imputing missing slices of DWIs in the incomplete part of the FOV.</p><p><strong>Results: </strong>For evaluating the imputed slices, on the Wisconsin Registry for Alzheimer's Prevention (WRAP) dataset, the proposed framework achieved <math> <mrow><msub><mi>PSNR</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>22.397</mn></mrow> </math> , <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>0.905</mn></mrow> </math> , <math> <mrow> <msub><mrow><mi>PSNR</mi></mrow> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>22.479</mn></mrow> </math> , and <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>0.893</mn></mrow> </math> ; on the National Alzheimer's Coordinating Center (NACC) dataset, it achieved <math> <mrow><msub><mi>PSNR</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>21.304</mn></mrow> </math> , <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>0.892</mn></mrow> </math> , <math> <mrow><msub><mi>PSNR</mi> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>21.599</mn></mrow> </math> , and <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>0.877</mn></mrow> </math> . The proposed framework improved the tractography accuracy, as demonstrated by an increased average Dice score for 72 tracts ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) on both the WRAP and NACC datasets.</p><p><strong>Conclusions: </strong>Results suggest that the proposed framework achieved sufficient imputation performance in brain dMRI data with an incomplete FOV for improving whole-brain tractography, thereby repairing the corrupted data. Our approach achieved more accurate whole-brain tractography results with an extended and complete FOV and reduced the uncertainty when analyzing bundles associa
目的:在脑弥散磁共振成像(dMRI)中,不完整的视场(FOV)会严重影响对全脑组织微观结构和连接性的容积和束状分析。我们的目标是开发一种方法,直接从现有的不完整视场的 dMRI 扫描中估算缺失的切片。我们假设,具有完整视场的估算图像可以改善具有不完整视场的损坏数据的全脑束学。因此,我们的方法提供了一种可取的替代方法,而不是丢弃有价值的脑部 dMRI 数据,使后续的牵引成像分析成为可能,否则这些分析将具有挑战性或无法通过损坏的数据实现:方法:我们提出了一个基于深度生成模型的框架,该模型可估算出不完整 FOV 的 dMRI 扫描中缺失的大脑区域。该模型能够学习扩散加权图像(DWIs)中的扩散特征和相应结构图像中明显的解剖学特征,从而有效地估算FOV不完整部分DWIs中缺失的切片:在威斯康星州阿尔茨海默氏症预防注册数据集(WRAP)上评估估算切片时,所提出的框架达到了 PSNR b 0 = 22.397 , SSIM b 0 = 0.905 , PSNR b 1300 = 22.479 ,SSIM b 1300 = 0.893 ;在国家阿尔茨海默氏症协调中心(NACC)数据集上,实现了 PSNR b 0 = 21.304 ,SSIM b 0 = 0.892 ,PSNR b 1300 = 21.599 ,SSIM b 1300 = 0.877 。在 WRAP 和 NACC 数据集上,拟议框架提高了 72 个神经束的平均 Dice 分数(P 0.001),从而提高了神经束绘制的准确性:结果表明,所提出的框架在具有不完整 FOV 的脑 dMRI 数据中实现了足够的估算性能,可用于改善全脑牵引成像,从而修复损坏的数据。在分析与阿尔茨海默病相关的脑束时,我们的方法在扩展的完整 FOV 下获得了更准确的全脑束图结果,并降低了不确定性。
{"title":"Field-of-view extension for brain diffusion MRI via deep generative models.","authors":"Chenyu Gao, Shunxing Bao, Michael E Kim, Nancy R Newlin, Praitayini Kanakaraj, Tianyuan Yao, Gaurav Rudravaram, Yuankai Huo, Daniel Moyer, Kurt Schilling, Walter A Kukull, Arthur W Toga, Derek B Archer, Timothy J Hohman, Bennett A Landman, Zhiyuan Li","doi":"10.1117/1.JMI.11.4.044008","DOIUrl":"10.1117/1.JMI.11.4.044008","url":null,"abstract":"<p><strong>Purpose: </strong>In brain diffusion magnetic resonance imaging (dMRI), the volumetric and bundle analyses of whole-brain tissue microstructure and connectivity can be severely impeded by an incomplete field of view (FOV). We aim to develop a method for imputing the missing slices directly from existing dMRI scans with an incomplete FOV. We hypothesize that the imputed image with a complete FOV can improve whole-brain tractography for corrupted data with an incomplete FOV. Therefore, our approach provides a desirable alternative to discarding the valuable brain dMRI data, enabling subsequent tractography analyses that would otherwise be challenging or unattainable with corrupted data.</p><p><strong>Approach: </strong>We propose a framework based on a deep generative model that estimates the absent brain regions in dMRI scans with an incomplete FOV. The model is capable of learning both the diffusion characteristics in diffusion-weighted images (DWIs) and the anatomical features evident in the corresponding structural images for efficiently imputing missing slices of DWIs in the incomplete part of the FOV.</p><p><strong>Results: </strong>For evaluating the imputed slices, on the Wisconsin Registry for Alzheimer's Prevention (WRAP) dataset, the proposed framework achieved <math> <mrow><msub><mi>PSNR</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>22.397</mn></mrow> </math> , <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>0.905</mn></mrow> </math> , <math> <mrow> <msub><mrow><mi>PSNR</mi></mrow> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>22.479</mn></mrow> </math> , and <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>0.893</mn></mrow> </math> ; on the National Alzheimer's Coordinating Center (NACC) dataset, it achieved <math> <mrow><msub><mi>PSNR</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>21.304</mn></mrow> </math> , <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>0</mn></mrow> </msub> <mo>=</mo> <mn>0.892</mn></mrow> </math> , <math> <mrow><msub><mi>PSNR</mi> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>21.599</mn></mrow> </math> , and <math> <mrow><msub><mi>SSIM</mi> <mrow><mi>b</mi> <mn>1300</mn></mrow> </msub> <mo>=</mo> <mn>0.877</mn></mrow> </math> . The proposed framework improved the tractography accuracy, as demonstrated by an increased average Dice score for 72 tracts ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) on both the WRAP and NACC datasets.</p><p><strong>Conclusions: </strong>Results suggest that the proposed framework achieved sufficient imputation performance in brain dMRI data with an incomplete FOV for improving whole-brain tractography, thereby repairing the corrupted data. Our approach achieved more accurate whole-brain tractography results with an extended and complete FOV and reduced the uncertainty when analyzing bundles associa","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"044008"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11344266/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142056922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Purpose: Endometrial cancer (EC) is one of the most common types of cancer affecting women. While the hematoxylin-and-eosin (H&E) staining remains the standard for histological analysis, the immunohistochemistry (IHC) method provides molecular-level visualizations. Our study proposes a digital staining method to generate the hematoxylin-3,3'-diaminobenzidine (H-DAB) IHC stain of Ki-67 for the whole slide image of the EC tumor from its H&E stain counterpart.
Approach: We employed a color unmixing technique to yield stain density maps from the optical density (OD) of the stains and utilized the U-Net for end-to-end inference. The effectiveness of the proposed method was evaluated using the Pearson correlation between the digital and physical stain's labeling index (LI), a key metric indicating tumor proliferation. Two different cross-validation schemes were designed in our study: intraslide validation and cross-case validation (CCV). In the widely used intraslide scheme, the training and validation sets might include different regions from the same slide. The rigorous CCV validation scheme strictly prohibited any validation slide from contributing to training.
Results: The proposed method yielded a high-resolution digital stain with preserved histological features, indicating a reliable correlation with the physical stain in terms of the Ki-67 LI. In the intraslide scheme, using intraslide patches resulted in a biased accuracy (e.g., ) significantly higher than that of CCV. The CCV scheme retained a fair correlation (e.g., ) between the LIs calculated from the digital stain and its physical IHC counterpart. Inferring the OD of the IHC stain from that of the H&E stain enhanced the correlation metric, outperforming that of the baseline model using the RGB space.
Conclusions: Our study revealed that molecule-level insights could be obtained from H&E images using deep learning. Furthermore, the improvement brought via OD inference indicated a possible method for creating more generalizable models for digital staining via per-stain analysis.
目的:子宫内膜癌(EC)是妇女最常见的癌症类型之一。虽然苏木精-伊红(H&E)染色仍是组织学分析的标准,但免疫组化(IHC)方法可提供分子水平的可视化。我们的研究提出了一种数字染色方法,通过 H&E 染色法生成 EC 肿瘤整张玻片图像中 Ki-67 的苏木精-3,3'-二氨基联苯胺(H-DAB)IHC 染色法:我们采用了一种颜色不混合技术,从染色剂的光密度(OD)得出染色剂密度图,并利用 U-Net 进行端到端推理。我们利用数字染色和物理染色的标记指数(LI)之间的皮尔逊相关性评估了所提方法的有效性。我们的研究设计了两种不同的交叉验证方案:滑动内验证和交叉案例验证(CCV)。在广泛使用的切片内验证方案中,训练集和验证集可能包括来自同一张切片的不同区域。严格的 CCV 验证方案严格禁止任何验证切片参与训练:结果:所提出的方法得到了保留组织学特征的高分辨率数字染色,表明在 Ki-67 LI 方面与物理染色具有可靠的相关性。在滑动内方案中,使用滑动内补丁的偏倚准确度(如 R = 0.98)明显高于 CCV。CCV 方案保留了数字染色与物理 IHC 计算的 LI 之间的相关性(如 R = 0.66)。从 H&E 染色结果推断 IHC 染色结果的 OD 增强了相关性指标,优于使用 RGB 空间的基线模型:我们的研究表明,利用深度学习可以从 H&E 图像中获得分子级的见解。此外,OD 推理带来的改进表明,通过每染色分析为数字染色创建更具通用性的模型是一种可行的方法。
{"title":"Transformation from hematoxylin-and-eosin staining to Ki-67 immunohistochemistry digital staining images using deep learning: experimental validation on the labeling index.","authors":"Cunyuan Ji, Kengo Oshima, Takumi Urata, Fumikazu Kimura, Keiko Ishii, Takeshi Uehara, Kenji Suzuki, Saori Takeyama, Masahiro Yamaguchi","doi":"10.1117/1.JMI.11.4.047501","DOIUrl":"10.1117/1.JMI.11.4.047501","url":null,"abstract":"<p><strong>Purpose: </strong>Endometrial cancer (EC) is one of the most common types of cancer affecting women. While the hematoxylin-and-eosin (H&E) staining remains the standard for histological analysis, the immunohistochemistry (IHC) method provides molecular-level visualizations. Our study proposes a digital staining method to generate the hematoxylin-3,3'-diaminobenzidine (H-DAB) IHC stain of Ki-67 for the whole slide image of the EC tumor from its H&E stain counterpart.</p><p><strong>Approach: </strong>We employed a color unmixing technique to yield stain density maps from the optical density (OD) of the stains and utilized the U-Net for end-to-end inference. The effectiveness of the proposed method was evaluated using the Pearson correlation between the digital and physical stain's labeling index (LI), a key metric indicating tumor proliferation. Two different cross-validation schemes were designed in our study: intraslide validation and cross-case validation (CCV). In the widely used intraslide scheme, the training and validation sets might include different regions from the same slide. The rigorous CCV validation scheme strictly prohibited any validation slide from contributing to training.</p><p><strong>Results: </strong>The proposed method yielded a high-resolution digital stain with preserved histological features, indicating a reliable correlation with the physical stain in terms of the Ki-67 LI. In the intraslide scheme, using intraslide patches resulted in a biased accuracy (e.g., <math><mrow><mi>R</mi> <mo>=</mo> <mn>0.98</mn></mrow> </math> ) significantly higher than that of CCV. The CCV scheme retained a fair correlation (e.g., <math><mrow><mi>R</mi> <mo>=</mo> <mn>0.66</mn></mrow> </math> ) between the LIs calculated from the digital stain and its physical IHC counterpart. Inferring the OD of the IHC stain from that of the H&E stain enhanced the correlation metric, outperforming that of the baseline model using the RGB space.</p><p><strong>Conclusions: </strong>Our study revealed that molecule-level insights could be obtained from H&E images using deep learning. Furthermore, the improvement brought via OD inference indicated a possible method for creating more generalizable models for digital staining via per-stain analysis.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"11 4","pages":"047501"},"PeriodicalIF":1.9,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11287056/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141861255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}