Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2023.12.001
Daniel Güllmar , Wei-Chan Hsu , Jürgen R. Reichenbach
Introduction
Multiple sclerosis (MS) is a complex neurodegenerative disorder that affects the brain and spinal cord. In this study, we applied a deep learning-based approach using the StyleGAN model to explore patterns related to MS and predict disease progression in magnetic resonance images (MRI).
Methods
We trained the StyleGAN model unsupervised using T1-weighted GRE MR images and diffusion-based ADC maps of MS patients and healthy controls. We then used the trained model to resample MR images from real input data and modified them by manipulations in the latent space to simulate MS progression. We analyzed the resulting simulation-related patterns mimicking disease progression by comparing the intensity profiles of the original and manipulated images and determined the brain parenchymal fraction (BPF).
Results
Our results show that MS progression can be simulated by manipulating MR images in the latent space, as evidenced by brain volume loss on both T1-weighted and ADC maps and increasing lesion extent on ADC maps.
Conclusion
Overall, this study demonstrates the potential of the StyleGAN model in medical imaging to study image markers and to shed more light on the relationship between brain atrophy and MS progression through corresponding manipulations in the latent space.
{"title":"Predicting disease-related MRI patterns of multiple sclerosis through GAN-based image editing","authors":"Daniel Güllmar , Wei-Chan Hsu , Jürgen R. Reichenbach","doi":"10.1016/j.zemedi.2023.12.001","DOIUrl":"10.1016/j.zemedi.2023.12.001","url":null,"abstract":"<div><h3>Introduction</h3><p>Multiple sclerosis (MS) is a complex neurodegenerative disorder that affects the brain and spinal cord. In this study, we applied a deep learning-based approach using the StyleGAN model to explore patterns related to MS and predict disease progression in magnetic resonance images (MRI).</p></div><div><h3>Methods</h3><p>We trained the StyleGAN model unsupervised using T<sub>1</sub>-weighted GRE MR images and diffusion-based ADC maps of MS patients and healthy controls. We then used the trained model to resample MR images from real input data and modified them by manipulations in the latent space to simulate MS progression. We analyzed the resulting simulation-related patterns mimicking disease progression by comparing the intensity profiles of the original and manipulated images and determined the brain parenchymal fraction (BPF).</p></div><div><h3>Results</h3><p>Our results show that MS progression can be simulated by manipulating MR images in the latent space, as evidenced by brain volume loss on both T<sub>1</sub>-weighted and ADC maps and increasing lesion extent on ADC maps.</p></div><div><h3>Conclusion</h3><p>Overall, this study demonstrates the potential of the StyleGAN model in medical imaging to study image markers and to shed more light on the relationship between brain atrophy and MS progression through corresponding manipulations in the latent space.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 318-329"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388923001484/pdfft?md5=41054e941858901ec78e1d44ca3d8f6d&pid=1-s2.0-S0939388923001484-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139030422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2022.10.006
Yewei Wang , Yaoying Liu , Yanlin Bai , Qichao Zhou , Shouping Xu , Xueying Pang
Purpose
During the radiation treatment planning process, one of the time-consuming procedures is the final high-resolution dose calculation, which obstacles the wide application of the emerging online adaptive radiotherapy techniques (OLART). There is an urgent desire for highly accurate and efficient dose calculation methods. This study aims to develop a dose super resolution-based deep learning model for fast and accurate dose prediction in clinical practice.
Method
A Multi-stage Dose Super-Resolution Network (MDSR Net) architecture with sparse masks module and multi-stage progressive dose distribution restoration method were developed to predict high-resolution dose distribution using low-resolution data. A total of 340 VMAT plans from different disease sites were used, among which 240 randomly selected nasopharyngeal, lung, and cervix cases were used for model training, and the remaining 60 cases from the same sites for model benchmark testing, and additional 40 cases from the unseen site (breast and rectum) was used for model generalizability evaluation. The clinical calculated dose with a grid size of 2 mm was used as baseline dose distribution. The input included the dose distribution with 4 mm grid size and CT images. The model performance was compared with HD U-Net and cubic interpolation methods using Dose-volume histograms (DVH) metrics and global gamma analysis with 1%/1 mm and 10% low dose threshold. The correlation between the prediction error and the dose, dose gradient, and CT values was also evaluated.
Results
The prediction errors of MDSR were 0.06–0.84% of Dmean indices, and the gamma passing rate was 83.1–91.0% on the benchmark testing dataset, and 0.02–1.03% and 71.3–90.3% for the generalization dataset respectively. The model performance was significantly higher than the HD U-Net and interpolation methods (p < 0.05). The mean errors of the MDSR model decreased (monotonously by 0.03–0.004%) with dose and increased (by 0.01–0.73%) with the dose gradient. There was no correlation between prediction errors and the CT values.
Conclusion
The proposed MDSR model achieved good agreement with the baseline high-resolution dose distribution, with small prediction errors for DVH indices and high gamma passing rate for both seen and unseen sites, indicating a robust and generalizable dose prediction model. The model can provide fast and accurate high-resolution dose distribution for clinical dose calculation, particularly for the routine practice of OLART.
{"title":"A generalization performance study on the boosting radiotherapy dose calculation engine based on super-resolution","authors":"Yewei Wang , Yaoying Liu , Yanlin Bai , Qichao Zhou , Shouping Xu , Xueying Pang","doi":"10.1016/j.zemedi.2022.10.006","DOIUrl":"10.1016/j.zemedi.2022.10.006","url":null,"abstract":"<div><h3>Purpose</h3><p>During the radiation treatment planning process, one of the time-consuming procedures is the final high-resolution dose calculation, which obstacles the wide application of the emerging online adaptive radiotherapy techniques (OLART). There is an urgent desire for highly accurate and efficient dose calculation methods. This study aims to develop a dose super resolution-based deep learning model for fast and accurate dose prediction in clinical practice.</p></div><div><h3>Method</h3><p>A Multi-stage Dose Super-Resolution Network (MDSR Net) architecture with sparse masks module and multi-stage progressive dose distribution restoration method were developed to predict high-resolution dose distribution using low-resolution data. A total of 340 VMAT plans from different disease sites were used, among which 240 randomly selected nasopharyngeal, lung, and cervix cases were used for model training, and the remaining 60 cases from the same sites for model benchmark testing, and additional 40 cases from the unseen site (breast and rectum) was used for model generalizability evaluation. The clinical calculated dose with a grid size of 2 mm was used as baseline dose distribution. The input included the dose distribution with 4 mm grid size and CT images. The model performance was compared with HD U-Net and cubic interpolation methods using Dose-volume histograms (DVH) metrics and global gamma analysis with 1%/1 mm and 10% low dose threshold. The correlation between the prediction error and the dose, dose gradient, and CT values was also evaluated.</p></div><div><h3>Results</h3><p>The prediction errors of MDSR were 0.06–0.84% of D<sub>mean</sub> indices, and the gamma passing rate was 83.1–91.0% on the benchmark testing dataset, and 0.02–1.03% and 71.3–90.3% for the generalization dataset respectively. The model performance was significantly higher than the HD U-Net and interpolation methods (<em>p</em> < 0.05). The mean errors of the MDSR model decreased (monotonously by 0.03–0.004%) with dose and increased (by 0.01–0.73%) with the dose gradient. There was no correlation between prediction errors and the CT values.</p></div><div><h3>Conclusion</h3><p>The proposed MDSR model achieved good agreement with the baseline high-resolution dose distribution, with small prediction errors for DVH indices and high gamma passing rate for both seen and unseen sites, indicating a robust and generalizable dose prediction model. The model can provide fast and accurate high-resolution dose distribution for clinical dose calculation, particularly for the routine practice of OLART.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 208-217"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388922001003/pdfft?md5=5beaf64e5d3600c18adc8f8420659d04&pid=1-s2.0-S0939388922001003-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10511675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2023.08.001
Anish Raj , Fabian Tollens , Anna Caroli , Dominik Nörenberg , Frank G. Zöllner
An accurate prognosis of renal function decline in Autosomal Dominant Polycystic Kidney Disease (ADPKD) is crucial for early intervention. Current biomarkers used are height-adjusted total kidney volume (HtTKV), estimated glomerular filtration rate (eGFR), and patient age. However, manually measuring kidney volume is time-consuming and subject to observer variability. Additionally, incorporating automatically generated features from kidney MRI images, along with conventional biomarkers, can enhance prognostic improvement. To address these issues, we developed two deep-learning algorithms. Firstly, an automated kidney volume segmentation model accurately calculates HtTKV. Secondly, we utilize segmented kidney volumes, predicted HtTKV, age, and baseline eGFR to predict chronic kidney disease (CKD) stages =3A, =3B, and a 30% decline in eGFR after 8 years from the baseline visit. Our approach combines a convolutional neural network (CNN) and a multi-layer perceptron (MLP). Our study included 135 subjects and the AUC scores obtained were 0.96, 0.96, and 0.95 for CKD stages =3A, =3B, and a 30% decline in eGFR, respectively. Furthermore, our algorithm achieved a Pearson correlation coefficient of 0.81 between predicted and measured eGFR decline. We extended our approach to predict distinct CKD stages after eight years with an AUC of 0.97. The proposed approach has the potential to enhance monitoring and facilitate prognosis in ADPKD patients, even in the early disease stages.
{"title":"Automated prognosis of renal function decline in ADPKD patients using deep learning","authors":"Anish Raj , Fabian Tollens , Anna Caroli , Dominik Nörenberg , Frank G. Zöllner","doi":"10.1016/j.zemedi.2023.08.001","DOIUrl":"10.1016/j.zemedi.2023.08.001","url":null,"abstract":"<div><p>An accurate prognosis of renal function decline in Autosomal Dominant Polycystic Kidney Disease (ADPKD) is crucial for early intervention. Current biomarkers used are height-adjusted total kidney volume (HtTKV), estimated glomerular filtration rate (eGFR), and patient age. However, manually measuring kidney volume is time-consuming and subject to observer variability. Additionally, incorporating automatically generated features from kidney MRI images, along with conventional biomarkers, can enhance prognostic improvement. To address these issues, we developed two deep-learning algorithms. Firstly, an automated kidney volume segmentation model accurately calculates HtTKV. Secondly, we utilize segmented kidney volumes, predicted HtTKV, age, and baseline eGFR to predict chronic kidney disease (CKD) stages <span><math><mrow><mo>></mo></mrow></math></span>=3A, <span><math><mrow><mo>></mo></mrow></math></span>=3B, and a 30% decline in eGFR after 8 years from the baseline visit. Our approach combines a convolutional neural network (CNN) and a multi-layer perceptron (MLP). Our study included 135 subjects and the AUC scores obtained were 0.96, 0.96, and 0.95 for CKD stages <span><math><mrow><mo>></mo></mrow></math></span>=3A, <span><math><mrow><mo>></mo></mrow></math></span>=3B, and a 30% decline in eGFR, respectively. Furthermore, our algorithm achieved a Pearson correlation coefficient of 0.81 between predicted and measured eGFR decline. We extended our approach to predict distinct CKD stages after eight years with an AUC of 0.97. The proposed approach has the potential to enhance monitoring and facilitate prognosis in ADPKD patients, even in the early disease stages.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 330-342"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388923000909/pdfft?md5=f7bc065601b8dfd2bfb240a0fa1328c0&pid=1-s2.0-S0939388923000909-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10056256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2022.11.008
Artem Zatcepin , Anna Kopczak , Adrien Holzgreve , Sandra Hein , Andreas Schindler , Marco Duering , Lena Kaiser , Simon Lindner , Martin Schidlowski , Peter Bartenstein , Nathalie Albert , Matthias Brendel , Sibylle I. Ziegler
Introduction
Neuroinflammation evaluation after acute ischemic stroke is a promising option for selecting an appropriate post-stroke treatment strategy. To assess neuroinflammation in vivo, translocator protein PET (TSPO PET) can be used. However, the gold standard TSPO PET quantification method includes a 90 min scan and continuous arterial blood sampling, which is challenging to perform on a routine basis. In this work, we determine what information is required for a simplified quantification approach using a machine learning algorithm.
Materials and Methods
We analyzed data from 18 patients with ischemic stroke who received 0–90 min [18F]GE-180 PET as well as T1-weigted (T1w), FLAIR, and arterial spin labeling (ASL) MRI scans. During PET scans, five manual venous blood samples at 5, 15, 30, 60, and 85 min post injection (p.i.) were drawn, and plasma activity concentration was measured. Total distribution volume (VT) was calculated using Logan plot with the full dynamic PET and an image-derived input function (IDIF) from the carotid arteries. IDIF was scaled by a calibration factor derived from all the measured plasma activity concentrations. The calculated VT values were used for training a random forest regressor. As input features for the model, we used three late PET frames (60–70, 70–80, and 80–90 min p.i.), the ASL image reflecting perfusion, the voxel coordinates, the lesion mask, and the five plasma activity concentrations. The algorithm was validated with the leave-one-out approach. To estimate the impact of the individual features on the algorithm’s performance, we used Shapley Additive Explanations (SHAP). Having determined that the three late PET frames and the plasma activity concentrations were the most important features, we tested a simplified quantification approach consisting of dividing a late PET frame by a plasma activity concentration. All the combinations of frames/samples were compared by means of concordance correlation coefficient and Bland-Altman plots.
Results
When using all the input features, the algorithm predicted VT values with high accuracy (87.8 ± 8.3%) for both lesion and non-lesion voxels. The SHAP values demonstrated high impact of the late PET frames (60–70, 70–80, and 80–90 min p.i.) and plasma activity concentrations on the VT prediction, while the influence of the ASL-derived perfusion, voxel coordinates, and the lesion mask was low. Among all the combinations of the late PET frames and plasma activity concentrations, the 70–80 min p.i. frame divided by the 30 min p.i. plasma sample produced the closest VT estimate in the ischemic lesion.
Conclusion
Reliable TSPO PET quantification is achievable by using a single late PET frame divided by a late blood sample activity concentration.
简介急性缺血性中风后的神经炎症评估是选择适当的中风后治疗策略的一个很有前景的选择。要评估体内的神经炎症,可以使用转运蛋白 PET(TSPO PET)。然而,金标准 TSPO PET 定量方法包括 90 分钟扫描和连续动脉血采样,这对常规操作具有挑战性。在这项工作中,我们利用机器学习算法确定了简化量化方法所需的信息:我们分析了 18 位缺血性中风患者的数据,这些患者接受了 0-90 分钟[18F]GE-180 PET 以及 T1 权衡 (T1w)、FLAIR 和动脉自旋标记 (ASL) MRI 扫描。在 PET 扫描期间,分别在注射后 5、15、30、60 和 85 分钟(p.i.)手动抽取五份静脉血样本,并测量血浆活性浓度。总分布容积(VT)是通过全动态 PET Logan 图和颈动脉图像输入函数(IDIF)计算得出的。IDIF 根据所有测量的血浆活性浓度得出的校准因子进行缩放。计算出的 VT 值用于训练随机森林回归器。作为模型的输入特征,我们使用了三个晚期 PET 帧(60-70、70-80 和 80-90 分钟 p.i.)、反映灌注的 ASL 图像、体素坐标、病灶掩膜和五个血浆活性浓度。该算法通过 "留一弃一 "方法进行了验证。为了估计各个特征对算法性能的影响,我们使用了夏普利相加解释(SHAP)。在确定三个晚期 PET 帧和血浆活性浓度是最重要的特征后,我们测试了一种简化的量化方法,即用晚期 PET 帧除以血浆活性浓度。我们通过一致性相关系数和布兰-阿尔特曼图对所有帧/样本组合进行了比较:结果:当使用所有输入特征时,该算法预测病变和非病变体素的 VT 值的准确率都很高(87.8 ± 8.3%)。SHAP值显示晚期PET帧(60-70、70-80和80-90分钟p.i.)和血浆活性浓度对VT预测的影响较大,而ASL衍生灌注、体素坐标和病变掩膜的影响较小。在PET晚期帧和血浆活动浓度的所有组合中,70-80分钟p.i.帧除以30分钟p.i.血浆样本得出的缺血性病变VT估计值最接近:结论:通过使用单个晚期 PET 帧除以晚期血样活性浓度,可以实现可靠的 TSPO PET 定量。
{"title":"Machine learning-based approach reveals essential features for simplified TSPO PET quantification in ischemic stroke patients","authors":"Artem Zatcepin , Anna Kopczak , Adrien Holzgreve , Sandra Hein , Andreas Schindler , Marco Duering , Lena Kaiser , Simon Lindner , Martin Schidlowski , Peter Bartenstein , Nathalie Albert , Matthias Brendel , Sibylle I. Ziegler","doi":"10.1016/j.zemedi.2022.11.008","DOIUrl":"10.1016/j.zemedi.2022.11.008","url":null,"abstract":"<div><h3>Introduction</h3><p>Neuroinflammation evaluation after acute ischemic stroke is a promising option for selecting an appropriate post-stroke treatment strategy. To assess neuroinflammation <em>in vivo</em>, translocator protein PET (TSPO PET) can be used. However, the gold standard TSPO PET quantification method includes a 90 min scan and continuous arterial blood sampling, which is challenging to perform on a routine basis. In this work, we determine what information is required for a simplified quantification approach using a machine learning algorithm.</p></div><div><h3>Materials and Methods</h3><p>We analyzed data from 18 patients with ischemic stroke who received 0–90 min [<sup>18</sup>F]GE-180 PET as well as T1-weigted (T1w), FLAIR, and arterial spin labeling (ASL) MRI scans. During PET scans, five manual venous blood samples at 5, 15, 30, 60, and 85 min post injection (p.i.) were drawn, and plasma activity concentration was measured. Total distribution volume (V<sub>T</sub>) was calculated using Logan plot with the full dynamic PET and an image-derived input function (IDIF) from the carotid arteries. IDIF was scaled by a calibration factor derived from all the measured plasma activity concentrations. The calculated V<sub>T</sub> values were used for training a random forest regressor. As input features for the model, we used three late PET frames (60–70, 70–80, and 80–90 min p.i.), the ASL image reflecting perfusion, the voxel coordinates, the lesion mask, and the five plasma activity concentrations. The algorithm was validated with the leave-one-out approach. To estimate the impact of the individual features on the algorithm’s performance, we used Shapley Additive Explanations (SHAP). Having determined that the three late PET frames and the plasma activity concentrations were the most important features, we tested a simplified quantification approach consisting of dividing a late PET frame by a plasma activity concentration. All the combinations of frames/samples were compared by means of concordance correlation coefficient and Bland-Altman plots.</p></div><div><h3>Results</h3><p>When using all the input features, the algorithm predicted V<sub>T</sub> values with high accuracy (87.8 ± 8.3%) for both lesion and non-lesion voxels. The SHAP values demonstrated high impact of the late PET frames (60–70, 70–80, and 80–90 min p.i.) and plasma activity concentrations on the V<sub>T</sub> prediction, while the influence of the ASL-derived perfusion, voxel coordinates, and the lesion mask was low. Among all the combinations of the late PET frames and plasma activity concentrations, the 70–80 min p.i. frame divided by the 30 min p.i. plasma sample produced the closest V<sub>T</sub> estimate in the ischemic lesion.</p></div><div><h3>Conclusion</h3><p>Reliable TSPO PET quantification is achievable by using a single late PET frame divided by a late blood sample activity concentration.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 218-230"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388922001283/pdfft?md5=f1499fef4d918f1109e5e15fcbad1787&pid=1-s2.0-S0939388922001283-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10567400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2023.05.003
Anika Strittmatter, Lothar R. Schad, Frank G. Zöllner
Multimodal image registration is applied in medical image analysis as it allows the integration of complementary data from multiple imaging modalities. In recent years, various neural network-based approaches for medical image registration have been presented in papers, but due to the use of different datasets, a fair comparison is not possible. In this research 20 different neural networks for an affine registration of medical images were implemented. The networks’ performance and the networks’ generalizability to new datasets were evaluated using two multimodal datasets - a synthetic and a real patient dataset - of three-dimensional CT and MR images of the liver. The networks were first trained semi-supervised using the synthetic dataset and then evaluated on the synthetic dataset and the unseen patient dataset. Afterwards, the networks were finetuned on the patient dataset and subsequently evaluated on the patient dataset. The networks were compared using our own developed CNN as benchmark and a conventional affine registration with SimpleElastix as baseline. Six networks improved the pre-registration Dice coefficient of the synthetic dataset significantly (p-value 0.05) and nine networks improved the pre-registration Dice coefficient of the patient dataset significantly and are therefore able to generalize to the new datasets used in our experiments. Many different machine learning-based methods have been proposed for affine multimodal medical image registration, but few are generalizable to new data and applications. It is therefore necessary to conduct further research in order to develop medical image registration techniques that can be applied more widely.
{"title":"Deep learning-based affine medical image registration for multimodal minimal-invasive image-guided interventions – A comparative study on generalizability","authors":"Anika Strittmatter, Lothar R. Schad, Frank G. Zöllner","doi":"10.1016/j.zemedi.2023.05.003","DOIUrl":"10.1016/j.zemedi.2023.05.003","url":null,"abstract":"<div><p>Multimodal image registration is applied in medical image analysis as it allows the integration of complementary data from multiple imaging modalities. In recent years, various neural network-based approaches for medical image registration have been presented in papers, but due to the use of different datasets, a fair comparison is not possible. In this research 20 different neural networks for an affine registration of medical images were implemented. The networks’ performance and the networks’ generalizability to new datasets were evaluated using two multimodal datasets - a synthetic and a real patient dataset - of three-dimensional CT and MR images of the liver. The networks were first trained semi-supervised using the synthetic dataset and then evaluated on the synthetic dataset and the unseen patient dataset. Afterwards, the networks were finetuned on the patient dataset and subsequently evaluated on the patient dataset. The networks were compared using our own developed CNN as benchmark and a conventional affine registration with SimpleElastix as baseline. Six networks improved the pre-registration Dice coefficient of the synthetic dataset significantly (<em>p</em>-value <span><math><mrow><mo><</mo></mrow></math></span> 0.05) and nine networks improved the pre-registration Dice coefficient of the patient dataset significantly and are therefore able to generalize to the new datasets used in our experiments. Many different machine learning-based methods have been proposed for affine multimodal medical image registration, but few are generalizable to new data and applications. It is therefore necessary to conduct further research in order to develop medical image registration techniques that can be applied more widely.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 291-317"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388923000715/pdfft?md5=8bc88c35e2779691cc7ef560e61e14e3&pid=1-s2.0-S0939388923000715-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9683952","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2022.11.010
Omer Turk , Davut Ozhan , Emrullah Acar , Tahir Cetin Akinci , Musa Yilmaz
Today, as in every life-threatening disease, early diagnosis of brain tumors plays a life-saving role. The brain tumor is formed by the transformation of brain cells from their normal structures into abnormal cell structures. These formed abnormal cells begin to form in masses in the brain regions. Nowadays, many different techniques are employed to detect these tumor masses, and the most common of these techniques is Magnetic Resonance Imaging (MRI). In this study, it is aimed to automatically detect brain tumors with the help of ensemble deep learning architectures (ResNet50, VGG19, InceptionV3 and MobileNet) and Class Activation Maps (CAMs) indicators by employing MRI images. The proposed system was implemented in three stages. In the first stage, it was determined whether there was a tumor in the MR images (Binary Approach). In the second stage, different tumor types (Normal, Glioma Tumor, Meningioma Tumor, Pituitary Tumor) were detected from MR images (Multi-class Approach). In the last stage, CAMs of each tumor group were created as an alternative tool to facilitate the work of specialists in tumor detection. The results showed that the overall accuracy of the binary approach was calculated as 100% on the ResNet50, InceptionV3 and MobileNet architectures, and 99.71% on the VGG19 architecture. Moreover, the accuracy values of 96.45% with ResNet50, 93.40% with VGG19, 85.03% with InceptionV3 and 89.34% with MobileNet architectures were obtained in the multi-class approach.
{"title":"Automatic detection of brain tumors with the aid of ensemble deep learning architectures and class activation map indicators by employing magnetic resonance images","authors":"Omer Turk , Davut Ozhan , Emrullah Acar , Tahir Cetin Akinci , Musa Yilmaz","doi":"10.1016/j.zemedi.2022.11.010","DOIUrl":"10.1016/j.zemedi.2022.11.010","url":null,"abstract":"<div><p>Today, as in every life-threatening disease, early diagnosis of brain tumors plays a life-saving role. The brain tumor is formed by the transformation of brain cells from their normal structures into abnormal cell structures. These formed abnormal cells begin to form in masses in the brain regions. Nowadays, many different techniques are employed to detect these tumor masses, and the most common of these techniques is Magnetic Resonance Imaging (MRI). In this study, it is aimed to automatically detect brain tumors with the help of ensemble deep learning architectures (ResNet50, VGG19, InceptionV3 and MobileNet) and Class Activation Maps (CAMs) indicators by employing MRI images. The proposed system was implemented in three stages. In the first stage, it was determined whether there was a tumor in the MR images (Binary Approach). In the second stage, different tumor types (Normal, Glioma Tumor, Meningioma Tumor, Pituitary Tumor) were detected from MR images (Multi-class Approach). In the last stage, CAMs of each tumor group were created as an alternative tool to facilitate the work of specialists in tumor detection. The results showed that the overall accuracy of the binary approach was calculated as 100% on the ResNet50, InceptionV3 and MobileNet architectures, and 99.71% on the VGG19 architecture. Moreover, the accuracy values of 96.45% with ResNet50, 93.40% with VGG19, 85.03% with InceptionV3 and 89.34% with MobileNet architectures were obtained in the multi-class approach.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 278-290"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388922001313/pdfft?md5=e55da35d209b688226a3577197edb180&pid=1-s2.0-S0939388922001313-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"10466945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2023.07.005
Tobit Führes , Marc Saake , Jennifer Lorenz , Hannes Seuss , Sebastian Bickelhaupt , Michael Uder , Frederik Bernd Laun
Purpose
This research aims to develop a feature-guided deep learning approach and compare it with an optimized conventional post-processing algorithm in order to enhance the image quality of diffusion-weighted liver images and, in particular, to reduce the pulsation-induced signal loss occurring predominantly in the left liver lobe.
Methods
Data from 40 patients with liver lesions were used. For the conventional approach, the best-suited out of five examined algorithms was chosen. For the deep learning approach, a U-Net was trained. Instead of learning “gold-standard” target images, the network was trained to optimize four image features (lesion CNR, vessel darkness, data consistency, and pulsation artifact reduction), which could be assessed quantitatively using manually drawn ROIs. A quality score was calculated from these four features. As an additional quality assessment, three radiologists rated different features of the resulting images.
Results
The conventional approach could substantially increase the lesion CNR and reduce the pulsation-induced signal loss. However, the vessel darkness was reduced. The deep learning approach increased the lesion CNR and reduced the signal loss to a slightly lower extent, but it could additionally increase the vessel darkness. According to the image quality score, the quality of the deep-learning images was higher than that of the images obtained using the conventional approach. The radiologist ratings were mostly consistent with the quantitative scores, but the overall quality ratings differed among the readers.
Conclusion
Unlike the conventional algorithm, the deep-learning algorithm increased the vessel darkness. Therefore, it may be a viable alternative to conventional algorithms.
目的 本研究旨在开发一种以特征为导向的深度学习方法,并将其与优化的传统后处理算法进行比较,以提高扩散加权肝脏图像的质量,尤其是减少主要发生在左肝叶的脉动引起的信号损失。在传统方法中,选择了五种已研究过的算法中最合适的一种。对于深度学习方法,则采用 U-Net 进行训练。该网络不是学习 "黄金标准 "目标图像,而是通过训练来优化四个图像特征(病变 CNR、血管暗度、数据一致性和脉动伪影减少),这些特征可通过手动绘制的 ROI 进行定量评估。根据这四个特征计算出质量分数。作为额外的质量评估,三位放射科医生对所得图像的不同特征进行了评分。但是,血管的暗度降低了。深度学习方法提高了病变 CNR,减少了信号损失,但程度略低,而且还增加了血管暗度。根据图像质量评分,深度学习图像的质量高于使用传统方法获得的图像。结论与传统算法不同,深度学习算法增加了血管暗度。因此,它可能是传统算法的可行替代方案。
{"title":"Feature-guided deep learning reduces signal loss and increases lesion CNR in diffusion-weighted imaging of the liver","authors":"Tobit Führes , Marc Saake , Jennifer Lorenz , Hannes Seuss , Sebastian Bickelhaupt , Michael Uder , Frederik Bernd Laun","doi":"10.1016/j.zemedi.2023.07.005","DOIUrl":"10.1016/j.zemedi.2023.07.005","url":null,"abstract":"<div><h3><strong>Purpose</strong></h3><p>This research aims to develop a feature-guided deep learning approach and compare it with an optimized conventional post-processing algorithm in order to enhance the image quality of diffusion-weighted liver images and, in particular, to reduce the pulsation-induced signal loss occurring predominantly in the left liver lobe.</p></div><div><h3><strong>Methods</strong></h3><p>Data from 40 patients with liver lesions were used. For the conventional approach, the best-suited out of five examined algorithms was chosen. For the deep learning approach, a U-Net was trained. Instead of learning “gold-standard” target images, the network was trained to optimize four image features (lesion CNR, vessel darkness, data consistency, and pulsation artifact reduction), which could be assessed quantitatively using manually drawn ROIs. A quality score was calculated from these four features. As an additional quality assessment, three radiologists rated different features of the resulting images.</p></div><div><h3><strong>Results</strong></h3><p>The conventional approach could substantially increase the lesion CNR and reduce the pulsation-induced signal loss. However, the vessel darkness was reduced. The deep learning approach increased the lesion CNR and reduced the signal loss to a slightly lower extent, but it could additionally increase the vessel darkness. According to the image quality score, the quality of the deep-learning images was higher than that of the images obtained using the conventional approach. The radiologist ratings were mostly consistent with the quantitative scores, but the overall quality ratings differed among the readers.</p></div><div><h3><strong>Conclusion</strong></h3><p>Unlike the conventional algorithm, the deep-learning algorithm increased the vessel darkness. Therefore, it may be a viable alternative to conventional algorithms.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 258-269"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388923000879/pdfft?md5=b3e5b6c0be696f64222a77e9bdedeec2&pid=1-s2.0-S0939388923000879-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9931929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2023.05.001
Marcel Nachbar , Monica lo Russo , Cihan Gani , Simon Boeke , Daniel Wegener , Frank Paulsen , Daniel Zips , Thais Roque , Nikos Paragios , Daniela Thorwarth
<div><h3>Background and purpose</h3><p>MR-guided radiotherapy (MRgRT) online plan adaptation accounts for tumor volume changes, interfraction motion and thus allows daily sparing of relevant organs at risk. Due to the high interfraction variability of bladder and rectum, patients with tumors in the pelvic region may strongly benefit from adaptive MRgRT. Currently, fast automatic annotation of anatomical structures is not available within the online MRgRT workflow. Therefore, the aim of this study was to train and validate a fast, accurate deep learning model for automatic MRI segmentation at the MR-Linac for future implementation in a clinical MRgRT workflow.</p></div><div><h3>Materials and methods</h3><p>For a total of 47 patients, T2w MRI data were acquired on a 1.5 T MR-Linac (Unity, Elekta) on five different days. Prostate, seminal vesicles, rectum, anal canal, bladder, penile bulb, body and bony structures were manually annotated. These training data consisting of 232 data sets in total was used for the generation of a deep learning based autocontouring model and validated on 20 unseen T2w-MRIs. For quantitative evaluation the validation set was contoured by a radiation oncologist as gold standard contours (GSC) and compared in MATLAB to the automatic contours (AIC). For the evaluation, dice similarity coefficients (DSC), and 95% Hausdorff distances (95% HD), added path length (APL) and surface DSC (sDSC) were calculated in a caudal-cranial window of <span><math><mrow><mo>±</mo></mrow></math></span> 4 cm with respect to the prostate ends. For qualitative evaluation, five radiation oncologists scored the AIC on the possible usage within an online adaptive workflow as follows: (1) no modifications needed, (2) minor adjustments needed, (3) major adjustments/ multiple minor adjustments needed, (4) not usable.</p></div><div><h3>Results</h3><p>The quantitative evaluation revealed a maximum median 95% HD of 6.9 mm for the rectum and minimum median 95% HD of 2.7 mm for the bladder. Maximal and minimal median DSC were detected for bladder with 0.97 and for penile bulb with 0.73, respectively. Using a tolerance level of 3 mm, the highest and lowest sDSC were determined for rectum (0.94) and anal canal (0.68), respectively.</p><p>Qualitative evaluation resulted in a mean score of 1.2 for AICs over all organs and patients across all expert ratings. For the different autocontoured structures, the highest mean score of 1.0 was observed for anal canal, sacrum, femur left and right, and pelvis left, whereas for prostate the lowest mean score of 2.0 was detected. In total, 80% of the contours were rated be clinically acceptable, 16% to require minor and 4% major adjustments for online adaptive MRgRT.</p></div><div><h3>Conclusion</h3><p>In this study, an AI-based autocontouring was successfully trained for online adaptive MR-guided radiotherapy on the 1.5 T MR-Linac system. The developed model can automatically generate contours accepted by physicians (80%) o
{"title":"Automatic AI-based contouring of prostate MRI for online adaptive radiotherapy","authors":"Marcel Nachbar , Monica lo Russo , Cihan Gani , Simon Boeke , Daniel Wegener , Frank Paulsen , Daniel Zips , Thais Roque , Nikos Paragios , Daniela Thorwarth","doi":"10.1016/j.zemedi.2023.05.001","DOIUrl":"10.1016/j.zemedi.2023.05.001","url":null,"abstract":"<div><h3>Background and purpose</h3><p>MR-guided radiotherapy (MRgRT) online plan adaptation accounts for tumor volume changes, interfraction motion and thus allows daily sparing of relevant organs at risk. Due to the high interfraction variability of bladder and rectum, patients with tumors in the pelvic region may strongly benefit from adaptive MRgRT. Currently, fast automatic annotation of anatomical structures is not available within the online MRgRT workflow. Therefore, the aim of this study was to train and validate a fast, accurate deep learning model for automatic MRI segmentation at the MR-Linac for future implementation in a clinical MRgRT workflow.</p></div><div><h3>Materials and methods</h3><p>For a total of 47 patients, T2w MRI data were acquired on a 1.5 T MR-Linac (Unity, Elekta) on five different days. Prostate, seminal vesicles, rectum, anal canal, bladder, penile bulb, body and bony structures were manually annotated. These training data consisting of 232 data sets in total was used for the generation of a deep learning based autocontouring model and validated on 20 unseen T2w-MRIs. For quantitative evaluation the validation set was contoured by a radiation oncologist as gold standard contours (GSC) and compared in MATLAB to the automatic contours (AIC). For the evaluation, dice similarity coefficients (DSC), and 95% Hausdorff distances (95% HD), added path length (APL) and surface DSC (sDSC) were calculated in a caudal-cranial window of <span><math><mrow><mo>±</mo></mrow></math></span> 4 cm with respect to the prostate ends. For qualitative evaluation, five radiation oncologists scored the AIC on the possible usage within an online adaptive workflow as follows: (1) no modifications needed, (2) minor adjustments needed, (3) major adjustments/ multiple minor adjustments needed, (4) not usable.</p></div><div><h3>Results</h3><p>The quantitative evaluation revealed a maximum median 95% HD of 6.9 mm for the rectum and minimum median 95% HD of 2.7 mm for the bladder. Maximal and minimal median DSC were detected for bladder with 0.97 and for penile bulb with 0.73, respectively. Using a tolerance level of 3 mm, the highest and lowest sDSC were determined for rectum (0.94) and anal canal (0.68), respectively.</p><p>Qualitative evaluation resulted in a mean score of 1.2 for AICs over all organs and patients across all expert ratings. For the different autocontoured structures, the highest mean score of 1.0 was observed for anal canal, sacrum, femur left and right, and pelvis left, whereas for prostate the lowest mean score of 2.0 was detected. In total, 80% of the contours were rated be clinically acceptable, 16% to require minor and 4% major adjustments for online adaptive MRgRT.</p></div><div><h3>Conclusion</h3><p>In this study, an AI-based autocontouring was successfully trained for online adaptive MR-guided radiotherapy on the 1.5 T MR-Linac system. The developed model can automatically generate contours accepted by physicians (80%) o","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 197-207"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388923000533/pdfft?md5=4c8a5787fe97a32ec18b4426b3597127&pid=1-s2.0-S0939388923000533-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"9562444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.zemedi.2024.02.001
Lorenzo Mercolli, Axel Rominger, Kuangyu Shi
The use of artificial intelligence systems in clinical routine is still hampered by the necessity of a medical device certification and/or by the difficulty of implementing these systems in a clinic’s quality management system. In this context, the key questions for a user are how to ensure robust model predictions and how to appraise the quality of a model’s results on a regular basis.
In this paper we discuss some conceptual foundation for a clinical implementation of a machine learning system and argue that both vendors and users should take certain responsibilities, as is already common practice for high-risk medical equipment.
We propose the methodology from AAPM Task Group 100 report No. 283 as a conceptual framework for developing risk-driven a quality management program for a clinical process that encompasses a machine learning system. This is illustrated with an example of a clinical workflow. Our analysis shows how the risk evaluation in this framework can accommodate artificial intelligence based systems independently of their robustness evaluation or the user’s in–house expertise. In particular, we highlight how the degree of interpretability of a machine learning system can be systematically accounted for within the risk evaluation and in the development of a quality management system.
{"title":"Towards quality management of artificial intelligence systems for medical applications","authors":"Lorenzo Mercolli, Axel Rominger, Kuangyu Shi","doi":"10.1016/j.zemedi.2024.02.001","DOIUrl":"10.1016/j.zemedi.2024.02.001","url":null,"abstract":"<div><p>The use of artificial intelligence systems in clinical routine is still hampered by the necessity of a medical device certification and/or by the difficulty of implementing these systems in a clinic’s quality management system. In this context, the key questions for a user are how to ensure robust model predictions and how to appraise the quality of a model’s results on a regular basis.</p><p>In this paper we discuss some conceptual foundation for a clinical implementation of a machine learning system and argue that both vendors and users should take certain responsibilities, as is already common practice for high-risk medical equipment.</p><p>We propose the methodology from AAPM Task Group 100 report No. 283 as a conceptual framework for developing risk-driven a quality management program for a clinical process that encompasses a machine learning system. This is illustrated with an example of a clinical workflow. Our analysis shows how the risk evaluation in this framework can accommodate artificial intelligence based systems independently of their robustness evaluation or the user’s in–house expertise. In particular, we highlight how the degree of interpretability of a machine learning system can be systematically accounted for within the risk evaluation and in the development of a quality management system.</p></div>","PeriodicalId":54397,"journal":{"name":"Zeitschrift fur Medizinische Physik","volume":"34 2","pages":"Pages 343-352"},"PeriodicalIF":2.0,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0939388924000242/pdfft?md5=309f6a0c3aedbe399d5a372c060278f6&pid=1-s2.0-S0939388924000242-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139984975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}