首页 > 最新文献

Medical physics最新文献

英文 中文
Optimization of sparse-view CT reconstruction based on convolutional neural network.
Pub Date : 2025-02-02 DOI: 10.1002/mp.17636
Liangliang Lv, Chang Li, Wenjing Wei, Shuyi Sun, Xiaoxuan Ren, Xiaodong Pan, Gongping Li

Background: Sparse-view CT shortens scan time and reduces radiation dose but results in severe streak artifacts due to insufficient sampling data. Deep learning methods can now suppress these artifacts and improve image quality in sparse-view CT reconstruction.

Purpose: The quality of sparse-view CT reconstructed images can still be improved. Additionally, the interpretability of deep learning-based optimization methods for these reconstruction images is lacking, and the role of different network layers in artifact removal requires further study. Moreover, the optimization capability of these methods for reconstruction images from various sparse views needs enhancement. This study aims to improve the network's optimization ability for sparse-view reconstructed images, enhance interpretability, and boost generalization by establishing multiple network structures and datasets.

Methods: In this paper, we developed a sparse-view CT reconstruction images improvement network (SRII-Net) based on U-Net. We added a copy pathway in the network and designed a residual image output block to boost the network's performance. Multiple networks with different connectivity structures were established using SRII-Net to analyze the contribution of each layer to artifact removal, improving the network's interpretability. Additionally, we created multiple datasets with reconstructed images of various sampling views to train and test the proposed network, investigating how these datasets from different sampling views affect the network's generalization ability.

Results: The results show that the proposed method outperforms current networks, with significant improvements in metrics like PSNR and SSIM. Image optimization time is at the millisecond level. By comparing the performance of different network structures, we've identified the impact of various hierarchical structures. The image detail information learned by shallow layers and the high-level abstract feature information learned by deep layers play a crucial role in optimizing sparse-view CT reconstruction images. Training the network with multiple mixed datasets revealed that, under a certain amount of data, selecting the appropriate categories of sampling views and their corresponding samples can effectively enhance the network's optimization ability for reconstructing images with different sampling views.

Conclusions: The network in this paper effectively suppresses artifacts in reconstructed images with different sparse views, improving generalization. We have also created diverse network structures and datasets to deepen the understanding of artifact removal in deep learning networks, offering insights for noise reduction and image enhancement in other imaging methods.

{"title":"Optimization of sparse-view CT reconstruction based on convolutional neural network.","authors":"Liangliang Lv, Chang Li, Wenjing Wei, Shuyi Sun, Xiaoxuan Ren, Xiaodong Pan, Gongping Li","doi":"10.1002/mp.17636","DOIUrl":"https://doi.org/10.1002/mp.17636","url":null,"abstract":"<p><strong>Background: </strong>Sparse-view CT shortens scan time and reduces radiation dose but results in severe streak artifacts due to insufficient sampling data. Deep learning methods can now suppress these artifacts and improve image quality in sparse-view CT reconstruction.</p><p><strong>Purpose: </strong>The quality of sparse-view CT reconstructed images can still be improved. Additionally, the interpretability of deep learning-based optimization methods for these reconstruction images is lacking, and the role of different network layers in artifact removal requires further study. Moreover, the optimization capability of these methods for reconstruction images from various sparse views needs enhancement. This study aims to improve the network's optimization ability for sparse-view reconstructed images, enhance interpretability, and boost generalization by establishing multiple network structures and datasets.</p><p><strong>Methods: </strong>In this paper, we developed a sparse-view CT reconstruction images improvement network (SRII-Net) based on U-Net. We added a copy pathway in the network and designed a residual image output block to boost the network's performance. Multiple networks with different connectivity structures were established using SRII-Net to analyze the contribution of each layer to artifact removal, improving the network's interpretability. Additionally, we created multiple datasets with reconstructed images of various sampling views to train and test the proposed network, investigating how these datasets from different sampling views affect the network's generalization ability.</p><p><strong>Results: </strong>The results show that the proposed method outperforms current networks, with significant improvements in metrics like PSNR and SSIM. Image optimization time is at the millisecond level. By comparing the performance of different network structures, we've identified the impact of various hierarchical structures. The image detail information learned by shallow layers and the high-level abstract feature information learned by deep layers play a crucial role in optimizing sparse-view CT reconstruction images. Training the network with multiple mixed datasets revealed that, under a certain amount of data, selecting the appropriate categories of sampling views and their corresponding samples can effectively enhance the network's optimization ability for reconstructing images with different sampling views.</p><p><strong>Conclusions: </strong>The network in this paper effectively suppresses artifacts in reconstructed images with different sparse views, improving generalization. We have also created diverse network structures and datasets to deepen the understanding of artifact removal in deep learning networks, offering insights for noise reduction and image enhancement in other imaging methods.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143082881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automating the optimization of proton PBS treatment planning for head and neck cancers using policy gradient-based deep reinforcement learning.
Pub Date : 2025-01-31 DOI: 10.1002/mp.17654
Qingqing Wang, Chang Chang

Background: Proton pencil beam scanning (PBS) treatment planning for head and neck (H&N) cancers is a time-consuming and experience-demanding task where a large number of potentially conflicting planning objectives are involved. Deep reinforcement learning (DRL) has recently been introduced to the planning processes of intensity-modulated radiation therapy (IMRT) and brachytherapy for prostate, lung, and cervical cancers. However, existing DRL planning models are built upon the Q-learning framework and rely on weighted linear combinations of clinical metrics for reward calculation. These approaches suffer from poor scalability and flexibility, that is, they are only capable of adjusting a limited number of planning objectives in discrete action spaces and therefore fail to generalize to more complex planning problems.

Purpose: Here we propose an automatic treatment planning model using the proximal policy optimization (PPO) algorithm in the policy gradient framework of DRL and a dose distribution-based reward function for proton PBS treatment planning of H&N cancers.

Methods: The planning process is formulated as an optimization problem. A set of empirical rules is used to create auxiliary planning structures from target volumes and organs-at-risk (OARs), along with their associated planning objectives. Special attention is given to overlapping structures with potentially conflicting objectives. These planning objectives are fed into an in-house optimization engine to generate the spot monitor unit (MU) values. A decision-making policy network trained using PPO is developed to iteratively adjust the involved planning objective parameters. The policy network predicts actions in a continuous action space and guides the treatment planning system to refine the PBS treatment plans using a novel dose distribution-based reward function. A total of 34 H&N patients (30 for training and 4 for test) and 26 liver patients (20 for training, 6 for test) are included in this study to train and verify the effectiveness and generalizability of the proposed method.

Results: Proton H&N treatment plans generated by the model show improved OAR sparing with equal or superior target coverage when compared with human-generated plans. Moreover, additional experiments on liver cancer demonstrate that the proposed method can be successfully generalized to other treatment sites.

Conclusions: The automatic treatment planning model can generate complex H&N plans with quality comparable or superior to those produced by experienced human planners. Compared with existing works, our method is capable of handling more planning objectives in continuous action spaces. To the best of our knowledge, this is the first DRL-based automatic treatment planning model capable of achieving human-level performance for H&N cancers.

{"title":"Automating the optimization of proton PBS treatment planning for head and neck cancers using policy gradient-based deep reinforcement learning.","authors":"Qingqing Wang, Chang Chang","doi":"10.1002/mp.17654","DOIUrl":"https://doi.org/10.1002/mp.17654","url":null,"abstract":"<p><strong>Background: </strong>Proton pencil beam scanning (PBS) treatment planning for head and neck (H&N) cancers is a time-consuming and experience-demanding task where a large number of potentially conflicting planning objectives are involved. Deep reinforcement learning (DRL) has recently been introduced to the planning processes of intensity-modulated radiation therapy (IMRT) and brachytherapy for prostate, lung, and cervical cancers. However, existing DRL planning models are built upon the Q-learning framework and rely on weighted linear combinations of clinical metrics for reward calculation. These approaches suffer from poor scalability and flexibility, that is, they are only capable of adjusting a limited number of planning objectives in discrete action spaces and therefore fail to generalize to more complex planning problems.</p><p><strong>Purpose: </strong>Here we propose an automatic treatment planning model using the proximal policy optimization (PPO) algorithm in the policy gradient framework of DRL and a dose distribution-based reward function for proton PBS treatment planning of H&N cancers.</p><p><strong>Methods: </strong>The planning process is formulated as an optimization problem. A set of empirical rules is used to create auxiliary planning structures from target volumes and organs-at-risk (OARs), along with their associated planning objectives. Special attention is given to overlapping structures with potentially conflicting objectives. These planning objectives are fed into an in-house optimization engine to generate the spot monitor unit (MU) values. A decision-making policy network trained using PPO is developed to iteratively adjust the involved planning objective parameters. The policy network predicts actions in a continuous action space and guides the treatment planning system to refine the PBS treatment plans using a novel dose distribution-based reward function. A total of 34 H&N patients (30 for training and 4 for test) and 26 liver patients (20 for training, 6 for test) are included in this study to train and verify the effectiveness and generalizability of the proposed method.</p><p><strong>Results: </strong>Proton H&N treatment plans generated by the model show improved OAR sparing with equal or superior target coverage when compared with human-generated plans. Moreover, additional experiments on liver cancer demonstrate that the proposed method can be successfully generalized to other treatment sites.</p><p><strong>Conclusions: </strong>The automatic treatment planning model can generate complex H&N plans with quality comparable or superior to those produced by experienced human planners. Compared with existing works, our method is capable of handling more planning objectives in continuous action spaces. To the best of our knowledge, this is the first DRL-based automatic treatment planning model capable of achieving human-level performance for H&N cancers.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143069544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterization of an MR-compatible motion platform for quality assurance of motion-compensated treatments on the 1.5 T MR-linac.
Pub Date : 2025-01-31 DOI: 10.1002/mp.17632
Stijn Oolbekkink, Pim T S Borman, Jochem W H Wolthaus, Bram van Asselen, Astrid L H M W van Lier, Stephanie Dunn, Grant R Koenig, Nick Hartman, Niusha Kheirkhah, Bas W Raaymakers, Martin F Fast

Background: Novel motion-compensated treatment techniques on the MR-linac can address adverse intra-fraction motion effects. These techniques involve beam gating or intra-fraction adaptations of the treatment plan based on real-time magnetic resonance imaging (MRI) performed during treatment. For quality assurance (QA) of these workflows, a multi-purpose motion platform is desirable. This platform should accommodate various phantoms, enabling multiple QA workflows.

Purpose: This study aims to evaluate the new IBA QUASAR Motion MR Platform for use in the 1.5 T MR-linac.

Methods: The motion platform was assessed for several magnetic resonance (MR) characteristics, including spurious noise generation and B0&B1 homogeneity. In addition, the motion platform's motion accuracy and beam attenuation were assessed. An application was shown with a ScandiDos Delta4 Phantom+ MR demonstrating patient-specific plan QA of gated treatments using time-resolved dosimetry that includes motion based on a patient's respiratory motion trace.

Results: All MR characterization measurements were within the set tolerances for MRI QA. The motion platform motion accuracy showed excellent agreement with the reference, with a standard deviation of the amplitude of 0.01  mm (20 kg load) for the motor's self-estimated positions and 0.22 mm (no load) for the images acquired with the electronic portal imager. Beam attenuation was found to be 11.8%. The combination of the motion platform and Delta4 demonstrated motion-included dosimetry at high temporal and spatial resolutions. Motion influenced the measured dose in non-gated treatments by up to -20.1%, while gated deliveries showed differences of up to -1.7% for selected diodes.

Conclusion: The motion platform was found to be usable in a 1.5 T magnetic field, and for all MR characterization experiments, no influence from the motion platform was observed. This motion platform enables to perform motion-included QA, with a measurement device of choice.

{"title":"Characterization of an MR-compatible motion platform for quality assurance of motion-compensated treatments on the 1.5 T MR-linac.","authors":"Stijn Oolbekkink, Pim T S Borman, Jochem W H Wolthaus, Bram van Asselen, Astrid L H M W van Lier, Stephanie Dunn, Grant R Koenig, Nick Hartman, Niusha Kheirkhah, Bas W Raaymakers, Martin F Fast","doi":"10.1002/mp.17632","DOIUrl":"https://doi.org/10.1002/mp.17632","url":null,"abstract":"<p><strong>Background: </strong>Novel motion-compensated treatment techniques on the MR-linac can address adverse intra-fraction motion effects. These techniques involve beam gating or intra-fraction adaptations of the treatment plan based on real-time magnetic resonance imaging (MRI) performed during treatment. For quality assurance (QA) of these workflows, a multi-purpose motion platform is desirable. This platform should accommodate various phantoms, enabling multiple QA workflows.</p><p><strong>Purpose: </strong>This study aims to evaluate the new IBA QUASAR Motion MR Platform for use in the 1.5 T MR-linac.</p><p><strong>Methods: </strong>The motion platform was assessed for several magnetic resonance (MR) characteristics, including spurious noise generation and B0&B1 homogeneity. In addition, the motion platform's motion accuracy and beam attenuation were assessed. An application was shown with a ScandiDos Delta4 Phantom+ MR demonstrating patient-specific plan QA of gated treatments using time-resolved dosimetry that includes motion based on a patient's respiratory motion trace.</p><p><strong>Results: </strong>All MR characterization measurements were within the set tolerances for MRI QA. The motion platform motion accuracy showed excellent agreement with the reference, with a standard deviation of the amplitude of 0.01  mm (20 kg load) for the motor's self-estimated positions and 0.22 mm (no load) for the images acquired with the electronic portal imager. Beam attenuation was found to be 11.8%. The combination of the motion platform and Delta4 demonstrated motion-included dosimetry at high temporal and spatial resolutions. Motion influenced the measured dose in non-gated treatments by up to -20.1%, while gated deliveries showed differences of up to -1.7% for selected diodes.</p><p><strong>Conclusion: </strong>The motion platform was found to be usable in a 1.5 T magnetic field, and for all MR characterization experiments, no influence from the motion platform was observed. This motion platform enables to perform motion-included QA, with a measurement device of choice.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143069831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TransAnaNet: Transformer-based anatomy change prediction network for head and neck cancer radiotherapy.
Pub Date : 2025-01-31 DOI: 10.1002/mp.17655
Meixu Chen, Kai Wang, Michael Dohopolski, Howard Morgan, David Sher, Jing Wang
<p><strong>Background: </strong>Adaptive radiotherapy (ART) can compensate for the dosimetric impact of anatomic change during radiotherapy of head-neck cancer (HNC) patients. However, implementing ART universally poses challenges in clinical workflow and resource allocation, given the variability in patient response and the constraints of available resources. Therefore, the prediction of anatomical change during radiotherapy for HNC patients is of importance to optimize patient clinical benefit and treatment resources. Current studies focus on developing binary ART eligibility classification models to identify patients who would experience significant anatomical change, but these models lack the ability to present the complex patterns and variations in anatomical changes over time. Vision Transformers (ViTs) represent a recent advancement in neural network architectures, utilizing self-attention mechanisms to process image data. Unlike traditional Convolutional Neural Networks (CNNs), ViTs can capture global contextual information more effectively, making them well-suited for image analysis and image generation tasks that involve complex patterns and structures, such as predicting anatomical changes in medical imaging.</p><p><strong>Purpose: </strong>The purpose of this study is to assess the feasibility of using a ViT-based neural network to predict radiotherapy-induced anatomic change of HNC patients.</p><p><strong>Methods: </strong>We retrospectively included 121 HNC patients treated with definitive chemoradiotherapy (CRT) or radiation alone. We collected the planning computed tomography image (pCT), planned dose, cone beam computed tomography images (CBCTs) acquired at the initial treatment (CBCT01) and Fraction 21 (CBCT21), and primary tumor volume (GTVp) and involved nodal volume (GTVn) delineated on both pCT and CBCTs of each patient for model construction and evaluation. A UNet-style Swin-Transformer-based ViT network was designed to learn the spatial correspondence and contextual information from embedded image patches of CT, dose, CBCT01, GTVp, and GTVn. The deformation vector field between CBCT01 and CBCT21 was estimated by the model as the prediction of anatomic change, and deformed CBCT01 was used as the prediction of CBCT21. We also generated binary masks of GTVp, GTVn, and patient body for volumetric change evaluation. We used data from 101 patients for training and validation, and the remaining 20 patients for testing. Image and volumetric similarity metrics including mean square error (MSE), peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), Dice coefficient, and average surface distance were used to measure the similarity between the target image and predicted CBCT. Anatomy change prediction performance of the proposed model was compared to a CNN-based prediction model and a traditional ViT-based prediction model.</p><p><strong>Results: </strong>The predicted image from the proposed method yielded the best s
{"title":"TransAnaNet: Transformer-based anatomy change prediction network for head and neck cancer radiotherapy.","authors":"Meixu Chen, Kai Wang, Michael Dohopolski, Howard Morgan, David Sher, Jing Wang","doi":"10.1002/mp.17655","DOIUrl":"10.1002/mp.17655","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Adaptive radiotherapy (ART) can compensate for the dosimetric impact of anatomic change during radiotherapy of head-neck cancer (HNC) patients. However, implementing ART universally poses challenges in clinical workflow and resource allocation, given the variability in patient response and the constraints of available resources. Therefore, the prediction of anatomical change during radiotherapy for HNC patients is of importance to optimize patient clinical benefit and treatment resources. Current studies focus on developing binary ART eligibility classification models to identify patients who would experience significant anatomical change, but these models lack the ability to present the complex patterns and variations in anatomical changes over time. Vision Transformers (ViTs) represent a recent advancement in neural network architectures, utilizing self-attention mechanisms to process image data. Unlike traditional Convolutional Neural Networks (CNNs), ViTs can capture global contextual information more effectively, making them well-suited for image analysis and image generation tasks that involve complex patterns and structures, such as predicting anatomical changes in medical imaging.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Purpose: &lt;/strong&gt;The purpose of this study is to assess the feasibility of using a ViT-based neural network to predict radiotherapy-induced anatomic change of HNC patients.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;We retrospectively included 121 HNC patients treated with definitive chemoradiotherapy (CRT) or radiation alone. We collected the planning computed tomography image (pCT), planned dose, cone beam computed tomography images (CBCTs) acquired at the initial treatment (CBCT01) and Fraction 21 (CBCT21), and primary tumor volume (GTVp) and involved nodal volume (GTVn) delineated on both pCT and CBCTs of each patient for model construction and evaluation. A UNet-style Swin-Transformer-based ViT network was designed to learn the spatial correspondence and contextual information from embedded image patches of CT, dose, CBCT01, GTVp, and GTVn. The deformation vector field between CBCT01 and CBCT21 was estimated by the model as the prediction of anatomic change, and deformed CBCT01 was used as the prediction of CBCT21. We also generated binary masks of GTVp, GTVn, and patient body for volumetric change evaluation. We used data from 101 patients for training and validation, and the remaining 20 patients for testing. Image and volumetric similarity metrics including mean square error (MSE), peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), Dice coefficient, and average surface distance were used to measure the similarity between the target image and predicted CBCT. Anatomy change prediction performance of the proposed model was compared to a CNN-based prediction model and a traditional ViT-based prediction model.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;The predicted image from the proposed method yielded the best s","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143070358","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid transformer-based model for mammogram classification by integrating prior and current images.
Pub Date : 2025-01-30 DOI: 10.1002/mp.17650
Afsana Ahsan Jeny, Sahand Hamzehei, Annie Jin, Stephen Andrew Baker, Tucker Van Rathe, Jun Bai, Clifford Yang, Sheida Nabavi

Background: Breast cancer screening via mammography plays a crucial role in early detection, significantly impacting women's health outcomes worldwide. However, the manual analysis of mammographic images is time-consuming and requires specialized expertise, presenting substantial challenges in medical practice.

Purpose: To address these challenges, we introduce a CNN-Transformer based model tailored for breast cancer classification through mammographic analysis. This model leverages both prior and current images to monitor temporal changes, aiming to enhance the efficiency and accuracy (ACC) of computer-aided diagnosis systems by mimicking the detailed examination process of radiologists.

Methods: In this study, our proposed model incorporates a novel integration of a position-wise feedforward network and multi-head self-attention, enabling it to detect abnormal or cancerous changes in mammograms over time. Additionally, the model employs positional encoding and channel attention methods to accurately highlight critical spatial features, thus precisely differentiating between normal and cancerous tissues. Our methodology utilizes focal loss (FL) to precisely address challenging instances that are difficult to classify, reducing false negatives and false positives to improve diagnostic ACC.

Results: We compared our model with eight baseline models; specifically, we utilized only current images for the single model ResNet50 while employing both prior and current images for the remaining models in terms of accuracy (ACC), sensitivity (SEN), precision (PRE), specificity (SPE), F1 score, and area under the curve (AUC). The results demonstrate that the proposed model outperforms the baseline models, achieving an ACC of 90.80%, SEN of 90.80%, PRE of 90.80%, SPE of 90.88%, an F1 score of 90.95%, and an AUC of 92.58%. The codes and related information are available at https://github.com/NabaviLab/PCTM.

Conclusions: Our proposed CNN-Transformer model integrates both prior and current images, removes long-range dependencies, and enhances its capability for nuanced classification. The application of FL reduces false positive rate (FPR) and false negative rates (FNR), improving both SEN and SPE. Furthermore, the model achieves the lowest false discovery rate and FNR across various abnormalities, including masses, calcification, and architectural distortions (ADs). These low error rates highlight the model's reliability and underscore its potential to improve early breast cancer detection in clinical practice.

{"title":"Hybrid transformer-based model for mammogram classification by integrating prior and current images.","authors":"Afsana Ahsan Jeny, Sahand Hamzehei, Annie Jin, Stephen Andrew Baker, Tucker Van Rathe, Jun Bai, Clifford Yang, Sheida Nabavi","doi":"10.1002/mp.17650","DOIUrl":"https://doi.org/10.1002/mp.17650","url":null,"abstract":"<p><strong>Background: </strong>Breast cancer screening via mammography plays a crucial role in early detection, significantly impacting women's health outcomes worldwide. However, the manual analysis of mammographic images is time-consuming and requires specialized expertise, presenting substantial challenges in medical practice.</p><p><strong>Purpose: </strong>To address these challenges, we introduce a CNN-Transformer based model tailored for breast cancer classification through mammographic analysis. This model leverages both prior and current images to monitor temporal changes, aiming to enhance the efficiency and accuracy (ACC) of computer-aided diagnosis systems by mimicking the detailed examination process of radiologists.</p><p><strong>Methods: </strong>In this study, our proposed model incorporates a novel integration of a position-wise feedforward network and multi-head self-attention, enabling it to detect abnormal or cancerous changes in mammograms over time. Additionally, the model employs positional encoding and channel attention methods to accurately highlight critical spatial features, thus precisely differentiating between normal and cancerous tissues. Our methodology utilizes focal loss (FL) to precisely address challenging instances that are difficult to classify, reducing false negatives and false positives to improve diagnostic ACC.</p><p><strong>Results: </strong>We compared our model with eight baseline models; specifically, we utilized only current images for the single model ResNet50 while employing both prior and current images for the remaining models in terms of accuracy (ACC), sensitivity (SEN), precision (PRE), specificity (SPE), F1 score, and area under the curve (AUC). The results demonstrate that the proposed model outperforms the baseline models, achieving an ACC of 90.80%, SEN of 90.80%, PRE of 90.80%, SPE of 90.88%, an F1 score of 90.95%, and an AUC of 92.58%. The codes and related information are available at https://github.com/NabaviLab/PCTM.</p><p><strong>Conclusions: </strong>Our proposed CNN-Transformer model integrates both prior and current images, removes long-range dependencies, and enhances its capability for nuanced classification. The application of FL reduces false positive rate (FPR) and false negative rates (FNR), improving both SEN and SPE. Furthermore, the model achieves the lowest false discovery rate and FNR across various abnormalities, including masses, calcification, and architectural distortions (ADs). These low error rates highlight the model's reliability and underscore its potential to improve early breast cancer detection in clinical practice.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143070147","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multilevel perception boundary-guided network for breast lesion segmentation in ultrasound images.
Pub Date : 2025-01-30 DOI: 10.1002/mp.17647
Xing Yang, Jian Zhang, Yingfeng Ou, Qijian Chen, Li Wang, Lihui Wang
<p><strong>Background: </strong>Automatic segmentation of breast tumors from the ultrasound images is essential for the subsequent clinical diagnosis and treatment plan. Although the existing deep learning-based methods have achieved considerable progress in automatic segmentation of breast tumors, their performance on tumors with similar intensity to the normal tissues is still not satisfactory, especially for the tumor boundaries.</p><p><strong>Purpose: </strong>To accurately segment the non-enhanced lesions with more accurate boundaries, a novel multilevel perception boundary-guided network (PBNet) is proposed to segment breast tumors from ultrasound images.</p><p><strong>Methods: </strong>PBNet consists of a multilevel global perception module (MGPM) and a boundary guided module (BGM). MGPM models long-range spatial dependencies by fusing both intra- and inter-level semantic information to enhance tumor recognition. In BGM, the tumor boundaries are extracted from the high-level semantic maps using the dilation and erosion effects of max pooling; such boundaries are then used to guide the fusion of low- and high-level features. Additionally, a multi-level boundary-enhanced segmentation (BS) loss is introduced to improve boundary segmentation performance. To evaluate the effectiveness of the proposed method, we compared it with state-of-the-art methods on two datasets, one publicly available datasets BUSI containing 780 images and one in-house dataset containing 995 images. To verify the robustness of each method, a 5-fold cross-validation method was used to train and test the models. Dice score (Dice), Jaccard coefficients (Jac), Hausdorff Distance (HD), Sensitivity (Sen), and specificity(Spe) were used to evaluate the segmentation performance quantitatively. The Wilcoxon test and Benjamini-Hochberg false discovery rate (FDR) multi-comparison correction were then performed to assess whether the proposed method presents statistically significant performance ( <math> <semantics><mrow><mi>p</mi> <mo>≤</mo> <mn>0.05</mn></mrow> <annotation>$ple 0.05$</annotation></semantics> </math> ) difference comparing with existing methods. In addition, to comprehensively demonstrate the difference between different methods, the Cohen's d effect size and compound p-value (c-Pvalue) obtained with Fisher's method were also calculated.</p><p><strong>Results: </strong>On the BUSI dataset, the mean Dice and Sen of PBNet was increased by 0.93% ( <math> <semantics><mrow><mi>p</mi> <mo>≤</mo> <mn>0.01</mn></mrow> <annotation>$ple 0.01$</annotation></semantics> </math> ) and 1.42% ( <math> <semantics><mrow><mi>p</mi> <mo>≤</mo> <mn>0.05</mn></mrow> <annotation>$ple 0.05$</annotation></semantics> </math> ), respectively, comparing against the corresponding suboptimal methods. On the in-house dataset, PBNet improved Dice, Jac and Spe by approximately 0.86% ( <math> <semantics><mrow><mi>p</mi> <mo>≤</mo> <mn>0.01</mn></mrow> <annotation>$ple 0.01$</annotation></semantics>
{"title":"Multilevel perception boundary-guided network for breast lesion segmentation in ultrasound images.","authors":"Xing Yang, Jian Zhang, Yingfeng Ou, Qijian Chen, Li Wang, Lihui Wang","doi":"10.1002/mp.17647","DOIUrl":"https://doi.org/10.1002/mp.17647","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Automatic segmentation of breast tumors from the ultrasound images is essential for the subsequent clinical diagnosis and treatment plan. Although the existing deep learning-based methods have achieved considerable progress in automatic segmentation of breast tumors, their performance on tumors with similar intensity to the normal tissues is still not satisfactory, especially for the tumor boundaries.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Purpose: &lt;/strong&gt;To accurately segment the non-enhanced lesions with more accurate boundaries, a novel multilevel perception boundary-guided network (PBNet) is proposed to segment breast tumors from ultrasound images.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;PBNet consists of a multilevel global perception module (MGPM) and a boundary guided module (BGM). MGPM models long-range spatial dependencies by fusing both intra- and inter-level semantic information to enhance tumor recognition. In BGM, the tumor boundaries are extracted from the high-level semantic maps using the dilation and erosion effects of max pooling; such boundaries are then used to guide the fusion of low- and high-level features. Additionally, a multi-level boundary-enhanced segmentation (BS) loss is introduced to improve boundary segmentation performance. To evaluate the effectiveness of the proposed method, we compared it with state-of-the-art methods on two datasets, one publicly available datasets BUSI containing 780 images and one in-house dataset containing 995 images. To verify the robustness of each method, a 5-fold cross-validation method was used to train and test the models. Dice score (Dice), Jaccard coefficients (Jac), Hausdorff Distance (HD), Sensitivity (Sen), and specificity(Spe) were used to evaluate the segmentation performance quantitatively. The Wilcoxon test and Benjamini-Hochberg false discovery rate (FDR) multi-comparison correction were then performed to assess whether the proposed method presents statistically significant performance ( &lt;math&gt; &lt;semantics&gt;&lt;mrow&gt;&lt;mi&gt;p&lt;/mi&gt; &lt;mo&gt;≤&lt;/mo&gt; &lt;mn&gt;0.05&lt;/mn&gt;&lt;/mrow&gt; &lt;annotation&gt;$ple 0.05$&lt;/annotation&gt;&lt;/semantics&gt; &lt;/math&gt; ) difference comparing with existing methods. In addition, to comprehensively demonstrate the difference between different methods, the Cohen's d effect size and compound p-value (c-Pvalue) obtained with Fisher's method were also calculated.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;On the BUSI dataset, the mean Dice and Sen of PBNet was increased by 0.93% ( &lt;math&gt; &lt;semantics&gt;&lt;mrow&gt;&lt;mi&gt;p&lt;/mi&gt; &lt;mo&gt;≤&lt;/mo&gt; &lt;mn&gt;0.01&lt;/mn&gt;&lt;/mrow&gt; &lt;annotation&gt;$ple 0.01$&lt;/annotation&gt;&lt;/semantics&gt; &lt;/math&gt; ) and 1.42% ( &lt;math&gt; &lt;semantics&gt;&lt;mrow&gt;&lt;mi&gt;p&lt;/mi&gt; &lt;mo&gt;≤&lt;/mo&gt; &lt;mn&gt;0.05&lt;/mn&gt;&lt;/mrow&gt; &lt;annotation&gt;$ple 0.05$&lt;/annotation&gt;&lt;/semantics&gt; &lt;/math&gt; ), respectively, comparing against the corresponding suboptimal methods. On the in-house dataset, PBNet improved Dice, Jac and Spe by approximately 0.86% ( &lt;math&gt; &lt;semantics&gt;&lt;mrow&gt;&lt;mi&gt;p&lt;/mi&gt; &lt;mo&gt;≤&lt;/mo&gt; &lt;mn&gt;0.01&lt;/mn&gt;&lt;/mrow&gt; &lt;annotation&gt;$ple 0.01$&lt;/annotation&gt;&lt;/semantics&gt;","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143070353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Impact of deep learning reconstructions on image quality and liver lesion detectability in dual-energy CT: An anthropomorphic phantom study.
Pub Date : 2025-01-30 DOI: 10.1002/mp.17651
Aurélie Pauthe, Milan Milliner, Hugo Pasquier, Lucie Campagnolo, Sébastien Mulé, Alain Luciani
<p><strong>Background: </strong>Deep learning image reconstruction (DLIR) algorithms allow strong noise reduction while preserving noise texture, which may potentially improve hypervascular focal liver lesions.</p><p><strong>Purpose: </strong>To assess the impact of DLIR on image quality (IQ) and detectability of simulated hypervascular hepatocellular carcinoma (HCC) in fast kV-switching dual-energy CT (DECT).</p><p><strong>Methods: </strong>An anthropomorphic phantom of a standard patient morphology (body mass index of 23 kg m<sup>-2</sup>) with customized liver, including mimickers of hypervascular lesions in both late arterial phase (AP) and portal venous phase (PVP) enhancement, was scanned on a DECT. Virtual monoenergetic images were reconstructed from raw data at four energy levels (40/50/60/70 keV) using filtered back-projection (FBP), adaptive statistical iterative reconstruction-V 50% and 100% (ASIRV-50 and ASIRV-100), DLIR low (DLIR-L), medium (DLIR-M), and high (DLIR-H). The contrast between the lesion and the liver parenchyma, the noise magnitude, the average and peak frequencies (f<sub>avg</sub> and f<sub>peak</sub>) of the noise power spectrum (NPS) reflecting noise texture, and the task-based measure of the modulation transfer function (MTF<sub>task</sub>) were measured to evaluate spatial resolution. A detectability index (d') was computed to model the detection of hypervascular lesions in both AP and PVP. Metrics were compared between reconstructions and between energy levels using a Friedman test with follow-up post-hoc multiple comparison.</p><p><strong>Results: </strong>Lesion-to-liver contrast significantly increased with decreasing energy level in both AP and PVP (p ≤ 0.042) but was not affected by reconstruction algorithm (p ≥ 0.57). Overall, noise magnitude increased with decreasing energy levels and was the lowest with ASIRV-100 at all energy levels in both AP and PVP (p ≤ 0.01) and significantly lower with DLIR-M and DLIR-H reconstructions compared to ASIRV-50 and DLIR-L (p < 0.001). For all reconstructions, noise texture within the liver tended to get smoother with decreasing energy; f<sub>avg</sub> significantly shifted towards lower frequencies from 70 to 40 keV (p ≤ 0.01). Noise texture was the smoothest with ASIRV-100 (p < 0.001) while DLIR-L had the noise texture closer to the one of FBP. The spatial resolution was not significantly affected by the energy level, but it was degraded when increasing the level of ASIRV and DLIR. For all reconstructions, the detectability indices increased with decreasing energy levels and peaked at 40 and 50 keV in AP and PVP, respectively. In both AP and PVP, the highest d' values were observed with ASIRV-100 and DLIR-H, whatever the energy level studied (p ≤ 0.01) without statistical significance between those two reconstructions.</p><p><strong>Conclusions: </strong>Compared to the routinely used level of iterative reconstruction, DLIR reduces noise without consequential noise textu
{"title":"Impact of deep learning reconstructions on image quality and liver lesion detectability in dual-energy CT: An anthropomorphic phantom study.","authors":"Aurélie Pauthe, Milan Milliner, Hugo Pasquier, Lucie Campagnolo, Sébastien Mulé, Alain Luciani","doi":"10.1002/mp.17651","DOIUrl":"https://doi.org/10.1002/mp.17651","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Deep learning image reconstruction (DLIR) algorithms allow strong noise reduction while preserving noise texture, which may potentially improve hypervascular focal liver lesions.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Purpose: &lt;/strong&gt;To assess the impact of DLIR on image quality (IQ) and detectability of simulated hypervascular hepatocellular carcinoma (HCC) in fast kV-switching dual-energy CT (DECT).&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;An anthropomorphic phantom of a standard patient morphology (body mass index of 23 kg m&lt;sup&gt;-2&lt;/sup&gt;) with customized liver, including mimickers of hypervascular lesions in both late arterial phase (AP) and portal venous phase (PVP) enhancement, was scanned on a DECT. Virtual monoenergetic images were reconstructed from raw data at four energy levels (40/50/60/70 keV) using filtered back-projection (FBP), adaptive statistical iterative reconstruction-V 50% and 100% (ASIRV-50 and ASIRV-100), DLIR low (DLIR-L), medium (DLIR-M), and high (DLIR-H). The contrast between the lesion and the liver parenchyma, the noise magnitude, the average and peak frequencies (f&lt;sub&gt;avg&lt;/sub&gt; and f&lt;sub&gt;peak&lt;/sub&gt;) of the noise power spectrum (NPS) reflecting noise texture, and the task-based measure of the modulation transfer function (MTF&lt;sub&gt;task&lt;/sub&gt;) were measured to evaluate spatial resolution. A detectability index (d') was computed to model the detection of hypervascular lesions in both AP and PVP. Metrics were compared between reconstructions and between energy levels using a Friedman test with follow-up post-hoc multiple comparison.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Lesion-to-liver contrast significantly increased with decreasing energy level in both AP and PVP (p ≤ 0.042) but was not affected by reconstruction algorithm (p ≥ 0.57). Overall, noise magnitude increased with decreasing energy levels and was the lowest with ASIRV-100 at all energy levels in both AP and PVP (p ≤ 0.01) and significantly lower with DLIR-M and DLIR-H reconstructions compared to ASIRV-50 and DLIR-L (p &lt; 0.001). For all reconstructions, noise texture within the liver tended to get smoother with decreasing energy; f&lt;sub&gt;avg&lt;/sub&gt; significantly shifted towards lower frequencies from 70 to 40 keV (p ≤ 0.01). Noise texture was the smoothest with ASIRV-100 (p &lt; 0.001) while DLIR-L had the noise texture closer to the one of FBP. The spatial resolution was not significantly affected by the energy level, but it was degraded when increasing the level of ASIRV and DLIR. For all reconstructions, the detectability indices increased with decreasing energy levels and peaked at 40 and 50 keV in AP and PVP, respectively. In both AP and PVP, the highest d' values were observed with ASIRV-100 and DLIR-H, whatever the energy level studied (p ≤ 0.01) without statistical significance between those two reconstructions.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;Compared to the routinely used level of iterative reconstruction, DLIR reduces noise without consequential noise textu","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143070209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implicit neural representation-based method for metal-induced beam hardening artifact reduction in X-ray CT imaging.
Pub Date : 2025-01-29 DOI: 10.1002/mp.17649
Hyoung Suk Park, Jin Keun Seo, Kiwan Jeon

Background: In X-ray computed tomography (CT), metal-induced beam hardening artifacts arise from the complex interactions between polychromatic X-ray beams and metallic objects, leading to degraded image quality and impeding accurate diagnosis. A previously proposed metal-induced beam hardening correction (MBHC) method provides a theoretical framework for addressing nonlinear artifacts through mathematical analysis, with its effectiveness demonstrated by numerical simulations and phantom experiments. However, in practical applications, this method relies on precise segmentation of highly attenuating materials and parameter estimations, which limit its ability to fully correct artifacts caused by the intricate interactions between metals and other dense materials, such as bone or teeth.

Purpose: This study aims to develop a parameter-free MBHC method that eliminates the need for accurate segmentation and parameter estimations, thereby addressing the limitations of the original MBHC approach.

Methods: The proposed method employs implicit neural representations (INR) to generate two tomographic images: one representing the monochromatic attenuation distribution at a specific energy level, and another capturing the nonlinear beam hardening effects caused by the polychromatic nature of X-ray beams. A loss function drives the generation of these images, where the predicted projection data is nonlinearly modeled by the combination of the two images. This approach eliminates the need for geometric and parameter estimation of metals, providing a more generalized solution.

Results: Numerical and phantom experiments demonstrates that the proposed method effectively reduces beam hardening artifacts caused by interactions between highly attenuating materials such as metals, bone, and teeth. Additionally, the proposed INR-based method demonstrates potential in addressing challenges related to data insufficiencies, such as photon starvation and truncated fields of view in CT imaging.

Conclusions: The proposed generalized MBHC method provides high-quality image reconstructions without requiring parameter estimations and segmentations, offering a robust solution for reducing metal-induced beam hardening artifacts in CT imaging.

{"title":"Implicit neural representation-based method for metal-induced beam hardening artifact reduction in X-ray CT imaging.","authors":"Hyoung Suk Park, Jin Keun Seo, Kiwan Jeon","doi":"10.1002/mp.17649","DOIUrl":"https://doi.org/10.1002/mp.17649","url":null,"abstract":"<p><strong>Background: </strong>In X-ray computed tomography (CT), metal-induced beam hardening artifacts arise from the complex interactions between polychromatic X-ray beams and metallic objects, leading to degraded image quality and impeding accurate diagnosis. A previously proposed metal-induced beam hardening correction (MBHC) method provides a theoretical framework for addressing nonlinear artifacts through mathematical analysis, with its effectiveness demonstrated by numerical simulations and phantom experiments. However, in practical applications, this method relies on precise segmentation of highly attenuating materials and parameter estimations, which limit its ability to fully correct artifacts caused by the intricate interactions between metals and other dense materials, such as bone or teeth.</p><p><strong>Purpose: </strong>This study aims to develop a parameter-free MBHC method that eliminates the need for accurate segmentation and parameter estimations, thereby addressing the limitations of the original MBHC approach.</p><p><strong>Methods: </strong>The proposed method employs implicit neural representations (INR) to generate two tomographic images: one representing the monochromatic attenuation distribution at a specific energy level, and another capturing the nonlinear beam hardening effects caused by the polychromatic nature of X-ray beams. A loss function drives the generation of these images, where the predicted projection data is nonlinearly modeled by the combination of the two images. This approach eliminates the need for geometric and parameter estimation of metals, providing a more generalized solution.</p><p><strong>Results: </strong>Numerical and phantom experiments demonstrates that the proposed method effectively reduces beam hardening artifacts caused by interactions between highly attenuating materials such as metals, bone, and teeth. Additionally, the proposed INR-based method demonstrates potential in addressing challenges related to data insufficiencies, such as photon starvation and truncated fields of view in CT imaging.</p><p><strong>Conclusions: </strong>The proposed generalized MBHC method provides high-quality image reconstructions without requiring parameter estimations and segmentations, offering a robust solution for reducing metal-induced beam hardening artifacts in CT imaging.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143070332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiogenomic explainable AI with neural ordinary differential equation for identifying post-SRS brain metastasis radionecrosis.
Pub Date : 2025-01-29 DOI: 10.1002/mp.17635
Jingtong Zhao, Eugene Vaios, Zhenyu Yang, Ke Lu, Scott Floyd, Deshan Yang, Hangjie Ji, Zachary J Reitman, Kyle J Lafata, Peter Fecci, John P Kirkpatrick, Chunhao Wang
<p><strong>Background: </strong>Stereotactic radiosurgery (SRS) is widely used for managing brain metastases (BMs), but an adverse effect, radionecrosis, complicates post-SRS management. Differentiating radionecrosis from tumor recurrence non-invasively remains a major clinical challenge, as conventional imaging techniques often necessitate surgical biopsy for accurate diagnosis. Machine learning and deep learning models have shown potential in distinguishing radionecrosis from tumor recurrence. However, their clinical adoption is hindered by a lack of explainability, limiting understanding and trust in their diagnostic decisions.</p><p><strong>Purpose: </strong>To utilize a novel neural ordinary differential equation (NODE) model for discerning BM post-SRS radionecrosis from recurrence. This approach integrates image-deep features, genomic biomarkers, and non-image clinical parameters within a synthesized latent feature space. The trajectory of each data sample towards the diagnosis decision can be visualized within this feature space, offering a new angle on radiogenomic data analysis foundational for AI explainability.</p><p><strong>Methods: </strong>By hypothesizing that deep feature extraction can be modeled as a spatiotemporally continuous process, we designed a novel model based on heavy ball NODE (HBNODE) in which deep feature extraction was governed by a second-order ODE. This approach enabled tracking of deep neural network (DNN) behavior by solving the HBNODE and observing the stepwise derivative evolution. Consequently, the trajectory of each sample within the Image-Genomic-Clinical (I-G-C) space became traceable. A decision-making field (F) was reconstructed within the feature space, with its gradient vectors directing the data samples' trajectories and intensities showing the potential. The evolution of F reflected the cumulative feature contributions at intermediate states to the final diagnosis, enabling quantitative and dynamic comparisons of the relative contribution of each feature category over time. A velocity curve was designed to determine key intermediate states (locoregional ∇F = 0) that are most predictive. Subsequently, a non-parametric model aggregated the optimal solutions from these key states to predict outcomes. Our dataset included 90 BMs from 62 NSCLC patients, and 3-month post-SRS T1+c MR image features, seven NSCLC genomic features, and seven clinical features were analyzed. An 8:2 train/test assignment was employed, and five independent models were trained to ensure robustness. Performance was benchmarked in sensitivity, specificity, accuracy, and ROC<sub>AUC</sub>, and results were compared against (1) a DNN using only image-based features, and (2) a combined "I+G+C" features without the HBNODE model.</p><p><strong>Results: </strong>The temporal evolution of gradient vectors and potential fields in F suggested that clinical features contribute the most during the initial stages of the HBNODE implementation, f
{"title":"Radiogenomic explainable AI with neural ordinary differential equation for identifying post-SRS brain metastasis radionecrosis.","authors":"Jingtong Zhao, Eugene Vaios, Zhenyu Yang, Ke Lu, Scott Floyd, Deshan Yang, Hangjie Ji, Zachary J Reitman, Kyle J Lafata, Peter Fecci, John P Kirkpatrick, Chunhao Wang","doi":"10.1002/mp.17635","DOIUrl":"https://doi.org/10.1002/mp.17635","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Background: &lt;/strong&gt;Stereotactic radiosurgery (SRS) is widely used for managing brain metastases (BMs), but an adverse effect, radionecrosis, complicates post-SRS management. Differentiating radionecrosis from tumor recurrence non-invasively remains a major clinical challenge, as conventional imaging techniques often necessitate surgical biopsy for accurate diagnosis. Machine learning and deep learning models have shown potential in distinguishing radionecrosis from tumor recurrence. However, their clinical adoption is hindered by a lack of explainability, limiting understanding and trust in their diagnostic decisions.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Purpose: &lt;/strong&gt;To utilize a novel neural ordinary differential equation (NODE) model for discerning BM post-SRS radionecrosis from recurrence. This approach integrates image-deep features, genomic biomarkers, and non-image clinical parameters within a synthesized latent feature space. The trajectory of each data sample towards the diagnosis decision can be visualized within this feature space, offering a new angle on radiogenomic data analysis foundational for AI explainability.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Methods: &lt;/strong&gt;By hypothesizing that deep feature extraction can be modeled as a spatiotemporally continuous process, we designed a novel model based on heavy ball NODE (HBNODE) in which deep feature extraction was governed by a second-order ODE. This approach enabled tracking of deep neural network (DNN) behavior by solving the HBNODE and observing the stepwise derivative evolution. Consequently, the trajectory of each sample within the Image-Genomic-Clinical (I-G-C) space became traceable. A decision-making field (F) was reconstructed within the feature space, with its gradient vectors directing the data samples' trajectories and intensities showing the potential. The evolution of F reflected the cumulative feature contributions at intermediate states to the final diagnosis, enabling quantitative and dynamic comparisons of the relative contribution of each feature category over time. A velocity curve was designed to determine key intermediate states (locoregional ∇F = 0) that are most predictive. Subsequently, a non-parametric model aggregated the optimal solutions from these key states to predict outcomes. Our dataset included 90 BMs from 62 NSCLC patients, and 3-month post-SRS T1+c MR image features, seven NSCLC genomic features, and seven clinical features were analyzed. An 8:2 train/test assignment was employed, and five independent models were trained to ensure robustness. Performance was benchmarked in sensitivity, specificity, accuracy, and ROC&lt;sub&gt;AUC&lt;/sub&gt;, and results were compared against (1) a DNN using only image-based features, and (2) a combined \"I+G+C\" features without the HBNODE model.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;The temporal evolution of gradient vectors and potential fields in F suggested that clinical features contribute the most during the initial stages of the HBNODE implementation, f","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143061632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Segmentation of coronary artery and calcification using prior knowledge based deep learning framework.
Pub Date : 2025-01-29 DOI: 10.1002/mp.17642
Jinda Wang, Qian Chen, Xingyu Jiang, Zeyu Zhang, Zhenyu Tang

Background: Computed tomography angiography (CTA) is used to screen for coronary artery calcification. As the coronary artery has complicated structure and tiny lumen, manual screening is a time-consuming task. Recently, many deep learning methods have been proposed for the segmentation (SEG) of coronary artery and calcification, however, they often neglect leveraging related anatomical prior knowledge, resulting in low accuracy and instability.

Purpose: This study aims to build a deep learning based SEG framework, which leverages anatomical prior knowledge of coronary artery and calcification, to improve the SEG accuracy. Moreover, based on the SEG results, this study also try to reveal the predictive ability of the volume ratio of coronary artery and calcification for rotational atherectomy (RA).

Methods: We present a new SEG framework, which is composed of four modules: the variational autoencoder based centerline extraction (CE) module, the self-attention (SA) module, the logic operation (LO) module, and the SEG module. Specifically, the CE module is used to crop a series of 3D CTA patches along the coronary artery, from which the continuous property of vessels can be utilized by the SA module to produce vessel-related features. According to the spatial relations between coronary artery lumen and calcification regions, the LO module with logic union and intersection is designed to refine the vessel-related features into lumen- and calcification-related features, based on which SEG results can be produced by the SEG module.

Results: Experimental results demonstrate that our framework outperforms the state-of-the-art methods on CTA image dataset of 72 patients with statistical significance. Ablation experiments confirm that the proposed modules have positive impacts to the SEG results. Moreover, based on the volume ratio of segmented coronary artery and calcification, the prediction accuracy of RA is 0.75.

Conclusions: Integrating anatomical prior knowledge of coronary artery and calcification into the deep learning based SEG framework can effectively enhance the performance. Moreover, the volume ratio of segmented coronary artery and calcification is a good predictive factor for RA.

{"title":"Segmentation of coronary artery and calcification using prior knowledge based deep learning framework.","authors":"Jinda Wang, Qian Chen, Xingyu Jiang, Zeyu Zhang, Zhenyu Tang","doi":"10.1002/mp.17642","DOIUrl":"https://doi.org/10.1002/mp.17642","url":null,"abstract":"<p><strong>Background: </strong>Computed tomography angiography (CTA) is used to screen for coronary artery calcification. As the coronary artery has complicated structure and tiny lumen, manual screening is a time-consuming task. Recently, many deep learning methods have been proposed for the segmentation (SEG) of coronary artery and calcification, however, they often neglect leveraging related anatomical prior knowledge, resulting in low accuracy and instability.</p><p><strong>Purpose: </strong>This study aims to build a deep learning based SEG framework, which leverages anatomical prior knowledge of coronary artery and calcification, to improve the SEG accuracy. Moreover, based on the SEG results, this study also try to reveal the predictive ability of the volume ratio of coronary artery and calcification for rotational atherectomy (RA).</p><p><strong>Methods: </strong>We present a new SEG framework, which is composed of four modules: the variational autoencoder based centerline extraction (CE) module, the self-attention (SA) module, the logic operation (LO) module, and the SEG module. Specifically, the CE module is used to crop a series of 3D CTA patches along the coronary artery, from which the continuous property of vessels can be utilized by the SA module to produce vessel-related features. According to the spatial relations between coronary artery lumen and calcification regions, the LO module with logic union and intersection is designed to refine the vessel-related features into lumen- and calcification-related features, based on which SEG results can be produced by the SEG module.</p><p><strong>Results: </strong>Experimental results demonstrate that our framework outperforms the state-of-the-art methods on CTA image dataset of 72 patients with statistical significance. Ablation experiments confirm that the proposed modules have positive impacts to the SEG results. Moreover, based on the volume ratio of segmented coronary artery and calcification, the prediction accuracy of RA is 0.75.</p><p><strong>Conclusions: </strong>Integrating anatomical prior knowledge of coronary artery and calcification into the deep learning based SEG framework can effectively enhance the performance. Moreover, the volume ratio of segmented coronary artery and calcification is a good predictive factor for RA.</p>","PeriodicalId":94136,"journal":{"name":"Medical physics","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143061577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Medical physics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1