Pub Date : 2026-01-23DOI: 10.1088/1361-6560/ae35c9
A Smolders, A J Lomax, F Albertini
Objective.Online adaptive proton therapy could benefit from reoptimization that considers the total dose delivered in previous fractions. However, the accumulated dose is uncertain because of deformable image registration (DIR) uncertainties. This work aims to evaluate the accuracy of a tool predicting the dose accumulation reliability of a treatment plan, allowing consideration of this reliability during treatment planning.Approach.A previously developed deep-learning-based DIR uncertainty model was extended to calculate theexpectedDIR uncertainty only from the planning computed tomography (CT) and theexpecteddose accumulation uncertainty by including the planned dose distribution. For 5 lung cancer patients, the expected dose accumulation uncertainty was compared to the uncertainty of the accumulated dose of 9 repeated CTs. The model was then applied to several alternative treatment plans for each patient to evaluate its potential for plan selection.Results.The average accumulated dose uncertainty was close to the expected dose uncertainty for a large range of expected uncertainties. For high expected uncertainties, the model slightly overestimated the uncertainty. For individual voxels, errors up to 5% of the prescribed dose were common, mainly due to the daily dose distribution deviating from the plan and not because of inaccuracies in the expected DIR uncertainty. Despite the voxel-wise inaccuracies, the method proved suitable to select and compare treatment plans with respect to their accumulation reliability.Significance.Using our tool to select reliably accumulatable treatment plans can facilitate the use of accumulated doses during online reoptimization.
{"title":"Predicting dose accumulation reliability at the planning stage, with an application to adaptive proton therapy.","authors":"A Smolders, A J Lomax, F Albertini","doi":"10.1088/1361-6560/ae35c9","DOIUrl":"10.1088/1361-6560/ae35c9","url":null,"abstract":"<p><p><i>Objective.</i>Online adaptive proton therapy could benefit from reoptimization that considers the total dose delivered in previous fractions. However, the accumulated dose is uncertain because of deformable image registration (DIR) uncertainties. This work aims to evaluate the accuracy of a tool predicting the dose accumulation reliability of a treatment plan, allowing consideration of this reliability during treatment planning.<i>Approach.</i>A previously developed deep-learning-based DIR uncertainty model was extended to calculate the<i>expected</i>DIR uncertainty only from the planning computed tomography (CT) and the<i>expected</i>dose accumulation uncertainty by including the planned dose distribution. For 5 lung cancer patients, the expected dose accumulation uncertainty was compared to the uncertainty of the accumulated dose of 9 repeated CTs. The model was then applied to several alternative treatment plans for each patient to evaluate its potential for plan selection.<i>Results.</i>The average accumulated dose uncertainty was close to the expected dose uncertainty for a large range of expected uncertainties. For high expected uncertainties, the model slightly overestimated the uncertainty. For individual voxels, errors up to 5% of the prescribed dose were common, mainly due to the daily dose distribution deviating from the plan and not because of inaccuracies in the expected DIR uncertainty. Despite the voxel-wise inaccuracies, the method proved suitable to select and compare treatment plans with respect to their accumulation reliability.<i>Significance.</i>Using our tool to select reliably accumulatable treatment plans can facilitate the use of accumulated doses during online reoptimization.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145934682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-22DOI: 10.1088/1361-6560/ae3c53
Jingyan Xu, Frédéric Noo
Objective:
We propose a new formulation for ideal observers (IOs) that incorporate stochastic object models (SOMs) for data acquisition optimization.
Approach:
A data acquisition system is considered as a (possibly nonlinear) discrete-to-discrete mapping from a finite-dimensional object space, x∈R^(n_d), to a finite-dimensional measurement space, y∈R^m. For binary tasks, the two underlying SOMs, H_0 and H_1, are specified by two probability density functions (PDFs) p_0 (x), p_1 (x). This leads to the notion of intrinsic likelihood ratio (LR) Λ_I (x)=p_1 (x)/p_0 (x) and intrinsic class separability (ICS), the latter quantifies the population separability that is independent of data acquisition. With respect to ICS, the IO employs the "extrinsic" LR Λ(y)=pr (y|H_1)/pr(y|H_0) of the data and quantifies the extrinsic class separability (ECS). The difference between ICS and ECS measures the efficiency of data acquisition. We show that the extrinsic LR Λ(y) is the expectation of the intrinsic LR Λ_I (x), where the expectation is with respect to the posterior PDF pr(x│y,H_0 ) under H_0.
Main results:
We use two examples, one to clarify the new IO and the second to demonstrate its potential for real world applications. Specifically, we apply the new IO to spectral optimization in dual-energy CT projection domain material decomposition (pMD), for which SOMs are used to describe variability of basis material line integrals. The performance rank orders obtained by IO agree with physics predictions.
Significance:
The main computation in the new IO involves sampling from the posterior PDF pr(x│y,H_0 ), which are similar to (fully) Bayesian reconstruction. Thus our IO computation is amenable to standard techniques already familiar to CT researchers. The example of dual-energy pMD serves as a prototype for other spectral optimization problems, e.g., for photon counting CT or multi-energy CT with multi-layer detectors.
.
{"title":"Ideal observer estimation for binary tasks with stochastic object models.","authors":"Jingyan Xu, Frédéric Noo","doi":"10.1088/1361-6560/ae3c53","DOIUrl":"https://doi.org/10.1088/1361-6560/ae3c53","url":null,"abstract":"<p><strong>Objective: </strong>
We propose a new formulation for ideal observers (IOs) that incorporate stochastic object models (SOMs) for data acquisition optimization.
Approach:
A data acquisition system is considered as a (possibly nonlinear) discrete-to-discrete mapping from a finite-dimensional object space, x∈R^(n_d), to a finite-dimensional measurement space, y∈R^m. For binary tasks, the two underlying SOMs, H_0 and H_1, are specified by two probability density functions (PDFs) p_0 (x), p_1 (x). This leads to the notion of intrinsic likelihood ratio (LR) Λ_I (x)=p_1 (x)/p_0 (x) and intrinsic class separability (ICS), the latter quantifies the population separability that is independent of data acquisition. With respect to ICS, the IO employs the \"extrinsic\" LR Λ(y)=pr (y|H_1)/pr(y|H_0) of the data and quantifies the extrinsic class separability (ECS). The difference between ICS and ECS measures the efficiency of data acquisition. We show that the extrinsic LR Λ(y) is the expectation of the intrinsic LR Λ_I (x), where the expectation is with respect to the posterior PDF pr(x│y,H_0 ) under H_0.
Main results:
We use two examples, one to clarify the new IO and the second to demonstrate its potential for real world applications. Specifically, we apply the new IO to spectral optimization in dual-energy CT projection domain material decomposition (pMD), for which SOMs are used to describe variability of basis material line integrals. The performance rank orders obtained by IO agree with physics predictions.
Significance:
The main computation in the new IO involves sampling from the posterior PDF pr(x│y,H_0 ), which are similar to (fully) Bayesian reconstruction. Thus our IO computation is amenable to standard techniques already familiar to CT researchers. The example of dual-energy pMD serves as a prototype for other spectral optimization problems, e.g., for photon counting CT or multi-energy CT with multi-layer detectors.
.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146030651","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-22DOI: 10.1088/1361-6560/ae3101
Le Yang, Haiyang Zhang, Lei Zheng, Tianfeng Zhang, Duojin Xia, Xuefei Song, Lei Zhou, Huifang Zhou
Objective.To develop an efficient deep learning framework for precise three-dimensional (3D) segmentation of complex orbital structures in multi-sequence magnetic resonance imaging (MRI) and robust assessment of thyroid eye disease (TED) activity, thereby addressing limitations in computational complexity, segmentation accuracy, and integration of multi-sequence features to support clinical decision-making.Approach.We propose RQNet, a U-shaped 3D segmentation network that incorporates the novel Refined Query Transformer Block with refined attention query multi-head self-attention. This design reduces attention complexity fromO(N2)toO(N⋅M)(M≪N) through pooled refined queries. High-quality segmentations then feed into a radiomics pipeline that extracts features per region of interest-including shape, first-order, and texture descriptors. The MRI features from the three sequences-T1-weighted imaging (T1WI), contrast-enhanced T1WI (T1CE), and T2-weighted imaging (T2WI)-are subsequently integrated, with support vector machine, random forest, and logistic regression models employed for assessment to distinguish between active and inactive TED phases.Main results.RQNet achieved Dice similarity coefficients of 83.34%-87.15% on TED datasets (T1WI, T2WI, T1CE), outperforming state-of-the-art models such as nnFormer, UNETR, SwinUNETR, SegResNet, and nnUNet. The radiomics fusion pipeline yielded area under the curve values of 84.65%-85.89% for TED activity assessment, surpassing single-sequence baselines and confirming the benefits of multi-sequence MRI feature fusion enhancements.Significance.The proposed RQNet establishes an efficient segmentation network for 3D orbital MRI, providing accurate depictions of TED structures, robust radiomics-based activity assessment, and enhanced TED assessment through multi-sequence MRI feature integration.
{"title":"Refined query network (RQNet) for precise MRI segmentation and robust TED activity assessment.","authors":"Le Yang, Haiyang Zhang, Lei Zheng, Tianfeng Zhang, Duojin Xia, Xuefei Song, Lei Zhou, Huifang Zhou","doi":"10.1088/1361-6560/ae3101","DOIUrl":"10.1088/1361-6560/ae3101","url":null,"abstract":"<p><p><i>Objective.</i>To develop an efficient deep learning framework for precise three-dimensional (3D) segmentation of complex orbital structures in multi-sequence magnetic resonance imaging (MRI) and robust assessment of thyroid eye disease (TED) activity, thereby addressing limitations in computational complexity, segmentation accuracy, and integration of multi-sequence features to support clinical decision-making.<i>Approach.</i>We propose RQNet, a U-shaped 3D segmentation network that incorporates the novel Refined Query Transformer Block with refined attention query multi-head self-attention. This design reduces attention complexity fromO(N2)toO(N⋅M)(M≪N) through pooled refined queries. High-quality segmentations then feed into a radiomics pipeline that extracts features per region of interest-including shape, first-order, and texture descriptors. The MRI features from the three sequences-T1-weighted imaging (T1WI), contrast-enhanced T1WI (T1CE), and T2-weighted imaging (T2WI)-are subsequently integrated, with support vector machine, random forest, and logistic regression models employed for assessment to distinguish between active and inactive TED phases.<i>Main results.</i>RQNet achieved Dice similarity coefficients of 83.34%-87.15% on TED datasets (T1WI, T2WI, T1CE), outperforming state-of-the-art models such as nnFormer, UNETR, SwinUNETR, SegResNet, and nnUNet. The radiomics fusion pipeline yielded area under the curve values of 84.65%-85.89% for TED activity assessment, surpassing single-sequence baselines and confirming the benefits of multi-sequence MRI feature fusion enhancements.<i>Significance.</i>The proposed RQNet establishes an efficient segmentation network for 3D orbital MRI, providing accurate depictions of TED structures, robust radiomics-based activity assessment, and enhanced TED assessment through multi-sequence MRI feature integration.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145827636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-22DOI: 10.1088/1361-6560/ae3658
Arman Gorji, Nima Sanati, Amir Hossein Pouria, Somayeh Sadat Mehrnia, Ilker Hacihaliloglu, Arman Rahmim, Mohammad R Salmanpour
Objective.Radiomics-based artificial intelligence (AI) models show potential in breast cancer diagnosis but lack interpretability. This study bridges the gap between radiomic features (RFs) and Breast Imaging Reporting and Data System (BI-RADS) descriptors through a clinically interpretable framework.Methods. We developed a dual-dictionary approach. First, a clinical mapping dictionary (CMD) was constructed by mapping 56 RFs to BI-RADS descriptors (shape, margin, internal enhancement (IE)) based on literature and expert review. Second, we applied this framework to a classification task to predict triple-negative (TNBC) versus non-TNBC subtypes using dynamic contrast-enhanced MRI data from a multi-institutional cohort of 1549 patients. We trained 27 machine learning classifiers with 27 feature selection methods. Using SHapley Additive exPlanations (SHAP), we interpreted the model's predictions and developed a Statistical Mapping Dictionary for 51 RFs, not included in the CMD.Results. The best-performing model (variance inflation factor feature selector + extra trees classifier) achieved an average cross-validation accuracy of 0.83 ± 0.02. Our dual-dictionary approach successfully translated predictive RFs into understandable clinical concepts. For example, higher values of 'Sphericity', corresponding to a round/oval shape, were predictive of TNBC. Similarly, lower values of 'Busyness', indicating more homogeneous IE, were also associated with TNBC, aligning with existing clinical observations. This framework confirmed known imaging biomarkers and identified novel, data-driven quantitative features.Conclusion.This study introduces a novel dual-dictionary framework (BM1.0) that bridges RFs and the BI-RADS clinical lexicon. By enhancing the interpretability and transparency of AI models, the framework supports greater clinical trust and paves the way for integrating RFs into breast cancer diagnosis and personalized care.
{"title":"Radiological and biological dictionary of radiomics features: addressing understandable AI issues in personalized breast cancer; dictionary version BM1.0.","authors":"Arman Gorji, Nima Sanati, Amir Hossein Pouria, Somayeh Sadat Mehrnia, Ilker Hacihaliloglu, Arman Rahmim, Mohammad R Salmanpour","doi":"10.1088/1361-6560/ae3658","DOIUrl":"10.1088/1361-6560/ae3658","url":null,"abstract":"<p><p><i>Objective.</i>Radiomics-based artificial intelligence (AI) models show potential in breast cancer diagnosis but lack interpretability. This study bridges the gap between radiomic features (RFs) and Breast Imaging Reporting and Data System (BI-RADS) descriptors through a clinically interpretable framework.<i>Methods</i>. We developed a dual-dictionary approach. First, a clinical mapping dictionary (CMD) was constructed by mapping 56 RFs to BI-RADS descriptors (shape, margin, internal enhancement (IE)) based on literature and expert review. Second, we applied this framework to a classification task to predict triple-negative (TNBC) versus non-TNBC subtypes using dynamic contrast-enhanced MRI data from a multi-institutional cohort of 1549 patients. We trained 27 machine learning classifiers with 27 feature selection methods. Using SHapley Additive exPlanations (SHAP), we interpreted the model's predictions and developed a Statistical Mapping Dictionary for 51 RFs, not included in the CMD.<i>Results</i>. The best-performing model (variance inflation factor feature selector + extra trees classifier) achieved an average cross-validation accuracy of 0.83 ± 0.02. Our dual-dictionary approach successfully translated predictive RFs into understandable clinical concepts. For example, higher values of 'Sphericity', corresponding to a round/oval shape, were predictive of TNBC. Similarly, lower values of 'Busyness', indicating more homogeneous IE, were also associated with TNBC, aligning with existing clinical observations. This framework confirmed known imaging biomarkers and identified novel, data-driven quantitative features.<i>Conclusion.</i>This study introduces a novel dual-dictionary framework (BM1.0) that bridges RFs and the BI-RADS clinical lexicon. By enhancing the interpretability and transparency of AI models, the framework supports greater clinical trust and paves the way for integrating RFs into breast cancer diagnosis and personalized care.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145945663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-21DOI: 10.1088/1361-6560/ae36e4
Harris Hamilton, Daniel Björkman, Antony Lomax, Jan Hrbacek
Purpose.Ocular torsion is a challenge occasionally encountered in ocular proton therapy (OPT) consisting of a rotation of the eye about the visual axis. This can result in the safety margin being compromised and reduced conformity of the dose field to the target. This note investigates the effect of ocular torsion on the lateral margin to verify and explore quantitative adaptation strategies to mitigate the adverse effect on this margin.Methods.OCULARIS, an in-house OPT research planning tool, was used to simulate 14 patients undergoing OPT. The lateral margin was determined for each patient at ocular torsion angles ranging from -8∘to 8∘in discrete steps of 2∘, with 19 collimator rotations simulated at each torsion angle.Results.Margin loss increases with greater ocular torsion, with significant inter-patient variability being influenced by the shape of the target. Aligning collimator rotation with ocular torsion nominal torsion matching (NTM) retains 61% of the margin, patient-specific adaptations achieve superior dose conformity to the target. A simple regression method, setting the collimator rotation to the ocular torsion angle minus 1∘for torsions greater than 2∘, offers some benefit over NTM in this cohort.Conclusions.Margin loss increases with ocular torsion, with the extent of loss being influenced by patient-specific geometry. The NTM collimator rotation strategy was found to adequately compensate for torsion-induced margin loss. Alternative collimator rotation strategies were also explored, including a framework for optimising collimator rotation in the event of ocular torsion.
{"title":"Mitigating ocular torsion induced margin loss in ocular proton therapy via collimator rotation.","authors":"Harris Hamilton, Daniel Björkman, Antony Lomax, Jan Hrbacek","doi":"10.1088/1361-6560/ae36e4","DOIUrl":"10.1088/1361-6560/ae36e4","url":null,"abstract":"<p><p><i>Purpose.</i>Ocular torsion is a challenge occasionally encountered in ocular proton therapy (OPT) consisting of a rotation of the eye about the visual axis. This can result in the safety margin being compromised and reduced conformity of the dose field to the target. This note investigates the effect of ocular torsion on the lateral margin to verify and explore quantitative adaptation strategies to mitigate the adverse effect on this margin.<i>Methods.</i>OCULARIS, an in-house OPT research planning tool, was used to simulate 14 patients undergoing OPT. The lateral margin was determined for each patient at ocular torsion angles ranging from -8<sup>∘</sup>to 8<sup>∘</sup>in discrete steps of 2<sup>∘</sup>, with 19 collimator rotations simulated at each torsion angle.<i>Results.</i>Margin loss increases with greater ocular torsion, with significant inter-patient variability being influenced by the shape of the target. Aligning collimator rotation with ocular torsion nominal torsion matching (NTM) retains 61% of the margin, patient-specific adaptations achieve superior dose conformity to the target. A simple regression method, setting the collimator rotation to the ocular torsion angle minus 1<sup>∘</sup>for torsions greater than 2<sup>∘</sup>, offers some benefit over NTM in this cohort.<i>Conclusions.</i>Margin loss increases with ocular torsion, with the extent of loss being influenced by patient-specific geometry. The NTM collimator rotation strategy was found to adequately compensate for torsion-induced margin loss. Alternative collimator rotation strategies were also explored, including a framework for optimising collimator rotation in the event of ocular torsion.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Objective.Accurate and personalized radiation dose estimation is crucial for effective targeted radionuclide therapy (TRT). Deep learning (DL) holds promise for this purpose. However, current DL-based dosimetry methods require large-scale supervised data, which is scarce in clinical practice.Approach.To address this challenge, we propose exploring semi-supervised learning (SSL) framework that leverages readily available pre-therapy positron emission tomography (PET) data, where only a small subset requires dose labels, to predict radiation doses, thereby reducing the dependency on extensive labeled datasets. In this study, traditional classification-based SSL approaches were adapted and extended in regression task specifically designed for dose prediction. To facilitate comprehensive testing and validation, we developed a synthetic dataset that simulates PET images and dose calculation using Monte Carlo simulations.Main results.In the experiment, several regression-adapted SSL methods were compared and evaluated under varying proportions of labeled data in the training set. The overall mean absolute percentage error of dose prediction remained between 9% and 11% across different organs, which achieved comparable performance than fully supervised ones.Significance.The preliminary experimental results demonstrated that the proposed SSL methods yield promising outcomes for organ-level dose prediction, particularly in scenarios where clinical data are not available in sufficient quantities.
{"title":"Semi-supervised learning for dose prediction in targeted radionuclide therapy: a synthetic data study.","authors":"Jing Zhang, Alexandre Bousse, Chi-Hieu Pham, Kuangyu Shi, Julien Bert","doi":"10.1088/1361-6560/ae36df","DOIUrl":"10.1088/1361-6560/ae36df","url":null,"abstract":"<p><p><i>Objective.</i>Accurate and personalized radiation dose estimation is crucial for effective targeted radionuclide therapy (TRT). Deep learning (DL) holds promise for this purpose. However, current DL-based dosimetry methods require large-scale supervised data, which is scarce in clinical practice.<i>Approach.</i>To address this challenge, we propose exploring semi-supervised learning (SSL) framework that leverages readily available pre-therapy positron emission tomography (PET) data, where only a small subset requires dose labels, to predict radiation doses, thereby reducing the dependency on extensive labeled datasets. In this study, traditional classification-based SSL approaches were adapted and extended in regression task specifically designed for dose prediction. To facilitate comprehensive testing and validation, we developed a synthetic dataset that simulates PET images and dose calculation using Monte Carlo simulations.<i>Main results.</i>In the experiment, several regression-adapted SSL methods were compared and evaluated under varying proportions of labeled data in the training set. The overall mean absolute percentage error of dose prediction remained between 9% and 11% across different organs, which achieved comparable performance than fully supervised ones.<i>Significance.</i>The preliminary experimental results demonstrated that the proposed SSL methods yield promising outcomes for organ-level dose prediction, particularly in scenarios where clinical data are not available in sufficient quantities.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145959765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-21DOI: 10.1088/1361-6560/ae35c5
Gang Lu, Xiangwen Wang, Mangang Xie, Xianghong Lin, Bowu Zhu, Yu Wei, Baoping Zhang, Jiao Du, Fuzhi Wu, Huazhong Shu
Objective. Cephalometric alandmark localization is of great clinical significance in diagnosing and treating patients with dental-maxillofacial deformities. Domain shifts across clinical centers significantly hinder model generalizability, causing existing methods to struggle with accurate and robust anatomical landmark localization due to insufficient alignment of high-level semantic features across domains. We aim to improve the cross-domain generalizability of cephalometric landmark detection by aligning semantic features and enhancing output resolution under an unsupervised domain adaptation (UDA) setting.Approach. In this paper, we propose bi-Level alignment with super-resolution head, an effective framework for precise and robust anatomical landmark detection under UDA. Specifically, we employ adaptive instance normalization to generate target-style images while preserving original anatomical spatial structure at the input level. At the output level, a Mean-Teacher framework leverages high-quality pseudo-labels from the teacher model to guide the student model's learning. Additionally, a lightweight super-resolution head enables the generation of high-resolution heatmaps from the multi-scale feature maps and the low-resolution heatmaps, reducing quantization errors with low computational cost.Results. The proposed method achieved a mean localization error of 1.64 mm, a successful detection rate of 72.68% within the clinically acceptable threshold of 2 mm, and an average classification accuracy of 81.81% for anatomical types.Significance. Extensive experiments on public cephalometric datasets demonstrate superiority over state-of-the-art UDA methods, highlighting its potential for clinical applications in cephalometric analysis and orthodontic surgery planning.
{"title":"Bi-level alignment with super-resolution head for unsupervised cephalometric landmark localization.","authors":"Gang Lu, Xiangwen Wang, Mangang Xie, Xianghong Lin, Bowu Zhu, Yu Wei, Baoping Zhang, Jiao Du, Fuzhi Wu, Huazhong Shu","doi":"10.1088/1361-6560/ae35c5","DOIUrl":"10.1088/1361-6560/ae35c5","url":null,"abstract":"<p><p><i>Objective</i>. Cephalometric alandmark localization is of great clinical significance in diagnosing and treating patients with dental-maxillofacial deformities. Domain shifts across clinical centers significantly hinder model generalizability, causing existing methods to struggle with accurate and robust anatomical landmark localization due to insufficient alignment of high-level semantic features across domains. We aim to improve the cross-domain generalizability of cephalometric landmark detection by aligning semantic features and enhancing output resolution under an unsupervised domain adaptation (UDA) setting.<i>Approach</i>. In this paper, we propose bi-Level alignment with super-resolution head, an effective framework for precise and robust anatomical landmark detection under UDA. Specifically, we employ adaptive instance normalization to generate target-style images while preserving original anatomical spatial structure at the input level. At the output level, a Mean-Teacher framework leverages high-quality pseudo-labels from the teacher model to guide the student model's learning. Additionally, a lightweight super-resolution head enables the generation of high-resolution heatmaps from the multi-scale feature maps and the low-resolution heatmaps, reducing quantization errors with low computational cost.<i>Results</i>. The proposed method achieved a mean localization error of 1.64 mm, a successful detection rate of 72.68% within the clinically acceptable threshold of 2 mm, and an average classification accuracy of 81.81% for anatomical types.<i>Significance</i>. Extensive experiments on public cephalometric datasets demonstrate superiority over state-of-the-art UDA methods, highlighting its potential for clinical applications in cephalometric analysis and orthodontic surgery planning.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145934709","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-21DOI: 10.1088/1361-6560/ae35c6
Zhuoran Jiang, Zhendong Zhang, Lei Xing, Lei Ren, Xianjin Dai
Objective.Unsupervised deep learning has shown great promise in deformable image registration (DIR). These methods update model weights to optimize image similarity without requiring ground truth deformation vector fields (DVFs). However, they inherently face the ill-conditioning challenges due to structural ambiguities. This study aims to address these issues by integrating the implicit anatomical understanding of vision foundation models (FMs) into a multi-scale unsupervised framework for accurate and robust DIR.Approach.Our method takes moving and fixed images as inputs and leverages a pre-trained encoder from a vision FM to extract latent features. These features are merged with those extracted by convolutional adaptors to incorporate inductive bias. Correlation-aware multi-layer perceptrons decode the features into DVFs. A pyramid architecture is implemented to capture multi-range dependencies, further enhancing the DIR robustness and accuracy. We evaluated our method using a multi-modality, cross-institutional database consisting of 150 cardiac cine MR and 40 liver CT.Main results.Our model generates realistic and accurate DVFs. Moving images deformed by our method showed excellent similarity to fixed images, achieving a registration Dice score of 0.869 ± 0.093 for cardiac MRI and an average landmark error of 1.60 ± 1.44 mm for liver CT, substantially surpassing the state-of-the-art methods. Ablation studies further verified the effectiveness of integrating foundation features to improve DIR accuracy (p< 0.05).Significance.Our novel approach demonstrates significant advancements in DIR for multi-modality images with complex structures and low contrasts, making it a powerful tool for a wide range of applications in medical image analysis.
{"title":"Foundation model-enhanced unsupervised 3D deformable medical image registration.","authors":"Zhuoran Jiang, Zhendong Zhang, Lei Xing, Lei Ren, Xianjin Dai","doi":"10.1088/1361-6560/ae35c6","DOIUrl":"10.1088/1361-6560/ae35c6","url":null,"abstract":"<p><p><i>Objective.</i>Unsupervised deep learning has shown great promise in deformable image registration (DIR). These methods update model weights to optimize image similarity without requiring ground truth deformation vector fields (DVFs). However, they inherently face the ill-conditioning challenges due to structural ambiguities. This study aims to address these issues by integrating the implicit anatomical understanding of vision foundation models (FMs) into a multi-scale unsupervised framework for accurate and robust DIR.<i>Approach.</i>Our method takes moving and fixed images as inputs and leverages a pre-trained encoder from a vision FM to extract latent features. These features are merged with those extracted by convolutional adaptors to incorporate inductive bias. Correlation-aware multi-layer perceptrons decode the features into DVFs. A pyramid architecture is implemented to capture multi-range dependencies, further enhancing the DIR robustness and accuracy. We evaluated our method using a multi-modality, cross-institutional database consisting of 150 cardiac cine MR and 40 liver CT.<i>Main results.</i>Our model generates realistic and accurate DVFs. Moving images deformed by our method showed excellent similarity to fixed images, achieving a registration Dice score of 0.869 ± 0.093 for cardiac MRI and an average landmark error of 1.60 ± 1.44 mm for liver CT, substantially surpassing the state-of-the-art methods. Ablation studies further verified the effectiveness of integrating foundation features to improve DIR accuracy (<i>p</i>< 0.05).<i>Significance.</i>Our novel approach demonstrates significant advancements in DIR for multi-modality images with complex structures and low contrasts, making it a powerful tool for a wide range of applications in medical image analysis.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145934731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-21DOI: 10.1088/1361-6560/ae365a
Xuesong Lu, Huaqiu Zhao, Hong Chen, Dandan Yang, Su Zhang, Qinlan Xie
Objective.Deformable registration plays a crucial role in motion estimation from a sequence of cardiac magnetic resonance (CMR) imaging, which is good for the diagnosis and treatment of heart diseases. To address the challenges posed by intensity inhomogeneity and complex deformation, we propose a novel convolutional neural network-Transformer framework for this task.Approach.In this study, a convolutional projection Transformer block that enables efficient self-attention computation was designed for modeling long-range spatial correspondences. Additionally, a cooperative learning pattern was adopted for fusing information from global and local features. Finally, multi-resolution strategy was employed for optimizing model parameters in coarse-to-fine manner.Main result.The proposed method was evaluated on three different CMR datasets for intra-subject registration. Experimental results show that the proposed method achieves better Dice overlap and lower surface distance, compared to four non-learning-based methods and three deep-learning-based methods.Significance.For the challenging task of CMR image registration, our method demonstrates superior performance, delivering more accurate results with lower complexity. It may thus facilitate cardiac motion estimation for clinical assessments of cardiac function.
{"title":"Deformable image registration using multi-resolution vision Transformer for cardiac motion estimation.","authors":"Xuesong Lu, Huaqiu Zhao, Hong Chen, Dandan Yang, Su Zhang, Qinlan Xie","doi":"10.1088/1361-6560/ae365a","DOIUrl":"10.1088/1361-6560/ae365a","url":null,"abstract":"<p><p><i>Objective.</i>Deformable registration plays a crucial role in motion estimation from a sequence of cardiac magnetic resonance (CMR) imaging, which is good for the diagnosis and treatment of heart diseases. To address the challenges posed by intensity inhomogeneity and complex deformation, we propose a novel convolutional neural network-Transformer framework for this task.<i>Approach.</i>In this study, a convolutional projection Transformer block that enables efficient self-attention computation was designed for modeling long-range spatial correspondences. Additionally, a cooperative learning pattern was adopted for fusing information from global and local features. Finally, multi-resolution strategy was employed for optimizing model parameters in coarse-to-fine manner.<i>Main result.</i>The proposed method was evaluated on three different CMR datasets for intra-subject registration. Experimental results show that the proposed method achieves better Dice overlap and lower surface distance, compared to four non-learning-based methods and three deep-learning-based methods.<i>Significance.</i>For the challenging task of CMR image registration, our method demonstrates superior performance, delivering more accurate results with lower complexity. It may thus facilitate cardiac motion estimation for clinical assessments of cardiac function.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145945724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-01-20DOI: 10.1088/1361-6560/ae1bf4
Stefan J van der Sar, David Leibold, Dennis R Schaart
Objective.We investigate scintillation detectors with silicon photomultipliers (SiPMs) as alternatives to direct-conversion detectors based on CdTe/Cd1-xZnxTe (CZT) for x-ray photon-counting imaging. Here, we measure counting and spectral performance of three scintillators and compare the results with performances reported in literature for CdTe/CZT detectors for diagnostic photon-counting computed tomography (PCCT).Approach.We built 1 × 1 mm2single-pixel detectors by coupling readily available LYSO:Ce, YAP:Ce, and LaBr3:Ce scintillators to ultrafast SiPMs. Pulse processing was optimized for rate capability rather than energy resolution. We exposed the detectors to three radioisotopes to determine energy response proportionality and energy resolution. Using an x-ray tube, we measured x-ray spectra and count rate curves, i.e. output count rate (OCR) versus input count rate (ICR).Main results.The energy resolutions of the LYSO:Ce and YAP:Ce detectors exceed 30% full-width-at-half-maximum (FWHM) at 60 keV, with YAP:Ce showing a more proportional response. For a 30 keV count-detection threshold, the maximum OCR of the YAP:Ce detector is 5.4 Mcps pixel-1for paralyzable-like counting, while the OCR approaches 12.5 Mcps pixel-1for nonparalyzable-like counting. The LYSO:Ce detector reaches 4.5 Mcps pixel-1and 10 Mcps pixel-1, respectively, and the LaBr3:Ce detector 10.4 Mcps pixel-1and 22 Mcps pixel-1. Thereby, the rate capability of the LaBr3:Ce detector is almost 80% of that reported for two CdTe/CZT detectors for diagnostic PCCT. Moreover, the LaBr3:Ce detector has high proportionality and an energy resolution of about 20% FWHM at 60 keV, which is comparable to at least one CdTe detector for diagnostic PCCT. The x-ray tube spectra measured using the scintillation detectors show reasonable agreement with incident spectra.Significance.This work indicates that LaBr3:Ce-based detectors may become an alternative to direct-conversion detectors for diagnostic PCCT, whereas LYSO:Ce- and YAP:Ce-based detectors appear better suited for applications with lower ICR, e.g. cone-beam PCCT in radiotherapy. Ways to further improve x-ray photon-counting scintillation detectors are also discussed.
{"title":"Experimental investigation of the potential of LaBr<sub>3</sub>:Ce, LYSO:Ce, and YAP:Ce for scintillator-based x-ray photon-counting detectors.","authors":"Stefan J van der Sar, David Leibold, Dennis R Schaart","doi":"10.1088/1361-6560/ae1bf4","DOIUrl":"10.1088/1361-6560/ae1bf4","url":null,"abstract":"<p><p><i>Objective.</i>We investigate scintillation detectors with silicon photomultipliers (SiPMs) as alternatives to direct-conversion detectors based on CdTe/Cd<sub>1-<i>x</i></sub>Zn<i><sub>x</sub></i>Te (CZT) for x-ray photon-counting imaging. Here, we measure counting and spectral performance of three scintillators and compare the results with performances reported in literature for CdTe/CZT detectors for diagnostic photon-counting computed tomography (PCCT).<i>Approach.</i>We built 1 × 1 mm<sup>2</sup>single-pixel detectors by coupling readily available LYSO:Ce, YAP:Ce, and LaBr<sub>3</sub>:Ce scintillators to ultrafast SiPMs. Pulse processing was optimized for rate capability rather than energy resolution. We exposed the detectors to three radioisotopes to determine energy response proportionality and energy resolution. Using an x-ray tube, we measured x-ray spectra and count rate curves, i.e. output count rate (OCR) versus input count rate (ICR).<i>Main results.</i>The energy resolutions of the LYSO:Ce and YAP:Ce detectors exceed 30% full-width-at-half-maximum (FWHM) at 60 keV, with YAP:Ce showing a more proportional response. For a 30 keV count-detection threshold, the maximum OCR of the YAP:Ce detector is 5.4 Mcps pixel<sup>-1</sup>for paralyzable-like counting, while the OCR approaches 12.5 Mcps pixel<sup>-1</sup>for nonparalyzable-like counting. The LYSO:Ce detector reaches 4.5 Mcps pixel<sup>-1</sup>and 10 Mcps pixel<sup>-1</sup>, respectively, and the LaBr<sub>3</sub>:Ce detector 10.4 Mcps pixel<sup>-1</sup>and 22 Mcps pixel<sup>-1</sup>. Thereby, the rate capability of the LaBr<sub>3</sub>:Ce detector is almost 80% of that reported for two CdTe/CZT detectors for diagnostic PCCT. Moreover, the LaBr<sub>3</sub>:Ce detector has high proportionality and an energy resolution of about 20% FWHM at 60 keV, which is comparable to at least one CdTe detector for diagnostic PCCT. The x-ray tube spectra measured using the scintillation detectors show reasonable agreement with incident spectra.<i>Significance.</i>This work indicates that LaBr<sub>3</sub>:Ce-based detectors may become an alternative to direct-conversion detectors for diagnostic PCCT, whereas LYSO:Ce- and YAP:Ce-based detectors appear better suited for applications with lower ICR, e.g. cone-beam PCCT in radiotherapy. Ways to further improve x-ray photon-counting scintillation detectors are also discussed.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.4,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145452673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}