首页 > 最新文献

Proceedings of SPIE--the International Society for Optical Engineering最新文献

英文 中文
Improving low-contrast liver metastasis detectability in deep-learning CT denoising using adaptive local fusion driven by total uncertainty and predictive mean. 利用全不确定性和预测均值驱动的自适应局部融合提高深度学习CT去噪中低对比肝转移的检测能力。
Pub Date : 2025-02-01 Epub Date: 2025-04-08 DOI: 10.1117/12.3047080
Hao Gong, Shravani A Kharat, Shuai Leng, Lifeng Yu, Scott S Hsieh, Joel G Fletcher, Cynthia H McCollough

Emerging deep-learning-based CT denoising techniques have the potential to improve diagnostic image quality in low-dose CT exams. However, aggressive radiation dose reduction and the intrinsic uncertainty in convolutional neural network (CNN) outputs are detrimental to detecting critical lesions (e.g., liver metastases) in CNN-denoised images. To tackle these issues, we characterized CNN output distribution via total uncertainty (i.e., data + model uncertainties) and predictive mean. Local mean-uncertainty-ratio (MUR) was calculated to detect highly unreliable regions in the denoised images. A MUR-driven adaptive local fusion (ALF) process was developed to adaptively merge local predictive means with the original noisy images, thereby improving image robustness. This process was incorporated into a previously validated deep-learning model observer to quantify liver metastasis detectability, using area under localization receiver operating characteristic curve (LAUC) as the figure-of-merit. For proof-of-concept, the proposed method was established and validated for a ResNet-based CT denoising method. A recent patient abdominal CT dataset was used in validation, involving 3 lesion sizes (7, 9, and 11 mm), 3 lesion contrasts (15, 20, and 25 HU), and 3 dose levels (25%, 50%, and 100% dose). Visual inspection and quantitative analyses were conducted. Statistical significance was tested. Total uncertainty at lesions and liver background generally increased as radiation dose decreased. With fixed dose, lesion-wise MUR showed no dependency on lesion size or contrast, but exhibited large variance across lesion locations (MUR range ~0.7 to 19). Compared to original ResNet-based denoising, the MUR-driven ALF consistently improved lesion detectability in challenging conditions such as lower dose, smaller lesion size, or lower contrast (range of absolute gain in LAUC: 0.04 to 0.1; P-value 0.008). The proposed method has the potential to improve reliability of deep-learning CT denoising and enhance lesion detection.

新兴的基于深度学习的CT去噪技术有可能提高低剂量CT检查的诊断图像质量。然而,卷积神经网络(CNN)输出中的侵袭性辐射剂量降低和固有不确定性不利于检测CNN去噪图像中的关键病变(例如肝转移)。为了解决这些问题,我们通过总不确定性(即数据+模型不确定性)和预测均值来表征CNN输出分布。计算局部平均不确定比(MUR)来检测去噪图像中高度不可靠的区域。提出了一种自适应局部融合(ALF)方法,将局部预测方法与原始噪声图像自适应融合,提高了图像的鲁棒性。该过程被整合到先前验证的深度学习模型观测器中,使用定位下接收者操作特征曲线(LAUC)作为优点值来量化肝转移的可检测性。为了进行概念验证,建立并验证了基于resnet的CT去噪方法。在验证中使用了最近的患者腹部CT数据集,包括3种病变大小(7,9和11mm), 3种病变对比(15,20和25hu)和3种剂量水平(25%,50%和100%剂量)。进行目视检查和定量分析。检验统计学显著性。随着辐射剂量的降低,病变和肝脏背景的总不确定性普遍增加。在固定剂量下,病变方向的MUR不依赖于病变大小或造影剂,但在病变位置之间表现出很大的差异(MUR范围为0.7至19)。与原始的基于resnet的去噪相比,在低剂量、小病变大小或低对比度等具有挑战性的条件下,由磁共振成像驱动的ALF持续提高了病变的可检测性(LAUC的绝对增益范围:0.04至0.1;假定值0.008)。该方法具有提高深度学习CT去噪可靠性和增强病灶检测能力的潜力。
{"title":"Improving low-contrast liver metastasis detectability in deep-learning CT denoising using adaptive local fusion driven by total uncertainty and predictive mean.","authors":"Hao Gong, Shravani A Kharat, Shuai Leng, Lifeng Yu, Scott S Hsieh, Joel G Fletcher, Cynthia H McCollough","doi":"10.1117/12.3047080","DOIUrl":"10.1117/12.3047080","url":null,"abstract":"<p><p>Emerging deep-learning-based CT denoising techniques have the potential to improve diagnostic image quality in low-dose CT exams. However, aggressive radiation dose reduction and the intrinsic uncertainty in convolutional neural network (CNN) outputs are detrimental to detecting critical lesions (e.g., liver metastases) in CNN-denoised images. To tackle these issues, we characterized CNN output distribution via total uncertainty (i.e., data + model uncertainties) and predictive mean. Local mean-uncertainty-ratio (MUR) was calculated to detect highly unreliable regions in the denoised images. A MUR-driven adaptive local fusion (ALF) process was developed to adaptively merge local predictive means with the original noisy images, thereby improving image robustness. This process was incorporated into a previously validated deep-learning model observer to quantify liver metastasis detectability, using area under localization receiver operating characteristic curve (LAUC) as the figure-of-merit. For proof-of-concept, the proposed method was established and validated for a ResNet-based CT denoising method. A recent patient abdominal CT dataset was used in validation, involving 3 lesion sizes (7, 9, and 11 mm), 3 lesion contrasts (15, 20, and 25 HU), and 3 dose levels (25%, 50%, and 100% dose). Visual inspection and quantitative analyses were conducted. Statistical significance was tested. Total uncertainty at lesions and liver background generally increased as radiation dose decreased. With fixed dose, lesion-wise MUR showed no dependency on lesion size or contrast, but exhibited large variance across lesion locations (MUR range ~0.7 to 19). Compared to original ResNet-based denoising, the MUR-driven ALF consistently improved lesion detectability in challenging conditions such as lower dose, smaller lesion size, or lower contrast (range of absolute gain in LAUC: 0.04 to 0.1; P-value 0.008). The proposed method has the potential to improve reliability of deep-learning CT denoising and enhance lesion detection.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13405 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12070600/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144058608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fair Text to Medical Image Diffusion Model with Subgroup Distribution Aligned Tuning. 具有子群分布对齐调整的公平文本到医学图像扩散模型。
Pub Date : 2025-02-01 Epub Date: 2025-04-10 DOI: 10.1117/12.3046450
Xu Han, Fangfang Fan, Jingzhao Rong, Zhen Li, Georges El Fakhri, Qingyu Chen, Xiaofeng Liu

The Text to Medical Image (T2MedI) approach using latent diffusion models holds significant promise for addressing the scarcity of medical imaging data and elucidating the appearance distribution of lesions corresponding to specific patient status descriptions. Like natural image synthesis models, our investigations reveal that the T2MedI model may exhibit biases towards certain subgroups, potentially neglecting minority groups present in the training dataset. In this study, we initially developed a T2MedI model adapted from the pre-trained Imagen framework. This model employs a fixed Contrastive Language-Image Pre-training (CLIP) text encoder, with its decoder fine-tuned using medical images from the Radiology Objects in Context (ROCO) dataset. We conduct both qualitative and quantitative analyses to examine its gender bias. To address this issue, we propose a subgroup distribution alignment method during fine-tuning on a target application dataset. Specifically, this process involves an alignment loss, guided by an off-the-shelf sensitivity-subgroup classifier, which aims to synchronize the classification probabilities between the generated images and those expected in the target dataset. Additionally, we preserve image quality through a CLIP-consistency regularization term, based on a knowledge distillation framework. For evaluation purposes, we designated the BraTS18 dataset as the target, and developed a gender classifier based on brain magnetic resonance (MR) imaging slices derived from it. Our methodology significantly mitigates gender representation inconsistencies in the generated MR images, aligning them more closely with the gender distribution in the BraTS18 dataset.

使用潜在扩散模型的文本到医学图像(T2MedI)方法在解决医学成像数据的稀缺性和阐明与特定患者状态描述相对应的病变外观分布方面具有重要的前景。与自然图像合成模型一样,我们的研究表明,T2MedI模型可能会对某些子群体产生偏见,可能会忽略训练数据集中的少数群体。在本研究中,我们首先根据预训练Imagen框架开发了T2MedI模型。该模型采用固定的对比语言图像预训练(CLIP)文本编码器,其解码器使用来自放射学对象上下文(ROCO)数据集的医学图像进行微调。我们进行了定性和定量分析,以检验其性别偏见。为了解决这个问题,我们提出了一种在目标应用程序数据集微调期间的子组分布对齐方法。具体来说,这个过程涉及对齐损失,由现成的灵敏度-子组分类器指导,其目的是同步生成的图像和目标数据集中预期的分类概率。此外,我们通过基于知识蒸馏框架的clip一致性正则化项来保持图像质量。为了评估目的,我们指定BraTS18数据集作为目标,并基于其衍生的脑磁共振(MR)成像切片开发了一个性别分类器。我们的方法显著减轻了生成的MR图像中性别代表的不一致性,使其与BraTS18数据集中的性别分布更接近。
{"title":"Fair Text to Medical Image Diffusion Model with Subgroup Distribution Aligned Tuning.","authors":"Xu Han, Fangfang Fan, Jingzhao Rong, Zhen Li, Georges El Fakhri, Qingyu Chen, Xiaofeng Liu","doi":"10.1117/12.3046450","DOIUrl":"10.1117/12.3046450","url":null,"abstract":"<p><p>The Text to Medical Image (T2MedI) approach using latent diffusion models holds significant promise for addressing the scarcity of medical imaging data and elucidating the appearance distribution of lesions corresponding to specific patient status descriptions. Like natural image synthesis models, our investigations reveal that the T2MedI model may exhibit biases towards certain subgroups, potentially neglecting minority groups present in the training dataset. In this study, we initially developed a T2MedI model adapted from the pre-trained Imagen framework. This model employs a fixed Contrastive Language-Image Pre-training (CLIP) text encoder, with its decoder fine-tuned using medical images from the Radiology Objects in Context (ROCO) dataset. We conduct both qualitative and quantitative analyses to examine its gender bias. To address this issue, we propose a subgroup distribution alignment method during fine-tuning on a target application dataset. Specifically, this process involves an alignment loss, guided by an off-the-shelf sensitivity-subgroup classifier, which aims to synchronize the classification probabilities between the generated images and those expected in the target dataset. Additionally, we preserve image quality through a CLIP-consistency regularization term, based on a knowledge distillation framework. For evaluation purposes, we designated the BraTS18 dataset as the target, and developed a gender classifier based on brain magnetic resonance (MR) imaging slices derived from it. Our methodology significantly mitigates gender representation inconsistencies in the generated MR images, aligning them more closely with the gender distribution in the BraTS18 dataset.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13411 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12360154/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144884423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Conditional Generative Diffusion Model of Trabecular Bone with Tunable Microstructure. 微结构可调的小梁骨条件生成扩散模型。
Pub Date : 2025-02-01 Epub Date: 2025-04-02 DOI: 10.1117/12.3049125
X Wang, G Shi, A Sivakumar, T Ye, A Sylvester, J W Stayman, W Zbijewski

Purpose: We developed a generative model capable of producing synthetic trabecular bone that can be precisely tuned to achieve specific structural characteristics, such as bone volume fraction (BV/TV), trabecular thickness (Tb.Th), and spacing (Tb.Sp).

Methods: The generative model is based on Diffusion Transformers (DiT), a latent diffusion approach employing a transformer architecture in the denoising network. To control the microstructure characteristics of the synthetic trabecular bone samples, the model is conditioned on BV/TV, Tb.Th, and Tb.Sp. The training data involved 29898 256×256-pixel Regions of Interest (ROIs) extracted from micro-CT volumes ( 50 μ m voxel size) of 20 femoral bone specimens, paired with trabecular metrics computed within each ROI; the training/validation split was 9:1. For testing, 3499 synthetic bone samples were generated over a wide range of condition (target) microstructure metrics. Results were evaluated in terms of (i) the ability to cover real-world distribution of trabecular structures (coverage), (ii) agreement with target metric values (Pearson Correlation), and (iii) consistency of the metrics across multiple realizations of the DiT model with fixed condition (Coefficient of Variation, CV).

Results: The model achieved good coverage of real-world bone microstructures and visual similarity to true trabecular ROIs. Pearson Correlations against the condition (target) metric values were high: 0.9540 for BV/TV, 0.9618 for Tb.Th, and 0.9835 Tb.Sp. Microstructural characteristics of the synthetic samples were stable across DiT realizations, with CV ranging from 3.37% to 11.78% for BV/TV, 2.27% to 3.22% for Tb.Th, and 2.53% to 5.00% for Tb.Sp.

Conclusion: The proposed generative model is capable of generating realistic digital trabecular bones that can be precisely tuned to achieve specified microstructural characteristics. Possible applications include virtual clinical trials of new skeletal image biomarkers and establishing priors for advanced image reconstruction.

目的:我们开发了一种能够生产合成骨小梁的生成模型,该模型可以精确调整以实现特定的结构特征,如骨体积分数(BV/TV),骨小梁厚度(Tb.Th)和间距(Tb.Sp)。方法:生成模型基于扩散变压器(Diffusion transformer, DiT),这是一种在去噪网络中采用变压器结构的潜在扩散方法。为了控制合成骨小梁样品的微观结构特征,模型的条件为BV/TV, Tb。这个,还有这个。训练数据涉及29898个256×256-pixel从20个股骨标本的微ct体积(50 μ m体素大小)中提取的感兴趣区域(ROI),并与每个ROI内计算的小梁指标配对;训练/验证的比例是9:1。为了进行测试,在广泛的条件(目标)微观结构指标下生成了3499个合成骨样品。评估结果的依据是:(i)覆盖小梁结构真实分布的能力(覆盖度),(ii)与目标度量值的一致性(Pearson相关性),以及(iii)固定条件下DiT模型在多个实现中的度量一致性(变异系数,CV)。结果:该模型实现了真实骨微结构的良好覆盖,视觉上与真实小梁roi相似。与条件(目标)度量值的Pearson相关性很高:BV/TV为0.9540,Tb为0.9618。Th和0.9835 Tb.Sp。合成样品的微观结构特征在DiT实现中稳定,BV/TV的变异系数为3.37% ~ 11.78%,Tb的变异系数为2.27% ~ 3.22%。结论:所提出的生成模型能够生成逼真的数字小梁骨,可以精确调整以达到特定的微观结构特征。可能的应用包括新的骨骼图像生物标志物的虚拟临床试验和建立高级图像重建的先验。
{"title":"A Conditional Generative Diffusion Model of Trabecular Bone with Tunable Microstructure.","authors":"X Wang, G Shi, A Sivakumar, T Ye, A Sylvester, J W Stayman, W Zbijewski","doi":"10.1117/12.3049125","DOIUrl":"10.1117/12.3049125","url":null,"abstract":"<p><strong>Purpose: </strong>We developed a generative model capable of producing synthetic trabecular bone that can be precisely tuned to achieve specific structural characteristics, such as bone volume fraction (BV/TV), trabecular thickness (Tb.Th), and spacing (Tb.Sp).</p><p><strong>Methods: </strong>The generative model is based on Diffusion Transformers (DiT), a latent diffusion approach employing a transformer architecture in the denoising network. To control the microstructure characteristics of the synthetic trabecular bone samples, the model is conditioned on BV/TV, Tb.Th, and Tb.Sp. The training data involved 29898 256×256-pixel Regions of Interest (ROIs) extracted from micro-CT volumes ( <math><mn>50</mn> <mspace></mspace> <mi>μ</mi> <mtext>m</mtext></math> voxel size) of 20 femoral bone specimens, paired with trabecular metrics computed within each ROI; the training/validation split was 9:1. For testing, 3499 synthetic bone samples were generated over a wide range of condition (target) microstructure metrics. Results were evaluated in terms of (i) the ability to cover real-world distribution of trabecular structures (coverage), (ii) agreement with target metric values (Pearson Correlation), and (iii) consistency of the metrics across multiple realizations of the DiT model with fixed condition (Coefficient of Variation, CV).</p><p><strong>Results: </strong>The model achieved good coverage of real-world bone microstructures and visual similarity to true trabecular ROIs. Pearson Correlations against the condition (target) metric values were high: 0.9540 for BV/TV, 0.9618 for Tb.Th, and 0.9835 Tb.Sp. Microstructural characteristics of the synthetic samples were stable across DiT realizations, with CV ranging from 3.37% to 11.78% for BV/TV, 2.27% to 3.22% for Tb.Th, and 2.53% to 5.00% for Tb.Sp.</p><p><strong>Conclusion: </strong>The proposed generative model is capable of generating realistic digital trabecular bones that can be precisely tuned to achieve specified microstructural characteristics. Possible applications include virtual clinical trials of new skeletal image biomarkers and establishing priors for advanced image reconstruction.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13410 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12302783/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144746429","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Black-box Optimization of CT Acquisition and Reconstruction Parameters: A Reinforcement Learning Approach. CT采集和重建参数的黑盒优化:一种强化学习方法。
Pub Date : 2025-02-01 Epub Date: 2025-04-08 DOI: 10.1117/12.3046807
David Fenwick, Navid NaderiAlizadeh, Vahid Tarokh, Darin Clark, Jayasai Rajagopal, Anuj Kapadia, Nicholas Felice, Ehsan Samei, Ehsan Abadi

Protocol optimization is critical in Computed Tomography (CT) for achieving desired diagnostic image quality while minimizing radiation dose. Due to the inter-effect of influencing CT parameters, traditional optimization methods rely on the testing of exhaustive combinations of these parameters. This poses a notable limitation due to the impracticality of exhaustive parameter testing. This study introduces a novel methodology leveraging Virtual Imaging Trials (VITs) and reinforcement learning to more efficiently optimize CT protocols. Computational phantoms with liver lesions were imaged using a validated CT simulator and reconstructed with a novel CT reconstruction Toolkit. The optimization parameter space included tube voltage, tube current, reconstruction kernel, slice thickness, and pixel size. The optimization process was done using a Proximal Policy Optimization (PPO) agent which was trained to maximize the Detectability Index (d') of the liver lesion for each reconstructed image. Results showed that our reinforcement learning approach found the absolute maximum d' across the test cases while requiring 79.7% fewer steps compared to an exhaustive search, demonstrating both accuracy and computational efficiency, offering a efficient and robust framework for CT protocol optimization. The flexibility of the proposed technique allows for use of varying image quality metrics as the objective metric to maximize for. Our findings highlight the advantages of combining VIT and reinforcement learning for CT protocol management.

在计算机断层扫描(CT)中,方案优化是实现所需诊断图像质量同时最小化辐射剂量的关键。由于影响连续油管参数的相互作用,传统的优化方法依赖于这些参数的穷举组合测试。由于穷举参数测试的不实用性,这造成了明显的限制。本研究介绍了一种利用虚拟成像试验(VITs)和强化学习的新方法,以更有效地优化CT协议。使用经过验证的CT模拟器对肝脏病变的计算幻影进行成像,并使用新的CT重建工具包进行重建。优化参数空间包括管电压、管电流、重构核、切片厚度和像素大小。优化过程使用近端策略优化(PPO)代理完成,该代理被训练为最大化每个重建图像的肝脏病变的可检测指数(d')。结果表明,我们的强化学习方法在测试用例中找到了绝对最大的d',而与穷举搜索相比,所需的步骤减少了79.7%,证明了准确性和计算效率,为CT协议优化提供了高效且稳健的框架。所提出的技术的灵活性允许使用不同的图像质量度量作为客观度量来最大化。我们的研究结果强调了将VIT和强化学习相结合用于CT协议管理的优势。
{"title":"Black-box Optimization of CT Acquisition and Reconstruction Parameters: A Reinforcement Learning Approach.","authors":"David Fenwick, Navid NaderiAlizadeh, Vahid Tarokh, Darin Clark, Jayasai Rajagopal, Anuj Kapadia, Nicholas Felice, Ehsan Samei, Ehsan Abadi","doi":"10.1117/12.3046807","DOIUrl":"10.1117/12.3046807","url":null,"abstract":"<p><p>Protocol optimization is critical in Computed Tomography (CT) for achieving desired diagnostic image quality while minimizing radiation dose. Due to the inter-effect of influencing CT parameters, traditional optimization methods rely on the testing of exhaustive combinations of these parameters. This poses a notable limitation due to the impracticality of exhaustive parameter testing. This study introduces a novel methodology leveraging Virtual Imaging Trials (VITs) and reinforcement learning to more efficiently optimize CT protocols. Computational phantoms with liver lesions were imaged using a validated CT simulator and reconstructed with a novel CT reconstruction Toolkit. The optimization parameter space included tube voltage, tube current, reconstruction kernel, slice thickness, and pixel size. The optimization process was done using a Proximal Policy Optimization (PPO) agent which was trained to maximize the Detectability Index (d') of the liver lesion for each reconstructed image. Results showed that our reinforcement learning approach found the absolute maximum d' across the test cases while requiring 79.7% fewer steps compared to an exhaustive search, demonstrating both accuracy and computational efficiency, offering a efficient and robust framework for CT protocol optimization. The flexibility of the proposed technique allows for use of varying image quality metrics as the objective metric to maximize for. Our findings highlight the advantages of combining VIT and reinforcement learning for CT protocol management.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13405 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12035822/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144045469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Intra- and inter-scanner CT variability and their impact on diagnostic tasks. 扫描仪内和扫描仪间的CT变异性及其对诊断任务的影响。
Pub Date : 2025-02-01 Epub Date: 2025-04-08 DOI: 10.1117/12.3047016
Isabel Montero, Saman Sotoudeh-Paima, Ehsan Abadi, Ehsan Samei

The increased development and production of Computed Tomography (CT) scanner technology has expanded patient access to advanced and affordable medical imaging technologies but has also introduced sources of variability in the clinical imaging landscape, which may influence patient care. This study examines the impact of intra-scanner and inter-scanner variability on image quality and quantitative imaging tasks, with a focus on the detectability index (d') as a measure of patient-specific task performance. We evaluated 813 clinical phantom image sets from the COPDGene study, aggregated by CT scanner make, model, and acquisition and reconstruction protocol. Each phantom image set was assessed for image quality metrics, including the Noise Power Spectrum (NPS) and in-plane Modulation Transfer Function (MTF). The d' index was calculated for 12 hypothetical lesion detection tasks, emulating clinically relevant lung and liver lesions of varying sizes and contrast levels. Qualitatively, analysis showed intra-scanner variability in NPS and MTF curves measured for identical acquisition and reconstruction settings. Inter-scanner comparisons demonstrated variability in d' measurements across different scanner makes and models, of similar acquisition and reconstruction settings. The study showed an intra-scanner variability of up to 13.7% and an inter-scanner variability of up to 19.3% in the d' index. These findings emphasize the need for considering scanner variability in patient-centered care and indicate that CT technology may influence the reliability of imaging tasks. The results of this study further motivate the development of virtual scanner models to better model and mitigate the variability observed in the clinical imaging landscape.

计算机断层扫描(CT)扫描仪技术的不断发展和生产扩大了患者获得先进和负担得起的医学成像技术的机会,但也引入了临床成像领域的变变性来源,这可能影响患者的护理。本研究考察了扫描仪内和扫描仪间的可变性对图像质量和定量成像任务的影响,重点关注可检测性指数(d')作为衡量患者特定任务表现的指标。我们评估了来自COPDGene研究的813个临床幻影图像集,这些图像集由CT扫描仪制造、模型、获取和重建方案汇总而成。评估每个幻影图像集的图像质量指标,包括噪声功率谱(NPS)和面内调制传递函数(MTF)。对12个假设的病变检测任务进行d'指数计算,模拟临床相关的不同大小和对比水平的肺和肝病变。定性分析显示,在相同的采集和重建设置下,测量的NPS和MTF曲线的扫描仪内变异性。扫描仪之间的比较表明,在类似的采集和重建设置下,不同扫描仪制造和型号的d'测量值存在差异。研究表明,d'指数的扫描仪内变异性高达13.7%,扫描仪间变异性高达19.3%。这些发现强调了在以患者为中心的护理中考虑扫描仪可变性的必要性,并表明CT技术可能会影响成像任务的可靠性。这项研究的结果进一步推动了虚拟扫描仪模型的发展,以更好地模拟和减轻临床成像领域观察到的可变性。
{"title":"Intra- and inter-scanner CT variability and their impact on diagnostic tasks.","authors":"Isabel Montero, Saman Sotoudeh-Paima, Ehsan Abadi, Ehsan Samei","doi":"10.1117/12.3047016","DOIUrl":"10.1117/12.3047016","url":null,"abstract":"<p><p>The increased development and production of Computed Tomography (CT) scanner technology has expanded patient access to advanced and affordable medical imaging technologies but has also introduced sources of variability in the clinical imaging landscape, which may influence patient care. This study examines the impact of intra-scanner and inter-scanner variability on image quality and quantitative imaging tasks, with a focus on the detectability index (d') as a measure of patient-specific task performance. We evaluated 813 clinical phantom image sets from the COPDGene study, aggregated by CT scanner make, model, and acquisition and reconstruction protocol. Each phantom image set was assessed for image quality metrics, including the Noise Power Spectrum (NPS) and in-plane Modulation Transfer Function (MTF). The d' index was calculated for 12 hypothetical lesion detection tasks, emulating clinically relevant lung and liver lesions of varying sizes and contrast levels. Qualitatively, analysis showed intra-scanner variability in NPS and MTF curves measured for identical acquisition and reconstruction settings. Inter-scanner comparisons demonstrated variability in d' measurements across different scanner makes and models, of similar acquisition and reconstruction settings. The study showed an intra-scanner variability of up to 13.7% and an inter-scanner variability of up to 19.3% in the d' index. These findings emphasize the need for considering scanner variability in patient-centered care and indicate that CT technology may influence the reliability of imaging tasks. The results of this study further motivate the development of virtual scanner models to better model and mitigate the variability observed in the clinical imaging landscape.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13405 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12035824/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144037036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Weighted Circle Fusion: Ensembling Circle Representation from Different Object Detection Results. 加权圆融合:不同目标检测结果的圆表示集成。
Pub Date : 2025-02-01 Epub Date: 2025-04-10 DOI: 10.1117/12.3047295
Jialin Yue, Tianyuan Yao, Ruining Deng, Quan Liu, Juming Xiong, Junlin Guo, Haichun Yang, Yuankai Huo

Recently, the use of circle representation has emerged as a method to improve the identification of spherical objects (such as glomeruli, cells, and nuclei) in medical imaging studies. In traditional bounding box-based object detection, combining results from multiple models improves accuracy, especially when real-time processing isn't crucial. Unfortunately, this widely adopted strategy is not readily available for combining circle representations. In this paper, we propose Weighted Circle Fusion (WCF), a simple approach for merging predictions from various circle detection models. Our method leverages confidence scores associated with each proposed bounding circle to generate averaged circles. We evaluate our method on a proprietary dataset for glomerular detection in whole slide imaging (WSI) and find a performance gain of 5% compared to existing ensemble methods. Additionally, we assess the efficiency of two annotation methods-fully manual annotation and a human-in-the-loop (HITL) approach-in labeling 200,000 glomeruli. The HITL approach, which integrates machine learning detection with human verification, demonstrated remarkable improvements in annotation efficiency. The Weighted Circle Fusion technique not only enhances object detection precision but also notably reduces false detections, presenting a promising direction for future research and application in pathological image analysis. The source code has been made publicly available at https://github.com/hrlblab/WeightedCircleFusion.

最近,在医学成像研究中,使用圆形表示已成为一种改进球形物体(如肾小球、细胞和细胞核)识别的方法。在传统的基于边界盒的目标检测中,结合多个模型的结果可以提高精度,特别是在实时处理不重要的情况下。不幸的是,这种被广泛采用的策略并不容易用于组合圆表示。在本文中,我们提出加权圆融合(WCF),这是一种简单的方法,用于合并来自各种圆检测模型的预测。我们的方法利用与每个提议的边界圆相关联的置信度得分来生成平均圆。我们在一个专有数据集上评估了我们的方法在全玻片成像(WSI)中肾小球检测,发现与现有的集成方法相比,性能提高了5%。此外,我们评估了两种标注方法的效率-完全手动标注和人在环(HITL)方法-在标记200,000个肾小球。HITL方法将机器学习检测与人工验证相结合,在标注效率上有了显著提高。加权圆融合技术不仅提高了目标检测精度,而且显著降低了误检率,在病理图像分析中具有广阔的研究和应用前景。源代码已在https://github.com/hrlblab/WeightedCircleFusion上公开提供。
{"title":"Weighted Circle Fusion: Ensembling Circle Representation from Different Object Detection Results.","authors":"Jialin Yue, Tianyuan Yao, Ruining Deng, Quan Liu, Juming Xiong, Junlin Guo, Haichun Yang, Yuankai Huo","doi":"10.1117/12.3047295","DOIUrl":"10.1117/12.3047295","url":null,"abstract":"<p><p>Recently, the use of circle representation has emerged as a method to improve the identification of spherical objects (such as glomeruli, cells, and nuclei) in medical imaging studies. In traditional bounding box-based object detection, combining results from multiple models improves accuracy, especially when real-time processing isn't crucial. Unfortunately, this widely adopted strategy is not readily available for combining circle representations. In this paper, we propose Weighted Circle Fusion (WCF), a simple approach for merging predictions from various circle detection models. Our method leverages confidence scores associated with each proposed bounding circle to generate averaged circles. We evaluate our method on a proprietary dataset for glomerular detection in whole slide imaging (WSI) and find a performance gain of 5% compared to existing ensemble methods. Additionally, we assess the efficiency of two annotation methods-fully manual annotation and a human-in-the-loop (HITL) approach-in labeling 200,000 glomeruli. The HITL approach, which integrates machine learning detection with human verification, demonstrated remarkable improvements in annotation efficiency. The Weighted Circle Fusion technique not only enhances object detection precision but also notably reduces false detections, presenting a promising direction for future research and application in pathological image analysis. The source code has been made publicly available at https://github.com/hrlblab/WeightedCircleFusion.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13409 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12515093/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145282187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward non-invasive diagnosis of Bankart lesions with deep learning. 基于深度学习的Bankart病变无创诊断研究。
Pub Date : 2025-02-01 Epub Date: 2025-04-04 DOI: 10.1117/12.3046251
Sahil Sethi, Sai Reddy, Mansi Sakarvadia, Jordan Serotte, Darlington Nwaudo, Nicholas Maassen, Lewis Shi

Purpose: Bankart lesions, or anterior-inferior glenoid labral tears, are diagnostically challenging on standard MRIs due to their subtle imaging features-often necessitating invasive MRI arthrograms (MRAs). This study develops deep learning (DL) models to detect Bankart lesions on both standard MRIs and MRAs, aiming to improve diagnostic accuracy and reduce reliance on MRAs.

Methods: We curated a dataset of 586 shoulder MRIs (335 standard, 251 MRAs) from 558 patients who underwent arthroscopy. Ground truth labels were derived from intraoperative findings, the gold standard for Bankart lesion diagnosis. Separate DL models for MRAs and standard MRIs were trained using the Swin Transformer architecture, pre-trained on a public knee MRI dataset. Predictions from sagittal, axial, and coronal views were ensembled to optimize performance. The models were evaluated on a 20% hold-out test set (117 MRIs: 46 MRAs, 71 standard MRIs).

Results: Bankart lesions were identified in 31.9% of MRAs and 8.6% of standard MRIs. The models achieved AUCs of 0.87 (86% accuracy, 83% sensitivity, 86% specificity) and 0.90 (85% accuracy, 82% sensitivity, 86% specificity) on standard MRIs and MRAs, respectively. These results match or surpass radiologist performance on our dataset and reported literature metrics. Notably, our model's performance on non-invasive standard MRIs matched or surpassed the radiologists interpreting MRAs.

Conclusion: This study demonstrates the feasibility of using DL to address the diagnostic challenges posed by subtle pathologies like Bankart lesions. Our models demonstrate potential to improve diagnostic confidence, reduce reliance on invasive imaging, and enhance accessibility to care.

目的:Bankart病变,或前下盂唇撕裂,由于其微妙的成像特征,通常需要有创性MRI关节造影(MRAs),在标准MRI上诊断具有挑战性。本研究开发了深度学习(DL)模型,在标准核磁共振成像和核磁共振成像上检测Bankart病变,旨在提高诊断准确性并减少对核磁共振成像的依赖。方法:我们从558名接受关节镜检查的患者中收集了586张肩部mri(335张标准mri, 251张mra)的数据集。基础真值标签来源于术中发现,这是Bankart病变诊断的金标准。使用Swin Transformer架构对mra和标准MRI的单独DL模型进行训练,该架构在公共膝盖MRI数据集上进行预训练。矢状面、轴状面和冠状面的预测被整合以优化性能。模型在20%滞留测试集上进行评估(117个核磁共振成像:46个核磁共振成像,71个标准核磁共振成像)。结果:在31.9%的mra和8.6%的标准mri中发现了Bankart病变。该模型在标准mri和MRAs上的auc分别为0.87(准确率86%,灵敏度83%,特异性86%)和0.90(准确率85%,灵敏度82%,特异性86%)。这些结果匹配或超过放射科医生在我们的数据集和报告的文献指标上的表现。值得注意的是,我们的模型在非侵入性标准核磁共振成像上的表现匹配或超过了放射科医生对核磁共振成像的解读。结论:本研究证明了使用DL来解决像Bankart病变这样的细微病理所带来的诊断挑战的可行性。我们的模型展示了提高诊断可信度、减少对侵入性影像学依赖和提高护理可及性的潜力。
{"title":"Toward non-invasive diagnosis of Bankart lesions with deep learning.","authors":"Sahil Sethi, Sai Reddy, Mansi Sakarvadia, Jordan Serotte, Darlington Nwaudo, Nicholas Maassen, Lewis Shi","doi":"10.1117/12.3046251","DOIUrl":"10.1117/12.3046251","url":null,"abstract":"<p><strong>Purpose: </strong>Bankart lesions, or anterior-inferior glenoid labral tears, are diagnostically challenging on standard MRIs due to their subtle imaging features-often necessitating invasive MRI arthrograms (MRAs). This study develops deep learning (DL) models to detect Bankart lesions on both standard MRIs and MRAs, aiming to improve diagnostic accuracy and reduce reliance on MRAs.</p><p><strong>Methods: </strong>We curated a dataset of 586 shoulder MRIs (335 standard, 251 MRAs) from 558 patients who underwent arthroscopy. Ground truth labels were derived from intraoperative findings, the gold standard for Bankart lesion diagnosis. Separate DL models for MRAs and standard MRIs were trained using the Swin Transformer architecture, pre-trained on a public knee MRI dataset. Predictions from sagittal, axial, and coronal views were ensembled to optimize performance. The models were evaluated on a 20% hold-out test set (117 MRIs: 46 MRAs, 71 standard MRIs).</p><p><strong>Results: </strong>Bankart lesions were identified in 31.9% of MRAs and 8.6% of standard MRIs. The models achieved AUCs of 0.87 (86% accuracy, 83% sensitivity, 86% specificity) and 0.90 (85% accuracy, 82% sensitivity, 86% specificity) on standard MRIs and MRAs, respectively. These results match or surpass radiologist performance on our dataset and reported literature metrics. Notably, our model's performance on non-invasive standard MRIs matched or surpassed the radiologists interpreting MRAs.</p><p><strong>Conclusion: </strong>This study demonstrates the feasibility of using DL to address the diagnostic challenges posed by subtle pathologies like Bankart lesions. Our models demonstrate potential to improve diagnostic confidence, reduce reliance on invasive imaging, and enhance accessibility to care.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13407 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12670170/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simulating scanner-and algorithm-specific 3D CT noise texture using physics-informed 2D and 2.5D generative neural network models. 使用物理信息2D和2.5D生成神经网络模型模拟扫描仪和算法特定的3D CT噪声纹理。
Pub Date : 2025-02-01 Epub Date: 2025-04-08 DOI: 10.1117/12.3047909
Hao Gong, Thomas M Huber, Timothy Winfree, Scott S Hsieh, Lifeng Yu, Shuai Leng, Cynthia H McCollough

Low-dose CT simulation is needed to assess reconstruction/denoising techniques and optimize dose. Projection-domain noise-insertion methods require manufacturers' proprietary tools. Image-domain noise-insertion methods face various challenges that affect generalizability, and few have been systematically validated for 3D noise synthesis. To improve generalizability, we presented a physics-informed model-based generative neural network for simulating scanner- and algorithm-specific low-dose CT exams (PALETTE). PALETTE included a noise-prior-generation process, a Noise2Noisier sub-network, and a noise-texture-synthesis sub-network. Custom regularization terms were developed to enforce 3D noise texture quality. Using PALETTE, one 2D and two 2.5D models (denoted as 2.5D N-N and N-1) were developed to conduct 2D and effective 3D noise modeling, respectively (input/output images: 2D - 1/1, 2.5D N-N - 3/3, 2.5D N-1 - 5/1). These models were trained and tested with an open-access abdominal CT dataset, including 20 testing cases reconstructed with two kernels and various field-of-view. In visual inspection, the 2D and 2.5D N-N models generated realistic local and global noise texture, while 2.5D N-1 showed more perceptual difference using the sharper kernel and coronal reformat. In quantitative evaluation, local noise level was compared using mean-absolute-percent-difference (MAPD), and global spectral similarity was assessed using spectral correlation mapper (SCM) and spectral angle mapper (SAM). The 2D model provided equivalent or relatively better performance than 2.5D models, showing well-matched local noise levels and high spectral similarity compared to the reference (sharper/smoother kernels): MAPD - 2D 1.5%/5.6% (p>0.05), 2.5D N-N 8.5%/7.9% (p<0.05), 2.5D N-1 12.3%/10.9% (p<0.05); mean SCM - 2D 0.97/0.97, 2.5D N-N 0.96/0.97, 2.5D N-1 0.85/0.97; mean SAM - 2D 0.12/0.12, 2.5D N-N 0.14/0.12, 2.5D N-1 0.37/0.12. With tripled model width, the 2.5D N-N outperformed N-1. This indicated 2.5D models need more learning capacity to further enhance 3D noise modeling. Using physics-based prior information, PALETTE can provide high-quality low-dose CT simulation to resemble scanner- and algorithm-specific 3D noise characteristics.

低剂量CT模拟需要评估重建/去噪技术和优化剂量。投影域噪声插入方法需要制造商的专有工具。图像域噪声插入方法面临着影响其泛化性的各种挑战,很少有方法对三维噪声合成进行系统验证。为了提高通用性,我们提出了一个基于物理信息模型的生成神经网络,用于模拟扫描仪和算法特定的低剂量CT检查(PALETTE)。调色板包括一个噪声先验生成过程、一个Noise2Noisier子网络和一个噪声-纹理合成子网络。开发了自定义正则化术语来增强3D噪声纹理质量。利用PALETTE分别建立了一个2D模型和两个2.5D模型(记为2.5D N-N和N-1),分别进行二维和有效的三维噪声建模(输入/输出图像:2D - 1/1、2.5D N-N - 3/3、2.5D N-1 - 5/1)。这些模型使用开放获取的腹部CT数据集进行训练和测试,其中包括20个用两个核和不同视场重建的测试案例。在视觉检测中,2D和2.5D N-N模型产生了真实的局部和全局噪声纹理,而2.5D N-1模型通过更清晰的核和冠状重构显示出更多的感知差异。在定量评价中,使用平均绝对百分比差(MAPD)比较局部噪声水平,使用光谱相关映射器(SCM)和光谱角映射器(SAM)评估全局光谱相似性。2D模型提供了与2.5D模型相当或相对更好的性能,与参考模型相比,显示出良好的局部噪声水平和高光谱相似性(更清晰/更平滑的内核):MAPD - 2D 1.5%/5.6% (p>0.05), 2.5D N-N 8.5%/7.9% (pN-1 12.3%/10.9% (pN-N 0.96/0.97, 2.5D N-1 0.85/0.97);平均SAM - 2D 0.12/0.12, 2.5D N-N 0.14/0.12, 2.5D N-1 0.37/0.12。随着模型宽度的三倍,2.5D N-N的性能优于N-1。这表明2.5D模型需要更多的学习能力来进一步增强三维噪声建模。利用基于物理的先验信息,PALETTE可以提供高质量的低剂量CT模拟,以模拟扫描仪和算法特定的3D噪声特征。
{"title":"Simulating scanner-and algorithm-specific 3D CT noise texture using physics-informed 2D and 2.5D generative neural network models.","authors":"Hao Gong, Thomas M Huber, Timothy Winfree, Scott S Hsieh, Lifeng Yu, Shuai Leng, Cynthia H McCollough","doi":"10.1117/12.3047909","DOIUrl":"10.1117/12.3047909","url":null,"abstract":"<p><p>Low-dose CT simulation is needed to assess reconstruction/denoising techniques and optimize dose. Projection-domain noise-insertion methods require manufacturers' proprietary tools. Image-domain noise-insertion methods face various challenges that affect generalizability, and few have been systematically validated for 3D noise synthesis. To improve generalizability, we presented a <b>p</b>hysics-informed model-based gener<b>a</b>tive neura<b>l</b> network for simulating scann<b>e</b>r- and algorithm-specific low-dose C<b>T e</b>xams (PALETTE). PALETTE included a noise-prior-generation process, a Noise2Noisier sub-network, and a noise-texture-synthesis sub-network. Custom regularization terms were developed to enforce 3D noise texture quality. Using PALETTE, one 2D and two 2.5D models (denoted as 2.5D <i>N-N</i> and <i>N-1</i>) were developed to conduct 2D and effective 3D noise modeling, respectively (input/output images: 2D - 1/1, 2.5D <i>N-N</i> - 3/3, 2.5D <i>N-1</i> - 5/1). These models were trained and tested with an open-access abdominal CT dataset, including 20 testing cases reconstructed with two kernels and various field-of-view. In visual inspection, the 2D and 2.5D <i>N-N</i> models generated realistic local and global noise texture, while 2.5D <i>N-1</i> showed more perceptual difference using the sharper kernel and coronal reformat. In quantitative evaluation, local noise level was compared using mean-absolute-percent-difference (MAPD), and global spectral similarity was assessed using spectral correlation mapper (SCM) and spectral angle mapper (SAM). The 2D model provided equivalent or relatively better performance than 2.5D models, showing well-matched local noise levels and high spectral similarity compared to the reference (sharper/smoother kernels): MAPD - 2D 1.5%/5.6% (p>0.05), 2.5D <i>N-N</i> 8.5%/7.9% (p<0.05), 2.5D <i>N-1</i> 12.3%/10.9% (p<0.05); mean SCM - 2D 0.97/0.97, 2.5D <i>N-N</i> 0.96/0.97, 2.5D <i>N-1</i> 0.85/0.97; mean SAM - 2D 0.12/0.12, 2.5D <i>N-N</i> 0.14/0.12, 2.5D <i>N-1</i> 0.37/0.12. With tripled model width, the 2.5D <i>N-N</i> outperformed <i>N-1</i>. This indicated 2.5D models need more learning capacity to further enhance 3D noise modeling. Using physics-based prior information, PALETTE can provide high-quality low-dose CT simulation to resemble scanner- and algorithm-specific 3D noise characteristics.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13405 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12070530/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144062947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Patient-specific Channelized Hotelling observer to estimate lesion detectability in CT. 患者特异性通道化Hotelling观察者在CT上评估病变的可检出性。
Pub Date : 2025-02-01 Epub Date: 2025-04-08 DOI: 10.1117/12.3047381
Zhongxing Zhou, Jarod Wellinghoff, Cynthia H McCollough, Lifeng Yu

Task-based image quality assessment is essential for CT protocol and radiation dose optimization. Despite many ongoing efforts, there is still an unmet need to measure and monitor the quality of images acquired from each patient exam. In this work, we developed a patient-specific channelized Hotelling observer (CHO)-based method to estimate the lesion detectability for each individual patient scan. The ensemble of background was created from patient images to include both relatively uniform regions and anatomically varying regions. Signals were modelled from lesions of different sizes and contrast levels after incorporating the effect of contrast-dependent spatial resolution. Index of detectability (d') was estimated using a CHO framework. This method was applied to clinical patient images obtained from a CT scanner at 3 different radiation dose levels. The d' for 5 different lesion size/contrast conditions was calculated across the scan range of each patient exam. The average noise levels and the d' averaged from 5 conditions were 13.2/3.78, 17.1/2.93 and 21.9/2.43 at 100%, 50% and 25% dose levels, respectively.

基于任务的图像质量评估对于CT方案和辐射剂量优化至关重要。尽管正在进行许多努力,但仍然需要测量和监测从每个患者检查中获得的图像的质量。在这项工作中,我们开发了一种基于患者特异性通道化霍特林观察者(CHO)的方法来估计每个患者扫描的病变可检测性。背景集合是由患者图像创建的,包括相对均匀的区域和解剖学上不同的区域。在纳入对比度相关空间分辨率的影响后,信号从不同大小和对比度水平的病变中建模。使用CHO框架估计可检测性指数(d')。该方法应用于CT扫描仪在3种不同辐射剂量水平下获得的临床患者图像。在每个患者检查的扫描范围内计算5种不同病变大小/造影剂条件的d'。在100%、50%和25%剂量水平下,5种条件下的平均噪声级和d′平均值分别为13.2/3.78、17.1/2.93和21.9/2.43。
{"title":"Patient-specific Channelized Hotelling observer to estimate lesion detectability in CT.","authors":"Zhongxing Zhou, Jarod Wellinghoff, Cynthia H McCollough, Lifeng Yu","doi":"10.1117/12.3047381","DOIUrl":"10.1117/12.3047381","url":null,"abstract":"<p><p>Task-based image quality assessment is essential for CT protocol and radiation dose optimization. Despite many ongoing efforts, there is still an unmet need to measure and monitor the quality of images acquired from each patient exam. In this work, we developed a patient-specific channelized Hotelling observer (CHO)-based method to estimate the lesion detectability for each individual patient scan. The ensemble of background was created from patient images to include both relatively uniform regions and anatomically varying regions. Signals were modelled from lesions of different sizes and contrast levels after incorporating the effect of contrast-dependent spatial resolution. Index of detectability (d') was estimated using a CHO framework. This method was applied to clinical patient images obtained from a CT scanner at 3 different radiation dose levels. The d' for 5 different lesion size/contrast conditions was calculated across the scan range of each patient exam. The average noise levels and the d' averaged from 5 conditions were 13.2/3.78, 17.1/2.93 and 21.9/2.43 at 100%, 50% and 25% dose levels, respectively.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13405 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12086740/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144103308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ensembled YOLO for multiorgan detection in chest x-rays. 用于胸部x光多器官检测的集成YOLO。
Pub Date : 2025-02-01 Epub Date: 2025-04-04 DOI: 10.1117/12.3047210
Sivaramakrishnan Rajaraman, Zhaohui Liang, Zhiyun Xue, Sameer Antani

Chest radiographs are a vital tool for identifying pathological changes within the thoracic cavity. Artificial intelligence (AI) and machine learning (ML) driven screening or diagnostic applications require accurate detection of anatomical structures within the Chest X-ray (CXR) image. The You Only Look Once (YOLO) object detection models have recently gained prominence for their efficacy in detecting anatomical structures in medical images. However, state-of-the-art results using it are typically for single anatomical organ detection. Advanced image analysis would benefit from simultaneous detection more than one anatomical organ. In this work we propose a multi-organ detection technique through two recent YOLO versions and their sub-variants. We evaluate their effectiveness in detecting lung and heart regions in CXRs simultaneously. We used the JSRT CXR dataset for internal training, validation, and testing. Further, the generalizability of the models is evaluated using two external test sets, viz., the Montgomery CXR dataset and a subset of the RSNA CXR dataset against available annotations therein. Our evaluation demonstrates that YOLOv9 models notably outperform YOLOv8 variants. We demonstrated further improvements in detection performance through ensemble approaches.

胸片是鉴别胸腔内病变的重要工具。人工智能(AI)和机器学习(ML)驱动的筛查或诊断应用需要准确检测胸部x光片(CXR)图像中的解剖结构。You Only Look Once (YOLO)目标检测模型最近因其在检测医学图像中的解剖结构方面的有效性而获得了突出的地位。然而,使用它的最先进的结果通常用于单个解剖器官检测。先进的图像分析将受益于同时检测多个解剖器官。在这项工作中,我们通过两个最新的YOLO版本及其子变体提出了一种多器官检测技术。我们评估了它们在cxr中同时检测肺和心脏区域的有效性。我们使用JSRT CXR数据集进行内部训练、验证和测试。此外,使用两个外部测试集(即Montgomery CXR数据集和RSNA CXR数据集的一个子集)对其中的可用注释评估模型的泛化性。我们的评估表明,YOLOv9模型明显优于YOLOv8变体。我们展示了通过集成方法进一步改进检测性能。
{"title":"Ensembled YOLO for multiorgan detection in chest x-rays.","authors":"Sivaramakrishnan Rajaraman, Zhaohui Liang, Zhiyun Xue, Sameer Antani","doi":"10.1117/12.3047210","DOIUrl":"10.1117/12.3047210","url":null,"abstract":"<p><p>Chest radiographs are a vital tool for identifying pathological changes within the thoracic cavity. Artificial intelligence (AI) and machine learning (ML) driven screening or diagnostic applications require accurate detection of anatomical structures within the Chest X-ray (CXR) image. The You Only Look Once (YOLO) object detection models have recently gained prominence for their efficacy in detecting anatomical structures in medical images. However, state-of-the-art results using it are typically for single anatomical organ detection. Advanced image analysis would benefit from simultaneous detection more than one anatomical organ. In this work we propose a multi-organ detection technique through two recent YOLO versions and their sub-variants. We evaluate their effectiveness in detecting lung and heart regions in CXRs simultaneously. We used the JSRT CXR dataset for internal training, validation, and testing. Further, the generalizability of the models is evaluated using two external test sets, viz., the Montgomery CXR dataset and a subset of the RSNA CXR dataset against available annotations therein. Our evaluation demonstrates that YOLOv9 models notably outperform YOLOv8 variants. We demonstrated further improvements in detection performance through ensemble approaches.</p>","PeriodicalId":74505,"journal":{"name":"Proceedings of SPIE--the International Society for Optical Engineering","volume":"13407 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11996225/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144060406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of SPIE--the International Society for Optical Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1