首页 > 最新文献

Journal of Medical Imaging最新文献

英文 中文
Analysis of intra- and inter-observer variability in 4D liver ultrasound landmark labeling. 四维肝脏超声标志物标记的观察者间和观察者内变异分析。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-06-30 DOI: 10.1117/1.JMI.12.5.051807
Daniel Wulff, Floris Ernst

Purpose: Four-dimensional (4D) ultrasound imaging is widely used in clinics for diagnostics and therapy guidance. Accurate target tracking in 4D ultrasound is crucial for autonomous therapy guidance systems, such as radiotherapy, where precise tumor localization ensures effective treatment. Supervised deep learning approaches rely on reliable ground truth, making accurate labels essential. We investigate the reliability of expert-labeled ground truth data by evaluating intra- and inter-observer variability in landmark labeling for 4D ultrasound imaging in the liver.

Approach: Eight 4D liver ultrasound sequences were labeled by eight expert observers, each labeling eight landmarks three times. Intra- and inter-observer variability was quantified, and observer survey and motion analysis were conducted to determine factors influencing labeling accuracy, such as ultrasound artifacts and motion amplitude.

Results: The mean intra-observer variability ranged from 1.58    mm ± 0.90    mm to 2.05    mm ± 1.22    mm depending on the observer. The inter-observer variability for the two observer groups was 2.68    mm ± 1.69    mm and 3.06    mm ± 1.74    mm . The observer survey and motion analysis revealed that ultrasound artifacts significantly affected labeling accuracy due to limited landmark visibility, whereas motion amplitude had no measurable effect. Our measured mean landmark motion was 11.56    mm ± 5.86    mm .

Conclusions: We highlight variability in expert-labeled ground truth data for 4D ultrasound imaging and identify ultrasound artifacts as a major source of labeling inaccuracies. These findings underscore the importance of addressing observer variability and artifact-related challenges to improve the reliability of ground truth data for evaluating target tracking algorithms in 4D ultrasound applications.

目的:四维超声成像广泛应用于临床诊断和治疗指导。在四维超声中精确的目标跟踪对于自主治疗引导系统至关重要,例如放射治疗,其中精确的肿瘤定位确保有效治疗。有监督的深度学习方法依赖于可靠的真实情况,因此准确的标签至关重要。我们通过评估肝脏四维超声成像中地标标记的观察者内部和观察者之间的可变性来研究专家标记的真实数据的可靠性。方法:8个4D肝脏超声序列由8名专家观察者标记,每个标记3次,标记8个地标。量化观察者内部和观察者之间的可变性,并进行观察者调查和运动分析,以确定影响标记准确性的因素,如超声伪影和运动幅度。结果:根据观察者的不同,观察者内部的平均变异范围为1.58 mm±0.90 mm至2.05 mm±1.22 mm。两个观察组的观察者间变异分别为2.68 mm±1.69 mm和3.06 mm±1.74 mm。观察者调查和运动分析显示,由于地标可见性有限,超声伪影显著影响标记准确性,而运动幅度没有可测量的影响。我们测量到的平均地标运动为11.56 mm±5.86 mm。结论:我们强调了专家标记的四维超声成像地面真值数据的可变性,并确定超声伪影是标记不准确的主要来源。这些发现强调了解决观测者可变性和伪影相关挑战的重要性,以提高评估4D超声应用中目标跟踪算法的地面真实数据的可靠性。
{"title":"Analysis of intra- and inter-observer variability in 4D liver ultrasound landmark labeling.","authors":"Daniel Wulff, Floris Ernst","doi":"10.1117/1.JMI.12.5.051807","DOIUrl":"10.1117/1.JMI.12.5.051807","url":null,"abstract":"<p><strong>Purpose: </strong>Four-dimensional (4D) ultrasound imaging is widely used in clinics for diagnostics and therapy guidance. Accurate target tracking in 4D ultrasound is crucial for autonomous therapy guidance systems, such as radiotherapy, where precise tumor localization ensures effective treatment. Supervised deep learning approaches rely on reliable ground truth, making accurate labels essential. We investigate the reliability of expert-labeled ground truth data by evaluating intra- and inter-observer variability in landmark labeling for 4D ultrasound imaging in the liver.</p><p><strong>Approach: </strong>Eight 4D liver ultrasound sequences were labeled by eight expert observers, each labeling eight landmarks three times. Intra- and inter-observer variability was quantified, and observer survey and motion analysis were conducted to determine factors influencing labeling accuracy, such as ultrasound artifacts and motion amplitude.</p><p><strong>Results: </strong>The mean intra-observer variability ranged from <math><mrow><mn>1.58</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>0.90</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> to <math><mrow><mn>2.05</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>1.22</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> depending on the observer. The inter-observer variability for the two observer groups was <math><mrow><mn>2.68</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>1.69</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> and <math><mrow><mn>3.06</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>1.74</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> . The observer survey and motion analysis revealed that ultrasound artifacts significantly affected labeling accuracy due to limited landmark visibility, whereas motion amplitude had no measurable effect. Our measured mean landmark motion was <math><mrow><mn>11.56</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>5.86</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> .</p><p><strong>Conclusions: </strong>We highlight variability in expert-labeled ground truth data for 4D ultrasound imaging and identify ultrasound artifacts as a major source of labeling inaccuracies. These findings underscore the importance of addressing observer variability and artifact-related challenges to improve the reliability of ground truth data for evaluating target tracking algorithms in 4D ultrasound applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"051807"},"PeriodicalIF":1.9,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12207815/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144545488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of algorithmic requirements for clinical application of material decomposition using a multi-layer flat panel detector. 多层平板探测器材料分解临床应用的算法要求评价。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-09-04 DOI: 10.1117/1.JMI.12.5.053501
Jamin Schaefer, Steffen Kappler, Ferdinand Lueck, Ludwig Ritschl, Thomas Weber, Georg Rose

Purpose: The combination of multi-layer flat panel detector (FPDT) X-ray imaging and physics-based material decomposition algorithms allows for the removal of anatomical structures. However, the reliability of these algorithms may be compromised by unaccounted materials or scattered radiation.

Approach: We investigated the two-material decomposition performance of a multi-layer FPDT in the context of 2D chest radiography without and with a 13:1 anti-scatter grid employed. A matrix-based material decomposition (MBMD) (equivalent to weighted logarithmic subtraction), a matrix-based material decomposition with polynomial beam hardening pre-correction (MBMD-PBC), and a projection domain decomposition were evaluated. The decomposition accuracy of simulated data was evaluated by comparing the bone and soft tissue images to the ground truth using the structural similarity index measure (SSIM). Simulation results were supported by experiments using a commercially available triple-layer FPDT retrofitted to a digital X-ray system.

Results: Independent of the selected decomposition algorithm, uncorrected scatter leads to negative bone estimates, resulting in small SSIM values and bone structures to remain visible in soft tissue images. Even with a 13:1 anti-scatter grid employed, bone images continue to show negative bone estimates, and bone structures appear in soft tissue images. Adipose tissue on the contrary has an almost negligible effect.

Conclusions: In a contact scan, scattered radiation leads to negative bone contrast estimates in the bone images and remaining bone contrast in the soft tissue images. Therefore, accurate scatter estimation and correction algorithms are essential when aiming for material decomposition using image data obtained with a multi-layer FPDT.

目的:多层平板探测器(FPDT) x射线成像和基于物理的材料分解算法相结合,可以去除解剖结构。然而,这些算法的可靠性可能会受到不明材料或散射辐射的影响。方法:我们研究了多层FPDT在不使用和使用13:1抗散射网格的二维胸片背景下的双材料分解性能。评估了基于矩阵的材料分解(MBMD)(相当于加权对数减法)、基于矩阵的材料分解与多项式光束硬化预校正(MBMD- pbc)以及投影域分解。利用结构相似指数度量(SSIM)将骨骼和软组织图像与地面真实情况进行比较,评估模拟数据的分解精度。仿真结果得到了商用三层FPDT改造成数字x射线系统的实验的支持。结果:与所选择的分解算法无关,未校正的散点会导致骨骼估计为负,导致软组织图像中SSIM值较小,骨骼结构仍然可见。即使采用13:1的反散射网格,骨骼图像仍然显示负骨估计,并且骨骼结构出现在软组织图像中。相反,脂肪组织的影响几乎可以忽略不计。结论:在接触扫描中,散射辐射导致骨骼图像中的骨对比度估计为负,软组织图像中的骨对比度估计为剩余。因此,在利用多层FPDT获得的图像数据进行材料分解时,精确的散射估计和校正算法是必不可少的。
{"title":"Evaluation of algorithmic requirements for clinical application of material decomposition using a multi-layer flat panel detector.","authors":"Jamin Schaefer, Steffen Kappler, Ferdinand Lueck, Ludwig Ritschl, Thomas Weber, Georg Rose","doi":"10.1117/1.JMI.12.5.053501","DOIUrl":"10.1117/1.JMI.12.5.053501","url":null,"abstract":"<p><strong>Purpose: </strong>The combination of multi-layer flat panel detector (FPDT) X-ray imaging and physics-based material decomposition algorithms allows for the removal of anatomical structures. However, the reliability of these algorithms may be compromised by unaccounted materials or scattered radiation.</p><p><strong>Approach: </strong>We investigated the two-material decomposition performance of a multi-layer FPDT in the context of 2D chest radiography without and with a 13:1 anti-scatter grid employed. A matrix-based material decomposition (MBMD) (equivalent to weighted logarithmic subtraction), a matrix-based material decomposition with polynomial beam hardening pre-correction (MBMD-PBC), and a projection domain decomposition were evaluated. The decomposition accuracy of simulated data was evaluated by comparing the bone and soft tissue images to the ground truth using the structural similarity index measure (SSIM). Simulation results were supported by experiments using a commercially available triple-layer FPDT retrofitted to a digital X-ray system.</p><p><strong>Results: </strong>Independent of the selected decomposition algorithm, uncorrected scatter leads to negative bone estimates, resulting in small SSIM values and bone structures to remain visible in soft tissue images. Even with a 13:1 anti-scatter grid employed, bone images continue to show negative bone estimates, and bone structures appear in soft tissue images. Adipose tissue on the contrary has an almost negligible effect.</p><p><strong>Conclusions: </strong>In a contact scan, scattered radiation leads to negative bone contrast estimates in the bone images and remaining bone contrast in the soft tissue images. Therefore, accurate scatter estimation and correction algorithms are essential when aiming for material decomposition using image data obtained with a multi-layer FPDT.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"053501"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12410756/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145015224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Statistical testing of agreement in overlap-based performance between an AI segmentation device and a multi-expert human panel without requiring a reference standard. 在不需要参考标准的情况下,对人工智能分割设备和多专家小组之间基于重叠的性能的一致性进行统计测试。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-22 DOI: 10.1117/1.JMI.12.5.055003
Tingting Hu, Berkman Sahiner, Shuyue Guan, Mike Mikailov, Kenny Cha, Frank Samuelson, Nicholas Petrick

Purpose: Artificial intelligence (AI)-based medical imaging devices often include lesion or organ segmentation capabilities. Existing methods for segmentation performance evaluation compare AI results with an aggregated reference standard using accuracy metrics such as the Dice coefficient or Hausdorff distance. However, these approaches are limited by lacking a gold standard and challenges in defining meaningful success criteria. To address this, we developed a statistical method to assess agreement between an AI device and multiple human experts without requiring a reference standard.

Approach: We propose a paired-testing method to evaluate whether an AI device's segmentation performance significantly differs from that of multiple human experts. The method compares device-to-expert dissimilarity with expert-to-expert dissimilarity, avoiding the need for a reference standard. We validated the method through (1) statistical simulations where the Dice coefficient performance is either shared ("overlap agreeable") or not shared ("overlap disagreeable") between the device and experts; (2) image-based simulations using 2D contours with shared or nonshared transformation parameters (transformation agreeable or disagreeable). We also applied the method to compare an AI segmentation algorithm with four radiologists using data from the Lung Image Database Consortium.

Results: Statistical simulations show the method controls type I error ( 0.05 ) for overlap-agreeable and type II error ( 0 ) for overlap-disagreeable scenarios. Image-based simulations show acceptable performance with a mean type I error of 0.07 (SD 0.03) for transformation-agreeable and a mean type II error of 0.07 (SD 0.18) for transformation-disagreeable cases.

Conclusions: The paired-testing method offers a new tool for assessing the agreement between an AI segmentation device and multiple human expert panelists without requiring a reference standard.

目的:基于人工智能(AI)的医学成像设备通常包括病变或器官分割功能。现有的分割性能评估方法使用精确度指标(如Dice系数或Hausdorff距离)将AI结果与汇总参考标准进行比较。然而,这些方法由于缺乏黄金标准和定义有意义的成功标准的挑战而受到限制。为了解决这个问题,我们开发了一种统计方法来评估人工智能设备和多个人类专家之间的一致性,而不需要参考标准。方法:我们提出了一种配对测试方法来评估人工智能设备的分割性能是否与多个人类专家的分割性能有显著差异。该方法比较了设备与专家之间的不相似性和专家与专家之间的不相似性,避免了对参考标准的需要。我们通过(1)统计模拟验证了该方法,其中Dice系数性能在设备和专家之间共享(“重叠一致”)或不共享(“重叠不一致”);(2)基于图像的模拟,使用具有共享或非共享变换参数的二维轮廓(变换符合或不符合)。我们还应用该方法将人工智能分割算法与四位放射科医生使用肺图像数据库联盟的数据进行比较。结果:统计模拟表明,该方法控制了重叠适宜情景的I型误差(~ 0.05)和重叠不适宜情景的II型误差(~ 0)。基于图像的模拟显示出可接受的性能,对于符合变换的情况,平均I类误差为0.07 (SD 0.03),对于不符合变换的情况,平均II类误差为0.07 (SD 0.18)。结论:配对测试方法为评估人工智能分割设备与多个人类专家小组成员之间的一致性提供了一种新工具,而无需参考标准。
{"title":"Statistical testing of agreement in overlap-based performance between an AI segmentation device and a multi-expert human panel without requiring a reference standard.","authors":"Tingting Hu, Berkman Sahiner, Shuyue Guan, Mike Mikailov, Kenny Cha, Frank Samuelson, Nicholas Petrick","doi":"10.1117/1.JMI.12.5.055003","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.055003","url":null,"abstract":"<p><strong>Purpose: </strong>Artificial intelligence (AI)-based medical imaging devices often include lesion or organ segmentation capabilities. Existing methods for segmentation performance evaluation compare AI results with an aggregated reference standard using accuracy metrics such as the Dice coefficient or Hausdorff distance. However, these approaches are limited by lacking a gold standard and challenges in defining meaningful success criteria. To address this, we developed a statistical method to assess agreement between an AI device and multiple human experts without requiring a reference standard.</p><p><strong>Approach: </strong>We propose a paired-testing method to evaluate whether an AI device's segmentation performance significantly differs from that of multiple human experts. The method compares device-to-expert dissimilarity with expert-to-expert dissimilarity, avoiding the need for a reference standard. We validated the method through (1) statistical simulations where the Dice coefficient performance is either shared (\"overlap agreeable\") or not shared (\"overlap disagreeable\") between the device and experts; (2) image-based simulations using 2D contours with shared or nonshared transformation parameters (transformation agreeable or disagreeable). We also applied the method to compare an AI segmentation algorithm with four radiologists using data from the Lung Image Database Consortium.</p><p><strong>Results: </strong>Statistical simulations show the method controls type I error ( <math><mrow><mo>∼</mo> <mn>0.05</mn></mrow> </math> ) for overlap-agreeable and type II error ( <math><mrow><mo>∼</mo> <mn>0</mn></mrow> </math> ) for overlap-disagreeable scenarios. Image-based simulations show acceptable performance with a mean type I error of 0.07 (SD 0.03) for transformation-agreeable and a mean type II error of 0.07 (SD 0.18) for transformation-disagreeable cases.</p><p><strong>Conclusions: </strong>The paired-testing method offers a new tool for assessing the agreement between an AI segmentation device and multiple human expert panelists without requiring a reference standard.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"055003"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12543030/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking 3D generative autoencoders for pseudo-healthy reconstruction of brain 18F-fluorodeoxyglucose positron emission tomography. 脑18f -氟脱氧葡萄糖正电子发射断层假健康重建的三维生成式自编码器的基准测试。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-16 DOI: 10.1117/1.JMI.12.5.054005
Ravi Hassanaly, Maëlys Solal, Olivier Colliot, Ninon Burgos

Purpose: Many deep generative models have been proposed to reconstruct pseudo-healthy images for anomaly detection. Among these models, the variational autoencoder (VAE) has emerged as both simple and efficient. Although significant progress has been made in refining the VAE within the field of computer vision, these advancements have not been extensively applied to medical imaging applications.

Approach: We present a benchmark that assesses the ability of multiple VAEs to reconstruct pseudo-healthy neuroimages for anomaly detection in the context of dementia. We first propose a rigorous methodology to define the optimal architecture of the vanilla VAE and select through random searches the best hyperparameters of the VAE variants. Relying on a simulation-based evaluation framework, we thoroughly assess the ability of 20 VAE models to reconstruct pseudo-healthy images for the detection of dementia-related anomalies in 3D brain F 18 -fluorodeoxyglucose (FDG) positron emission tomography (PET) and compare their performance.

Results: This benchmark demonstrated that the majority of the VAE models tested were able to reconstruct images of good quality and generate healthy-looking images from simulated images presenting anomalies.

Conclusions: Even if no model clearly outperformed all the others, the benchmark allowed identifying a few models that perform slightly better than the vanilla VAE. It further showed that many VAE-based models can generalize to the detection of anomalies of various intensities, shapes, and locations in 3D brain FDG PET.

目的:提出了许多深度生成模型来重建用于异常检测的伪健康图像。在这些模型中,变分自编码器(VAE)以其简单和高效的特点而出现。尽管在细化计算机视觉领域的VAE方面取得了重大进展,但这些进展尚未广泛应用于医学成像应用。方法:我们提出了一个基准,评估多个VAEs重建伪健康神经图像的能力,以便在痴呆症的背景下进行异常检测。我们首先提出了一种严格的方法来定义香草VAE的最优架构,并通过随机搜索选择VAE变体的最佳超参数。基于基于模拟的评估框架,我们全面评估了20种VAE模型重建伪健康图像的能力,以检测3D脑f18 -氟脱氧葡萄糖(FDG)正电子发射断层扫描(PET)中与痴呆症相关的异常,并比较了它们的性能。结果:该基准测试表明,大多数测试的VAE模型能够重建质量良好的图像,并从呈现异常的模拟图像中生成看起来健康的图像。结论:即使没有模型明显优于所有其他模型,基准测试允许识别一些比普通VAE稍微好一点的模型。进一步表明,许多基于vae的模型可以推广到3D脑FDG PET中各种强度、形状和位置的异常检测。
{"title":"Benchmarking 3D generative autoencoders for pseudo-healthy reconstruction of brain <sup>18</sup>F-fluorodeoxyglucose positron emission tomography.","authors":"Ravi Hassanaly, Maëlys Solal, Olivier Colliot, Ninon Burgos","doi":"10.1117/1.JMI.12.5.054005","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.054005","url":null,"abstract":"<p><strong>Purpose: </strong>Many deep generative models have been proposed to reconstruct pseudo-healthy images for anomaly detection. Among these models, the variational autoencoder (VAE) has emerged as both simple and efficient. Although significant progress has been made in refining the VAE within the field of computer vision, these advancements have not been extensively applied to medical imaging applications.</p><p><strong>Approach: </strong>We present a benchmark that assesses the ability of multiple VAEs to reconstruct pseudo-healthy neuroimages for anomaly detection in the context of dementia. We first propose a rigorous methodology to define the optimal architecture of the vanilla VAE and select through random searches the best hyperparameters of the VAE variants. Relying on a simulation-based evaluation framework, we thoroughly assess the ability of 20 VAE models to reconstruct pseudo-healthy images for the detection of dementia-related anomalies in 3D brain <math> <mrow> <mmultiscripts><mrow><mi>F</mi></mrow> <mprescripts></mprescripts> <none></none> <mrow><mn>18</mn></mrow> </mmultiscripts> </mrow> </math> -fluorodeoxyglucose (FDG) positron emission tomography (PET) and compare their performance.</p><p><strong>Results: </strong>This benchmark demonstrated that the majority of the VAE models tested were able to reconstruct images of good quality and generate healthy-looking images from simulated images presenting anomalies.</p><p><strong>Conclusions: </strong>Even if no model clearly outperformed all the others, the benchmark allowed identifying a few models that perform slightly better than the vanilla VAE. It further showed that many VAE-based models can generalize to the detection of anomalies of various intensities, shapes, and locations in 3D brain FDG PET.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"054005"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12543032/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DMM-UNet: dual-path multi-scale Mamba UNet for medical image segmentation. DMM-UNet:用于医学图像分割的双路径多尺度曼巴UNet。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-09-29 DOI: 10.1117/1.JMI.12.5.054003
Liquan Zhao, Mingxia Cao, Yanfei Jia

Purpose: State space models have shown promise in medical image segmentation by modeling long-range dependencies with linear complexity. However, they are limited in their ability to capture local features, which hinders their capacity to extract multiscale details and integrate global and local contextual information effectively. To address these shortcomings, we propose the dual-path multi-scale Mamba UNet (DMM-UNet) model.

Approach: This architecture facilitates deep fusion of local and global features through multi-scale modules within a U-shaped encoder-decoder framework. First, we introduce the multi-scale channel attention selective scanning block in the encoder, which combines global selective scanning with multi-scale channel attention to model both long-range and local dependencies simultaneously. Second, we design the spatial attention selective scanning block for the decoder. This block integrates global scanning with spatial attention mechanisms, enabling precise aggregation of semantic features through gated weighting. Finally, we develop the multi-dimensional collaborative attention layer to extract complementary attention weights across height, width, and channel dimensions, facilitating cross-space-channel feature interactions.

Results: Experiments were conducted on the ISIC17, ISIC18, Synapse, and ACDC datasets. One of the indicators, Dice similarity coefficient, achieved 89.88% on the ISIC17 dataset, 90.52% on the ISIC18 dataset, 83.07% on the Synapse dataset, and 92.60% on the ACDC dataset. There are also other indicators that perform well on this model.

Conclusions: The DMM-UNet model effectively addresses the shortcomings of state space models by enabling the integration of both local and global features, improving segmentation performance, and offering enhanced multiscale feature fusion for medical image segmentation tasks.

目的:状态空间模型通过对具有线性复杂性的远程依赖关系进行建模,在医学图像分割中显示出良好的前景。然而,它们捕获局部特征的能力有限,这阻碍了它们提取多尺度细节和有效整合全局和局部上下文信息的能力。为了解决这些缺点,我们提出了双路径多尺度曼巴UNet (DMM-UNet)模型。方法:该架构通过u型编码器-解码器框架内的多尺度模块促进局部和全局特征的深度融合。首先,我们在编码器中引入了多尺度通道注意选择性扫描块,将全局选择性扫描与多尺度通道注意相结合,同时对远程和局部依赖关系进行建模。其次,设计了译码器的空间注意选择扫描块。该块集成了全局扫描和空间注意机制,通过门控加权实现语义特征的精确聚合。最后,我们开发了多维协同关注层,以提取跨高度、宽度和通道维度的互补关注权重,促进跨空间通道特征交互。结果:在ISIC17、ISIC18、Synapse和ACDC数据集上进行了实验。其中Dice相似系数在ISIC17数据集上达到89.88%,在ISIC18数据集上达到90.52%,在Synapse数据集上达到83.07%,在ACDC数据集上达到92.60%。还有其他一些指标在这个模型上表现良好。结论:DMM-UNet模型有效地解决了状态空间模型的不足,实现了局部和全局特征的融合,提高了分割性能,并为医学图像分割任务提供了增强的多尺度特征融合。
{"title":"DMM-UNet: dual-path multi-scale Mamba UNet for medical image segmentation.","authors":"Liquan Zhao, Mingxia Cao, Yanfei Jia","doi":"10.1117/1.JMI.12.5.054003","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.054003","url":null,"abstract":"<p><strong>Purpose: </strong>State space models have shown promise in medical image segmentation by modeling long-range dependencies with linear complexity. However, they are limited in their ability to capture local features, which hinders their capacity to extract multiscale details and integrate global and local contextual information effectively. To address these shortcomings, we propose the dual-path multi-scale Mamba UNet (DMM-UNet) model.</p><p><strong>Approach: </strong>This architecture facilitates deep fusion of local and global features through multi-scale modules within a U-shaped encoder-decoder framework. First, we introduce the multi-scale channel attention selective scanning block in the encoder, which combines global selective scanning with multi-scale channel attention to model both long-range and local dependencies simultaneously. Second, we design the spatial attention selective scanning block for the decoder. This block integrates global scanning with spatial attention mechanisms, enabling precise aggregation of semantic features through gated weighting. Finally, we develop the multi-dimensional collaborative attention layer to extract complementary attention weights across height, width, and channel dimensions, facilitating cross-space-channel feature interactions.</p><p><strong>Results: </strong>Experiments were conducted on the ISIC17, ISIC18, Synapse, and ACDC datasets. One of the indicators, Dice similarity coefficient, achieved 89.88% on the ISIC17 dataset, 90.52% on the ISIC18 dataset, 83.07% on the Synapse dataset, and 92.60% on the ACDC dataset. There are also other indicators that perform well on this model.</p><p><strong>Conclusions: </strong>The DMM-UNet model effectively addresses the shortcomings of state space models by enabling the integration of both local and global features, improving segmentation performance, and offering enhanced multiscale feature fusion for medical image segmentation tasks.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"054003"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12480969/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145208018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using a limited field of view to improve training for pulmonary nodule detection on radiographs. 利用有限视场改进x线片上肺结节检测的培训。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-04-25 DOI: 10.1117/1.JMI.12.5.051804
Samual K Zenger, Rishabh Agarwal, William F Auffermann

Purpose: Perceptual error is a significant cause of medical errors in radiology. Given the amount of information in a medical image, an image interpreter may become distracted by information unrelated to their search pattern. This may be especially challenging for novices. We aim to examine teaching medical trainees to evaluate chest radiographs (CXRs) for pulmonary nodules on limited field-of-view (LFOV) images, with the field of view (FOV) restricted to the lungs and mediastinum.

Approach: Healthcare trainees with limited exposure to interpreting images were asked to identify pulmonary nodules on CXRs, half of which contained nodules. The control and experimental groups evaluated two sets of CXRs. After the first set, the experimental group was trained to evaluate LFOV images, and both groups were again asked to assess CXRs for pulmonary nodules. Participants were given surveys after this educational session to determine their thoughts about the training and symptoms of computer vision syndrome (CVS).

Results: There was a significant improvement in performance in pulmonary nodule identification for both the experimental and control groups, but the improvement was more considerable in the experimental group ( p - value = 0.022 ). Survey responses were uniformly positive, and each question was statistically significant (all p - values < 0.001 ).

Conclusions: Our results show that using LFOV images may be helpful when teaching trainees specific high-yield perceptual tasks, such as nodule identification. The use of LFOV images was associated with reduced symptoms of CVS.

目的:感知错误是导致放射学医疗错误的重要原因。给定医学图像中的信息量,图像解释器可能会被与其搜索模式无关的信息分散注意力。这对新手来说尤其具有挑战性。我们的目的是检查医学培训生在有限视场(LFOV)图像上评估胸片(cxr)对肺结节的诊断,视场(FOV)仅限于肺和纵隔。方法:医疗保健培训生与有限的接触解释图像被要求识别肺结节的cxr,其中一半包含结节。对照组和实验组分别评价两组cxr。在第一组之后,实验组接受训练以评估LFOV图像,两组再次被要求评估肺结节的cxr。在这一教育课程结束后,参与者接受了调查,以确定他们对训练和计算机视觉综合征(CVS)症状的看法。结果:实验组与对照组肺结节识别能力均有显著提高,但实验组提高更明显(p值= 0.022)。调查结果一致是肯定的,每个问题都有统计学意义(p值均为0.001)。结论:我们的研究结果表明,在教授受训者特定的高收益感知任务(如结节识别)时,使用LFOV图像可能有所帮助。LFOV图像的使用与CVS症状的减轻有关。
{"title":"Using a limited field of view to improve training for pulmonary nodule detection on radiographs.","authors":"Samual K Zenger, Rishabh Agarwal, William F Auffermann","doi":"10.1117/1.JMI.12.5.051804","DOIUrl":"10.1117/1.JMI.12.5.051804","url":null,"abstract":"<p><strong>Purpose: </strong>Perceptual error is a significant cause of medical errors in radiology. Given the amount of information in a medical image, an image interpreter may become distracted by information unrelated to their search pattern. This may be especially challenging for novices. We aim to examine teaching medical trainees to evaluate chest radiographs (CXRs) for pulmonary nodules on limited field-of-view (LFOV) images, with the field of view (FOV) restricted to the lungs and mediastinum.</p><p><strong>Approach: </strong>Healthcare trainees with limited exposure to interpreting images were asked to identify pulmonary nodules on CXRs, half of which contained nodules. The control and experimental groups evaluated two sets of CXRs. After the first set, the experimental group was trained to evaluate LFOV images, and both groups were again asked to assess CXRs for pulmonary nodules. Participants were given surveys after this educational session to determine their thoughts about the training and symptoms of computer vision syndrome (CVS).</p><p><strong>Results: </strong>There was a significant improvement in performance in pulmonary nodule identification for both the experimental and control groups, but the improvement was more considerable in the experimental group ( <math><mrow><mi>p</mi> <mtext>-</mtext> <mtext>value</mtext> <mo>=</mo> <mn>0.022</mn></mrow> </math> ). Survey responses were uniformly positive, and each question was statistically significant (all <math><mrow><mi>p</mi> <mtext>-</mtext> <mtext>values</mtext> <mo><</mo> <mn>0.001</mn></mrow> </math> ).</p><p><strong>Conclusions: </strong>Our results show that using LFOV images may be helpful when teaching trainees specific high-yield perceptual tasks, such as nodule identification. The use of LFOV images was associated with reduced symptoms of CVS.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"051804"},"PeriodicalIF":1.9,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12023444/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144021338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep-learning-based washout classification for decision support in contrast-enhanced ultrasound examinations of the liver. 基于深度学习的洗刷分类在肝脏超声造影检查中的决策支持。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-07-01 Epub Date: 2025-07-22 DOI: 10.1117/1.JMI.12.4.044502
Hannah Strohm, Sven Rothlübbers, Jürgen Jenne, Dirk-André Clevert, Thomas Fischer, Niklas Hitschrich, Bernhard Mumm, Paul Spiesecke, Matthias Günther

Purpose: Contrast-enhanced ultrasound (CEUS) is a reliable tool to diagnose focal liver lesions, which appear ambiguous in normal B-mode ultrasound. However, interpretation of the dynamic contrast sequences can be challenging, hindering the widespread application of CEUS. We investigate the use of a deep-learning-based image classifier for determining the diagnosis-relevant feature washout from CEUS acquisitions.

Approach: We introduce a data representation, which is agnostic to data heterogeneity regarding lesion size, subtype, and length of the sequences. Then, an image-based classifier is exploited for washout classification. Strategies to cope with sparse annotations and motion are systematically evaluated, as well as the potential benefits of using a perfusion model to cover missing time points.

Results: Results indicate decent performance comparable to studies found in the literature, with a maximum balanced accuracy of 84.0% on the validation and 82.0% on the test set. Correlation-based frame selection yielded improvements in classification performance, whereas further motion compensation did not show any benefit in the conducted experiments.

Conclusions: It is shown that deep-learning-based washout classification is feasible in principle. It offers a simple form of interpretability compared with benign versus malignant classifications. The concept of classifying individual features instead of the diagnosis itself could be extended to other features such as the arterial inflow behavior. The main factors distinguishing it from existing approaches are the data representation and task formulation, as well as a large dataset size with 500 liver lesions from two centers for algorithmic development and testing.

目的:对比增强超声(CEUS)是诊断局灶性肝脏病变的可靠工具,在正常b超中表现不明确。然而,动态对比序列的解释可能具有挑战性,阻碍了超声造影的广泛应用。我们研究了基于深度学习的图像分类器的使用,以确定从CEUS获取的诊断相关特征冲洗。方法:我们引入了一种数据表示,它与病变大小、亚型和序列长度的数据异质性无关。然后,利用基于图像的分类器进行冲洗分类。系统地评估了处理稀疏注释和运动的策略,以及使用灌注模型覆盖缺失时间点的潜在好处。结果:结果表明,与文献中发现的研究相比,性能良好,验证的最大平衡精度为84.0%,测试集的最大平衡精度为82.0%。基于相关性的帧选择提高了分类性能,而进一步的运动补偿在实验中没有显示出任何好处。结论:基于深度学习的水洗分类在原则上是可行的。与良性与恶性分类相比,它提供了一种简单的可解释性形式。分类个体特征而不是诊断本身的概念可以扩展到其他特征,如动脉流入行为。将其与现有方法区分开来的主要因素是数据表示和任务制定,以及来自两个算法开发和测试中心的500个肝脏病变的大型数据集。
{"title":"Deep-learning-based washout classification for decision support in contrast-enhanced ultrasound examinations of the liver.","authors":"Hannah Strohm, Sven Rothlübbers, Jürgen Jenne, Dirk-André Clevert, Thomas Fischer, Niklas Hitschrich, Bernhard Mumm, Paul Spiesecke, Matthias Günther","doi":"10.1117/1.JMI.12.4.044502","DOIUrl":"https://doi.org/10.1117/1.JMI.12.4.044502","url":null,"abstract":"<p><strong>Purpose: </strong>Contrast-enhanced ultrasound (CEUS) is a reliable tool to diagnose focal liver lesions, which appear ambiguous in normal B-mode ultrasound. However, interpretation of the dynamic contrast sequences can be challenging, hindering the widespread application of CEUS. We investigate the use of a deep-learning-based image classifier for determining the diagnosis-relevant feature washout from CEUS acquisitions.</p><p><strong>Approach: </strong>We introduce a data representation, which is agnostic to data heterogeneity regarding lesion size, subtype, and length of the sequences. Then, an image-based classifier is exploited for washout classification. Strategies to cope with sparse annotations and motion are systematically evaluated, as well as the potential benefits of using a perfusion model to cover missing time points.</p><p><strong>Results: </strong>Results indicate decent performance comparable to studies found in the literature, with a maximum balanced accuracy of 84.0% on the validation and 82.0% on the test set. Correlation-based frame selection yielded improvements in classification performance, whereas further motion compensation did not show any benefit in the conducted experiments.</p><p><strong>Conclusions: </strong>It is shown that deep-learning-based washout classification is feasible in principle. It offers a simple form of interpretability compared with benign versus malignant classifications. The concept of classifying individual features instead of the diagnosis itself could be extended to other features such as the arterial inflow behavior. The main factors distinguishing it from existing approaches are the data representation and task formulation, as well as a large dataset size with 500 liver lesions from two centers for algorithmic development and testing.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044502"},"PeriodicalIF":1.9,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12279466/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144700098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Harnessing chemically crosslinked microbubble clusters using deep learning for ultrasound contrast imaging. 利用化学交联微泡簇进行超声对比成像的深度学习。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-07-01 Epub Date: 2025-07-12 DOI: 10.1117/1.JMI.12.4.047001
Teja Pathour, Ghazal Rastegar, Shashank R Sirsi, Baowei Fei

Purpose: We aim to investigate and isolate the distinctive acoustic properties generated by chemically crosslinked microbubble clusters (CCMCs) using machine learning (ML) techniques, specifically using an anomaly detection model based on autoencoders.

Approach: CCMCs were synthesized via copper-free click chemistry and subjected to acoustic analysis using a clinical transducer. Radiofrequency data were acquired, processed, and organized into training and testing datasets for the ML models. We trained an anomaly detection model with the nonclustered microbubbles (MBs) and tested the model on the CCMCs to isolate the unique acoustics. We also had a separate set of control experiments that was performed to validate the anomaly detection model.

Results: The anomaly detection model successfully identified frames exhibiting unique acoustic signatures associated with CCMCs. Frequency domain analysis further confirmed that these frames displayed higher amplitude and energy, suggesting the occurrence of potential coalescence events. The specificity of the model was validated through control experiments, in which both groups contained only individual MBs without clustering. As anticipated, no anomalies were detected in this control dataset, reinforcing the model's ability to distinguish clustered MBs from nonclustered ones.

Conclusions: We highlight the feasibility of detecting and distinguishing the unique acoustic characteristics of CCMCs, thereby improving the detectability and localization of contrast agents in ultrasound imaging. The elevated acoustic amplitudes produced by CCMCs offer potential advantages for more effective contrast agent detection, which is particularly valuable in super-resolution ultrasound imaging. Both the contrast agent and the ML-based analysis approach hold promise for a wide range of applications.

目的:我们的目标是利用机器学习(ML)技术,特别是基于自动编码器的异常检测模型,研究和分离化学交联微泡团簇(CCMCs)产生的独特声学特性。方法:通过无铜点击化学合成ccmc,并使用临床换能器进行声学分析。射频数据被采集、处理并组织成ML模型的训练和测试数据集。我们用非聚类微气泡(mb)训练了一个异常检测模型,并在ccmc上测试了该模型,以隔离独特的声学。我们还进行了一组单独的控制实验来验证异常检测模型。结果:异常检测模型成功地识别出与ccmc相关的具有独特声学特征的帧。频域分析进一步证实,这些帧显示出更高的振幅和能量,表明存在潜在的聚并事件。通过对照实验验证了模型的特异性,在对照实验中,两组均仅包含单个MBs,未聚类。正如预期的那样,在这个控制数据集中没有检测到异常,这加强了模型区分集群mb和非集群mb的能力。结论:我们强调了检测和区分ccmc独特声学特征的可行性,从而提高了造影剂在超声成像中的可检测性和定位性。ccmc产生的声振幅升高为更有效的造影剂检测提供了潜在的优势,这在超分辨率超声成像中特别有价值。造影剂和基于ml的分析方法都有广泛的应用前景。
{"title":"Harnessing chemically crosslinked microbubble clusters using deep learning for ultrasound contrast imaging.","authors":"Teja Pathour, Ghazal Rastegar, Shashank R Sirsi, Baowei Fei","doi":"10.1117/1.JMI.12.4.047001","DOIUrl":"10.1117/1.JMI.12.4.047001","url":null,"abstract":"<p><strong>Purpose: </strong>We aim to investigate and isolate the distinctive acoustic properties generated by chemically crosslinked microbubble clusters (CCMCs) using machine learning (ML) techniques, specifically using an anomaly detection model based on autoencoders.</p><p><strong>Approach: </strong>CCMCs were synthesized via copper-free click chemistry and subjected to acoustic analysis using a clinical transducer. Radiofrequency data were acquired, processed, and organized into training and testing datasets for the ML models. We trained an anomaly detection model with the nonclustered microbubbles (MBs) and tested the model on the CCMCs to isolate the unique acoustics. We also had a separate set of control experiments that was performed to validate the anomaly detection model.</p><p><strong>Results: </strong>The anomaly detection model successfully identified frames exhibiting unique acoustic signatures associated with CCMCs. Frequency domain analysis further confirmed that these frames displayed higher amplitude and energy, suggesting the occurrence of potential coalescence events. The specificity of the model was validated through control experiments, in which both groups contained only individual MBs without clustering. As anticipated, no anomalies were detected in this control dataset, reinforcing the model's ability to distinguish clustered MBs from nonclustered ones.</p><p><strong>Conclusions: </strong>We highlight the feasibility of detecting and distinguishing the unique acoustic characteristics of CCMCs, thereby improving the detectability and localization of contrast agents in ultrasound imaging. The elevated acoustic amplitudes produced by CCMCs offer potential advantages for more effective contrast agent detection, which is particularly valuable in super-resolution ultrasound imaging. Both the contrast agent and the ML-based analysis approach hold promise for a wide range of applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"047001"},"PeriodicalIF":1.9,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12255354/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144627413","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GRN+: a simplified generative reinforcement network for tissue layer analysis in 3D ultrasound images for chronic low-back pain. GRN+:用于慢性腰痛三维超声图像组织层分析的简化生成强化网络。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-07-01 Epub Date: 2025-07-31 DOI: 10.1117/1.JMI.12.4.044001
Zixue Zeng, Xiaoyan Zhao, Matthew Cartier, Xin Meng, Jiantao Pu

Purpose: 3D ultrasound delivers high-resolution, real-time images of soft tissues, which are essential for pain research. However, manually distinguishing various tissues for quantitative analysis is labor-intensive. We aimed to automate multilayer segmentation in 3D ultrasound volumes using minimal annotated data by developing generative reinforcement network plus (GRN+), a semi-supervised multi-model framework.

Approach: GRN+ integrates a ResNet-based generator and a U-Net segmentation model. Through a method called segmentation-guided enhancement (SGE), the generator produces new images under the guidance of the segmentation model, with its weights adjusted according to the segmentation loss gradient. To prevent gradient explosion and secure stable training, a two-stage backpropagation strategy was implemented: the first stage propagates the segmentation loss through both the generator and segmentation model, whereas the second stage concentrates on optimizing the segmentation model alone, thereby refining mask prediction using the generated images.

Results: Tested on 69 fully annotated 3D ultrasound scans from 29 subjects with six manually labeled tissue layers, GRN+ outperformed all other semi-supervised methods in terms of the Dice coefficient using only 5% labeled data, despite not using unlabeled data for unsupervised training. In addition, when applied to fully annotated datasets, GRN+ with SGE achieved a 2.16% higher Dice coefficient while incurring lower computational costs compared to other models.

Conclusions: GRN+ provides accurate tissue segmentation while reducing both computational expenses and the dependency on extensive annotations, making it an effective tool for 3D ultrasound analysis in patients with chronic lower back pain.

目的:三维超声提供高分辨率、实时的软组织图像,这对疼痛研究至关重要。然而,手工区分各种组织进行定量分析是劳动密集型的。我们的目标是通过开发生成强化网络+ (GRN+),一种半监督多模型框架,使用最少的注释数据,在3D超声体积中自动进行多层分割。方法:GRN+集成了基于resnet的生成器和U-Net分割模型。通过一种称为分割引导增强(SGE)的方法,生成器在分割模型的指导下生成新图像,并根据分割损失梯度调整其权重。为了防止梯度爆炸和保证训练的稳定性,采用了两阶段反向传播策略:第一阶段通过生成器和分割模型传播分割损失,而第二阶段专注于单独优化分割模型,从而利用生成的图像改进掩码预测。结果:对来自29名受试者的69个完全注释的3D超声扫描进行了测试,其中包含6个手动标记的组织层,尽管没有使用未标记的数据进行无监督训练,但仅使用5%的标记数据,GRN+在Dice系数方面优于所有其他半监督方法。此外,当应用于完全注释的数据集时,与其他模型相比,具有SGE的GRN+的Dice系数提高了2.16%,而计算成本更低。结论:GRN+提供了准确的组织分割,同时减少了计算费用和对大量注释的依赖,使其成为慢性下背痛患者三维超声分析的有效工具。
{"title":"GRN+: a simplified generative reinforcement network for tissue layer analysis in 3D ultrasound images for chronic low-back pain.","authors":"Zixue Zeng, Xiaoyan Zhao, Matthew Cartier, Xin Meng, Jiantao Pu","doi":"10.1117/1.JMI.12.4.044001","DOIUrl":"10.1117/1.JMI.12.4.044001","url":null,"abstract":"<p><strong>Purpose: </strong>3D ultrasound delivers high-resolution, real-time images of soft tissues, which are essential for pain research. However, manually distinguishing various tissues for quantitative analysis is labor-intensive. We aimed to automate multilayer segmentation in 3D ultrasound volumes using minimal annotated data by developing generative reinforcement network plus (GRN+), a semi-supervised multi-model framework.</p><p><strong>Approach: </strong>GRN+ integrates a ResNet-based generator and a U-Net segmentation model. Through a method called segmentation-guided enhancement (SGE), the generator produces new images under the guidance of the segmentation model, with its weights adjusted according to the segmentation loss gradient. To prevent gradient explosion and secure stable training, a two-stage backpropagation strategy was implemented: the first stage propagates the segmentation loss through both the generator and segmentation model, whereas the second stage concentrates on optimizing the segmentation model alone, thereby refining mask prediction using the generated images.</p><p><strong>Results: </strong>Tested on 69 fully annotated 3D ultrasound scans from 29 subjects with six manually labeled tissue layers, GRN+ outperformed all other semi-supervised methods in terms of the Dice coefficient using only 5% labeled data, despite not using unlabeled data for unsupervised training. In addition, when applied to fully annotated datasets, GRN+ with SGE achieved a 2.16% higher Dice coefficient while incurring lower computational costs compared to other models.</p><p><strong>Conclusions: </strong>GRN+ provides accurate tissue segmentation while reducing both computational expenses and the dependency on extensive annotations, making it an effective tool for 3D ultrasound analysis in patients with chronic lower back pain.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"044001"},"PeriodicalIF":1.7,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12310559/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144761789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Influence of phantom design on evaluation metrics in photon counting spectral head CT: a simulation study. 光体设计对光子计数谱头CT评价指标影响的模拟研究。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-07-01 Epub Date: 2025-07-12 DOI: 10.1117/1.JMI.12.4.043501
Bahaa Ghammraoui, Mridul Bhattarai, Harsha Marupudi, Stephen J Glick

Purpose: Accurate iodine quantification in contrast-enhanced head CT is crucial for precise diagnosis and treatment planning. Traditional CT methods, which use energy-integrating detectors and dual-exposure techniques for material discrimination, often increase patient radiation exposure and are susceptible to motion artifacts and spectral resolution loss. Photon counting detectors (PCDs), capable of acquiring multiple energy windows in a single exposure with superior energy resolution, offer a promising alternative. However, the adoption of these technological advancements requires corresponding developments in evaluation methodologies to ensure their safe and effective implementation. One critical area of concern is the accuracy of iodine quantification, which is commonly assessed using cylindrical phantoms that neither replicate the shape of the human head nor incorporate skull-mimicking materials. These phantoms are widely used not only for testing but also for calibration, which may contribute to an overestimation of system performance in clinical applications. We address the impact of phantom design on evaluation metrics in spectral head CT, comparing conventional cylindrical phantoms to anatomically realistic elliptical phantoms with skull simulants.

Approach: We conducted simulations using a photon-counting spectral CT system equipped with cadmium telluride (CdTe) detectors, utilizing the Photon Counting Toolkit and Tigre CT software for detector response and CT geometry simulations. We compared cylindrical phantoms (20 cm diameter) to elliptical phantoms in three different sizes, incorporating skull materials with major/minor diameters and skull thicknesses of 18/14/0.5, 20/16/0.6, and 23/18/0.7 cm. Iodine inserts at concentrations of 0, 2, 5, and 10    mg / mL with diameters of 1, 0.5, and 0.3 cm were used. We evaluated the influence of bowtie filters, various tube currents, and operating voltages. Image reconstruction was performed after beam hardening correction using the signal-to-thickness calibration (STC) method with standard filtered back projection, followed by both image-based and projection-based material decomposition.

Results: The results showed that image-based methods were more sensitive to phantom design, with cylindrical phantoms exhibiting enhanced performance compared with anatomically realistic designs across key metrics, including systematic error, root mean square error (RMSE), and precision. By contrast, the projection-based material decomposition method demonstrated greater consistency across different phantom designs and improved accuracy and precision. This highlights its potential for more reliable iodine quantification in complex geometries.

Conclusions: These findings underscore the critical importance of phantom design, especially the inclusion

目的:增强头部CT碘定量对准确诊断和制定治疗方案至关重要。传统的CT方法使用能量积分检测器和双曝光技术进行材料识别,通常会增加患者的辐射暴露,并且容易产生运动伪影和光谱分辨率损失。光子计数探测器(PCDs)能够在单次曝光中以优异的能量分辨率获取多个能量窗口,提供了一个有希望的替代方案。然而,采用这些技术进步需要在评价方法方面作出相应的发展,以确保其安全和有效的执行。一个值得关注的关键领域是碘定量的准确性,通常使用圆柱形的模型进行评估,这些模型既不能复制人类头部的形状,也不能使用模拟头骨的材料。这些幻影不仅广泛用于测试,而且还用于校准,这可能导致在临床应用中对系统性能的高估。我们讨论了幽灵设计对光谱头部CT评估指标的影响,比较了传统的圆柱形幽灵与颅骨模拟的解剖学逼真的椭圆形幽灵。方法:我们使用配备碲化镉(CdTe)探测器的光子计数光谱CT系统进行模拟,利用光子计数工具包和Tigre CT软件进行探测器响应和CT几何模拟。我们比较了三种不同尺寸的圆柱形幻影(直径20 cm)和椭圆形幻影,颅骨材料的主要/小直径和颅骨厚度分别为18/14/0.5、20/16/0.6和23/18/0.7 cm。碘插入物浓度分别为0、2、5和10 mg / mL,直径分别为1、0.5和0.3 cm。我们评估了领结滤波器、各种管电流和工作电压的影响。在使用标准滤波后投影的信号-厚度校准(STC)方法进行光束硬化校正后,进行图像重建,然后进行基于图像和基于投影的材料分解。结果:结果表明,基于图像的方法对模型设计更敏感,在系统误差、均方根误差(RMSE)和精度等关键指标上,圆柱形模型比解剖学真实的模型表现出更高的性能。相比之下,基于投影的材料分解方法在不同的模体设计中表现出更大的一致性,并提高了精度和精度。这突出了它在复杂几何形状中更可靠的碘定量的潜力。结论:这些发现强调了假体设计的重要性,特别是在定量结果评估中包含模拟颅骨的材料。通常用于校准和测试的圆柱形幻影,由于其简化的几何形状,可能会高估头部CT碘定量的性能。我们强调需要采用解剖学上真实的幻像设计,例如带有颅骨模拟物的椭圆幻像,以便对光谱光子计数头部CT系统进行更具临床相关性和准确性的评估。
{"title":"Influence of phantom design on evaluation metrics in photon counting spectral head CT: a simulation study.","authors":"Bahaa Ghammraoui, Mridul Bhattarai, Harsha Marupudi, Stephen J Glick","doi":"10.1117/1.JMI.12.4.043501","DOIUrl":"https://doi.org/10.1117/1.JMI.12.4.043501","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate iodine quantification in contrast-enhanced head CT is crucial for precise diagnosis and treatment planning. Traditional CT methods, which use energy-integrating detectors and dual-exposure techniques for material discrimination, often increase patient radiation exposure and are susceptible to motion artifacts and spectral resolution loss. Photon counting detectors (PCDs), capable of acquiring multiple energy windows in a single exposure with superior energy resolution, offer a promising alternative. However, the adoption of these technological advancements requires corresponding developments in evaluation methodologies to ensure their safe and effective implementation. One critical area of concern is the accuracy of iodine quantification, which is commonly assessed using cylindrical phantoms that neither replicate the shape of the human head nor incorporate skull-mimicking materials. These phantoms are widely used not only for testing but also for calibration, which may contribute to an overestimation of system performance in clinical applications. We address the impact of phantom design on evaluation metrics in spectral head CT, comparing conventional cylindrical phantoms to anatomically realistic elliptical phantoms with skull simulants.</p><p><strong>Approach: </strong>We conducted simulations using a photon-counting spectral CT system equipped with cadmium telluride (CdTe) detectors, utilizing the Photon Counting Toolkit and Tigre CT software for detector response and CT geometry simulations. We compared cylindrical phantoms (20 cm diameter) to elliptical phantoms in three different sizes, incorporating skull materials with major/minor diameters and skull thicknesses of 18/14/0.5, 20/16/0.6, and 23/18/0.7 cm. Iodine inserts at concentrations of 0, 2, 5, and <math><mrow><mn>10</mn> <mtext>  </mtext> <mi>mg</mi> <mo>/</mo> <mi>mL</mi></mrow> </math> with diameters of 1, 0.5, and 0.3 cm were used. We evaluated the influence of bowtie filters, various tube currents, and operating voltages. Image reconstruction was performed after beam hardening correction using the signal-to-thickness calibration (STC) method with standard filtered back projection, followed by both image-based and projection-based material decomposition.</p><p><strong>Results: </strong>The results showed that image-based methods were more sensitive to phantom design, with cylindrical phantoms exhibiting enhanced performance compared with anatomically realistic designs across key metrics, including systematic error, root mean square error (RMSE), and precision. By contrast, the projection-based material decomposition method demonstrated greater consistency across different phantom designs and improved accuracy and precision. This highlights its potential for more reliable iodine quantification in complex geometries.</p><p><strong>Conclusions: </strong>These findings underscore the critical importance of phantom design, especially the inclusion","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 4","pages":"043501"},"PeriodicalIF":1.9,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12254834/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144627414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Medical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1