首页 > 最新文献

Journal of Medical Imaging最新文献

英文 中文
Deep-learning-based estimation of left ventricle myocardial strain from echocardiograms with occlusion artifacts. 基于深度学习的超声心动图左心室心肌应变估计。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-09-27 DOI: 10.1117/1.JMI.12.5.054002
Alan Romero-Pacheco, Nidiyare Hevia-Montiel, Blanca Vazquez, Fernando Arámbula Cosío, Jorge Perez-Gonzalez

Purpose: We present a deep-learning-based methodology for estimating deformation in 2D echocardiograms. The goal is to automatically estimate the longitudinal strain of the left ventricle (LV) walls in images affected by speckle noise and acoustic occlusions.

Approach: The proposed methodology integrates algorithms for converting sparse to dense flow, a Res-UNet architecture for automatic myocardium segmentation, flow estimation using a global motion aggregation network, and the computation of longitudinal strain curves and the global longitudinal strain (GLS) index. The approach was evaluated using two echocardiographic datasets in apical four-chamber view, both modified with noise and acoustic shadows. The CAMUS dataset ( N = 250 ) was used for LV wall segmentation, whereas a synthetic image database ( N = 2037 ) was employed for flow estimation.

Results: Among the main performance metrics achieved are 98% [96 to 99] of correlation in the conversion from sparse to dense flow, a Dice index of 88.2 % ± 3.8 % for myocardial segmentation, an endpoint error of 0.133 [0.13 to 0.14] pixels in flow estimation, and an error of 1.34% [0.94 to 2.09] in the estimation of the GLS index.

Conclusions: The results demonstrate improvements over previously reported performances while maintaining stability in echocardiograms with acoustic shadows. This methodology could be useful in clinical practice for the analysis of echocardiograms with noise artifacts and acoustic occlusions. Our code and trained models are publicly available at https://github.com/ArBioIIMAS/echo-gma.

目的:我们提出了一种基于深度学习的方法来估计二维超声心动图的变形。目标是在受斑点噪声和声闭塞影响的图像中自动估计左心室(LV)壁的纵向应变。方法:该方法集成了将稀疏流转换为密集流的算法,用于自动心肌分割的Res-UNet架构,使用全局运动聚集网络的流量估计,以及纵向应变曲线和全局纵向应变(GLS)指数的计算。该方法使用两个超声心动图数据集在根尖四室视图中进行评估,都使用噪声和声学阴影进行修改。CAMUS数据集(N = 250)用于左室壁分割,而合成图像数据库(N = 2037)用于流量估计。结果:在实现的主要性能指标中,从稀疏到密集流转换的相关性为98%[96至99],心肌分割的Dice指数为89.2%±3.8%,流量估计的终点误差为0.133[0.13至0.14]像素,GLS指数估计的误差为1.34%[0.94至2.09]。结论:结果表明,在保持超声心动图稳定性的同时,比先前报道的性能有所改善。该方法可用于临床超声心动图噪声伪影和声学阻塞的分析。我们的代码和训练过的模型可以在https://github.com/ArBioIIMAS/echo-gma上公开获得。
{"title":"Deep-learning-based estimation of left ventricle myocardial strain from echocardiograms with occlusion artifacts.","authors":"Alan Romero-Pacheco, Nidiyare Hevia-Montiel, Blanca Vazquez, Fernando Arámbula Cosío, Jorge Perez-Gonzalez","doi":"10.1117/1.JMI.12.5.054002","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.054002","url":null,"abstract":"<p><strong>Purpose: </strong>We present a deep-learning-based methodology for estimating deformation in 2D echocardiograms. The goal is to automatically estimate the longitudinal strain of the left ventricle (LV) walls in images affected by speckle noise and acoustic occlusions.</p><p><strong>Approach: </strong>The proposed methodology integrates algorithms for converting sparse to dense flow, a Res-UNet architecture for automatic myocardium segmentation, flow estimation using a global motion aggregation network, and the computation of longitudinal strain curves and the global longitudinal strain (GLS) index. The approach was evaluated using two echocardiographic datasets in apical four-chamber view, both modified with noise and acoustic shadows. The CAMUS dataset ( <math><mrow><mi>N</mi> <mo>=</mo> <mn>250</mn></mrow> </math> ) was used for LV wall segmentation, whereas a synthetic image database ( <math><mrow><mi>N</mi> <mo>=</mo> <mn>2037</mn></mrow> </math> ) was employed for flow estimation.</p><p><strong>Results: </strong>Among the main performance metrics achieved are 98% [96 to 99] of correlation in the conversion from sparse to dense flow, a Dice index of <math><mrow><mn>88.2</mn> <mo>%</mo> <mo>±</mo> <mn>3.8</mn> <mo>%</mo></mrow> </math> for myocardial segmentation, an endpoint error of 0.133 [0.13 to 0.14] pixels in flow estimation, and an error of 1.34% [0.94 to 2.09] in the estimation of the GLS index.</p><p><strong>Conclusions: </strong>The results demonstrate improvements over previously reported performances while maintaining stability in echocardiograms with acoustic shadows. This methodology could be useful in clinical practice for the analysis of echocardiograms with noise artifacts and acoustic occlusions. Our code and trained models are publicly available at https://github.com/ArBioIIMAS/echo-gma.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"054002"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12476231/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145187211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving personalized federated learning to optimize site-specific performance in computer-aided detection/diagnosis. 改进个性化联合学习,优化计算机辅助检测/诊断中特定站点的性能。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-22 DOI: 10.1117/1.JMI.12.5.054503
Aiki Yamada, Shouhei Hanaoka, Tomomi Takenaga, Soichiro Miki, Takeharu Yoshikawa, Osamu Abe, Toshiya Nakaguchi, Yukihiro Nomura

Purpose: Personalized federated learning (PFL) has been explored to address data heterogeneity while preserving privacy, and its application in computer-aided detection/diagnosis (CAD) software has been investigated. Ditto, a commonly studied PFL method, trains global and personalized models but is limited by instability in model updates and high hyperparameter tuning costs. We proposed Improved Ditto, a PFL method that dynamically adjusts the proportion of global model weights during personalized model updates to enhance stability and reduce hyperparameter tuning costs.

Approach: We introduced a personalized model update rule in Improved Ditto that dynamically determines the proportion of global model weights based on the L2-norm of the gradient-derived and global-model-derived terms. This method was evaluated using three types of CAD software: cerebral aneurysm detection in magnetic resonance (MR) angiography images (segmentation), brain metastasis detection in contrast-enhanced T1-weighted MR images (object detection), and liver lesion classification in gadolinium-ethoxybenzyl-diethylenetriamine pentaacetic acid-enhanced MR images (classification). The proposed method was compared with several conventional methods.

Results: In two out of three CAD software, the performance of Improved Ditto was competitive with Ditto and other federated-learning-based methods. The proposed method achieved a narrower hyperparameter search space, which contributed to reducing the tuning costs. In addition, it improved the stability of personalized model updates, suggesting enhanced adaptability to diverse datasets and tasks.

Conclusions: We demonstrate that dynamically adjusting global model weights during personalized model updates can improve the stability and adaptability of PFL. The proposed method reduces the hyperparameter tuning costs and offers potential benefits for CAD software.

目的:探讨个性化联邦学习(PFL)在保护隐私的同时解决数据异质性问题,并研究其在计算机辅助检测/诊断(CAD)软件中的应用。Ditto是一种常用的PFL方法,它训练全局和个性化模型,但受模型更新不稳定和高超参数调整成本的限制。我们提出了一种PFL方法Improved Ditto,该方法在个性化模型更新过程中动态调整全局模型权重的比例,以提高稳定性并降低超参数调整成本。方法:我们在改进同上中引入了个性化的模型更新规则,该规则基于梯度衍生项和全局模型衍生项的l2范数动态确定全局模型权重的比例。使用三种CAD软件对该方法进行评估:磁共振(MR)血管造影图像中的脑动脉瘤检测(分割),对比增强t1加权MR图像中的脑转移检测(目标检测),钆-乙氧基苄基-二乙烯三胺五乙酸增强MR图像中的肝脏病变分类(分类)。将该方法与几种传统方法进行了比较。结果:在三个CAD软件中的两个中,改进的Ditto的性能与Ditto和其他基于联邦学习的方法具有竞争力。该方法实现了更窄的超参数搜索空间,有助于降低调优成本。此外,它提高了个性化模型更新的稳定性,增强了对不同数据集和任务的适应性。结论:在个性化模型更新过程中动态调整全局模型权值可以提高PFL的稳定性和自适应性。该方法降低了超参数整定成本,为CAD软件提供了潜在的优势。
{"title":"Improving personalized federated learning to optimize site-specific performance in computer-aided detection/diagnosis.","authors":"Aiki Yamada, Shouhei Hanaoka, Tomomi Takenaga, Soichiro Miki, Takeharu Yoshikawa, Osamu Abe, Toshiya Nakaguchi, Yukihiro Nomura","doi":"10.1117/1.JMI.12.5.054503","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.054503","url":null,"abstract":"<p><strong>Purpose: </strong>Personalized federated learning (PFL) has been explored to address data heterogeneity while preserving privacy, and its application in computer-aided detection/diagnosis (CAD) software has been investigated. Ditto, a commonly studied PFL method, trains global and personalized models but is limited by instability in model updates and high hyperparameter tuning costs. We proposed Improved Ditto, a PFL method that dynamically adjusts the proportion of global model weights during personalized model updates to enhance stability and reduce hyperparameter tuning costs.</p><p><strong>Approach: </strong>We introduced a personalized model update rule in Improved Ditto that dynamically determines the proportion of global model weights based on the L2-norm of the gradient-derived and global-model-derived terms. This method was evaluated using three types of CAD software: cerebral aneurysm detection in magnetic resonance (MR) angiography images (segmentation), brain metastasis detection in contrast-enhanced T1-weighted MR images (object detection), and liver lesion classification in gadolinium-ethoxybenzyl-diethylenetriamine pentaacetic acid-enhanced MR images (classification). The proposed method was compared with several conventional methods.</p><p><strong>Results: </strong>In two out of three CAD software, the performance of Improved Ditto was competitive with Ditto and other federated-learning-based methods. The proposed method achieved a narrower hyperparameter search space, which contributed to reducing the tuning costs. In addition, it improved the stability of personalized model updates, suggesting enhanced adaptability to diverse datasets and tasks.</p><p><strong>Conclusions: </strong>We demonstrate that dynamically adjusting global model weights during personalized model updates can improve the stability and adaptability of PFL. The proposed method reduces the hyperparameter tuning costs and offers potential benefits for CAD software.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"054503"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12543029/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145355763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DABS-MS: deep atlas-based segmentation using the Mumford-Shah functional. dads - ms:使用Mumford-Shah函数的基于深度图谱的分割。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-21 DOI: 10.1117/1.JMI.12.5.055002
Hannah G Mason, Jack H Noble

Purpose: Cochlear implants (CIs) are neural prosthetics used to treat patients with severe-to-profound hearing loss. Patient-specific modeling of CI stimulation of the auditory nerve fiber (ANF) can help audiologists improve the CI programming. These models require localization of the ANFs relative to the surrounding anatomy and the CI. Localization is challenging because the ANFs are so small that they are not directly visible in clinical imaging. We hypothesize that the position of the ANFs can be accurately inferred from the location of the internal auditory canal (IAC), which has high contrast in CT because the ANFs pass through this canal between the cochlea and the brain.

Approach: Inspired by VoxelMorph, we propose a deep atlas-based IAC segmentation network. We create a single atlas in which the IAC and ANFs are pre-localized. Our network is trained to produce deformation fields (DFs) mapping coordinates from the atlas to new target volumes and that accurately segment the IAC. We hypothesize that DFs that accurately segment the IAC in target images will also facilitate accurate atlas-based localization of the ANFs. As opposed to VoxelMorph, which aims to produce DFs that accurately register the entire volume, our contribution is an entirely self-supervised training scheme that aims to produce DFs that accurately segment the target structure. This self-supervision is facilitated using a loss function inspired by the Mumford-Shah functional. We call our method Deep Atlas-Based Segmentation using Mumford-Shah (DABS-MS).

Results: Results show that DABS-MS outperforms VoxelMorph for IAC segmentation. Tests with publicly available datasets for trachea and kidney segmentation also show significant improvement in segmentation accuracy, demonstrating the generalizability of the method.

Conclusions: Our proposed DABS-MS method can accurately segment the IAC, which can then facilitate the localization of the ANFs. This patient-specific modeling of CI stimulation of the ANFs can help audiologists improve the CI programming, leading to better outcomes for patients with severe-to-profound hearing loss.

目的:人工耳蜗是一种用于治疗重度至重度听力损失患者的神经义肢。听觉神经纤维(ANF)的CI刺激的患者特异性建模可以帮助听力学家改进CI编程。这些模型需要定位相对于周围解剖结构和CI的anf。定位具有挑战性,因为anf很小,在临床成像中不能直接看到。我们假设anf的位置可以准确地从内耳道(IAC)的位置推断出来,内耳道在CT上具有高对比度,因为anf通过耳蜗和大脑之间的内耳道。方法:受VoxelMorph的启发,我们提出了一种基于深度地图集的IAC分割网络。我们创建了一个单一的图谱,其中IAC和anf是预先定位的。我们的网络被训练来产生从地图集到新目标体的变形场(df)映射坐标,并准确地分割IAC。我们假设,在目标图像中准确分割IAC的DFs也将有助于准确的基于地图集的anf定位。与VoxelMorph相反,VoxelMorph旨在生成准确注册整个体积的df,我们的贡献是一个完全自监督的训练方案,旨在生成准确分割目标结构的df。这种自我监督是由Mumford-Shah泛函启发的损失函数促进的。我们称我们的方法为使用Mumford-Shah (DABS-MS)的基于深度地图集的分割。结果:DABS-MS在IAC分割上优于VoxelMorph。使用公开可用的气管和肾脏分割数据集进行的测试也显示出分割精度的显着提高,证明了该方法的可泛化性。结论:DABS-MS方法可以准确地分割IAC,从而有助于anf的定位。这种对anf的CI刺激的患者特异性建模可以帮助听力学家改进CI编程,从而为重度到重度听力损失的患者带来更好的结果。
{"title":"DABS-MS: deep atlas-based segmentation using the Mumford-Shah functional.","authors":"Hannah G Mason, Jack H Noble","doi":"10.1117/1.JMI.12.5.055002","DOIUrl":"10.1117/1.JMI.12.5.055002","url":null,"abstract":"<p><strong>Purpose: </strong>Cochlear implants (CIs) are neural prosthetics used to treat patients with severe-to-profound hearing loss. Patient-specific modeling of CI stimulation of the auditory nerve fiber (ANF) can help audiologists improve the CI programming. These models require localization of the ANFs relative to the surrounding anatomy and the CI. Localization is challenging because the ANFs are so small that they are not directly visible in clinical imaging. We hypothesize that the position of the ANFs can be accurately inferred from the location of the internal auditory canal (IAC), which has high contrast in CT because the ANFs pass through this canal between the cochlea and the brain.</p><p><strong>Approach: </strong>Inspired by VoxelMorph, we propose a deep atlas-based IAC segmentation network. We create a single atlas in which the IAC and ANFs are pre-localized. Our network is trained to produce deformation fields (DFs) mapping coordinates from the atlas to new target volumes and that accurately segment the IAC. We hypothesize that DFs that accurately segment the IAC in target images will also facilitate accurate atlas-based localization of the ANFs. As opposed to VoxelMorph, which aims to produce DFs that accurately register the entire volume, our contribution is an entirely self-supervised training scheme that aims to produce DFs that accurately segment the target structure. This self-supervision is facilitated using a loss function inspired by the Mumford-Shah functional. We call our method Deep Atlas-Based Segmentation using Mumford-Shah (DABS-MS).</p><p><strong>Results: </strong>Results show that DABS-MS outperforms VoxelMorph for IAC segmentation. Tests with publicly available datasets for trachea and kidney segmentation also show significant improvement in segmentation accuracy, demonstrating the generalizability of the method.</p><p><strong>Conclusions: </strong>Our proposed DABS-MS method can accurately segment the IAC, which can then facilitate the localization of the ANFs. This patient-specific modeling of CI stimulation of the ANFs can help audiologists improve the CI programming, leading to better outcomes for patients with severe-to-profound hearing loss.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"055002"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12539791/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145349288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Medical Image Perception and Quality Assessment Through Technology and Human Factors Research. 通过技术与人因研究推进医学图像感知与质量评价。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-23 DOI: 10.1117/1.JMI.12.5.051801
Frank Tong, Elizabeth A Krupinski

The editorial introduces the Special Section on Medical Image Perception and Observer Performance for JMI Volume 12 Issue 5.

该社论介绍了JMI第12卷第5期的医学图像感知和观察者表现特别部分。
{"title":"Advancing Medical Image Perception and Quality Assessment Through Technology and Human Factors Research.","authors":"Frank Tong, Elizabeth A Krupinski","doi":"10.1117/1.JMI.12.5.051801","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.051801","url":null,"abstract":"<p><p>The editorial introduces the Special Section on Medical Image Perception and Observer Performance for JMI Volume 12 Issue 5.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"051801"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12547280/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145372992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Analysis of intra- and inter-observer variability in 4D liver ultrasound landmark labeling. 四维肝脏超声标志物标记的观察者间和观察者内变异分析。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-06-30 DOI: 10.1117/1.JMI.12.5.051807
Daniel Wulff, Floris Ernst

Purpose: Four-dimensional (4D) ultrasound imaging is widely used in clinics for diagnostics and therapy guidance. Accurate target tracking in 4D ultrasound is crucial for autonomous therapy guidance systems, such as radiotherapy, where precise tumor localization ensures effective treatment. Supervised deep learning approaches rely on reliable ground truth, making accurate labels essential. We investigate the reliability of expert-labeled ground truth data by evaluating intra- and inter-observer variability in landmark labeling for 4D ultrasound imaging in the liver.

Approach: Eight 4D liver ultrasound sequences were labeled by eight expert observers, each labeling eight landmarks three times. Intra- and inter-observer variability was quantified, and observer survey and motion analysis were conducted to determine factors influencing labeling accuracy, such as ultrasound artifacts and motion amplitude.

Results: The mean intra-observer variability ranged from 1.58    mm ± 0.90    mm to 2.05    mm ± 1.22    mm depending on the observer. The inter-observer variability for the two observer groups was 2.68    mm ± 1.69    mm and 3.06    mm ± 1.74    mm . The observer survey and motion analysis revealed that ultrasound artifacts significantly affected labeling accuracy due to limited landmark visibility, whereas motion amplitude had no measurable effect. Our measured mean landmark motion was 11.56    mm ± 5.86    mm .

Conclusions: We highlight variability in expert-labeled ground truth data for 4D ultrasound imaging and identify ultrasound artifacts as a major source of labeling inaccuracies. These findings underscore the importance of addressing observer variability and artifact-related challenges to improve the reliability of ground truth data for evaluating target tracking algorithms in 4D ultrasound applications.

目的:四维超声成像广泛应用于临床诊断和治疗指导。在四维超声中精确的目标跟踪对于自主治疗引导系统至关重要,例如放射治疗,其中精确的肿瘤定位确保有效治疗。有监督的深度学习方法依赖于可靠的真实情况,因此准确的标签至关重要。我们通过评估肝脏四维超声成像中地标标记的观察者内部和观察者之间的可变性来研究专家标记的真实数据的可靠性。方法:8个4D肝脏超声序列由8名专家观察者标记,每个标记3次,标记8个地标。量化观察者内部和观察者之间的可变性,并进行观察者调查和运动分析,以确定影响标记准确性的因素,如超声伪影和运动幅度。结果:根据观察者的不同,观察者内部的平均变异范围为1.58 mm±0.90 mm至2.05 mm±1.22 mm。两个观察组的观察者间变异分别为2.68 mm±1.69 mm和3.06 mm±1.74 mm。观察者调查和运动分析显示,由于地标可见性有限,超声伪影显著影响标记准确性,而运动幅度没有可测量的影响。我们测量到的平均地标运动为11.56 mm±5.86 mm。结论:我们强调了专家标记的四维超声成像地面真值数据的可变性,并确定超声伪影是标记不准确的主要来源。这些发现强调了解决观测者可变性和伪影相关挑战的重要性,以提高评估4D超声应用中目标跟踪算法的地面真实数据的可靠性。
{"title":"Analysis of intra- and inter-observer variability in 4D liver ultrasound landmark labeling.","authors":"Daniel Wulff, Floris Ernst","doi":"10.1117/1.JMI.12.5.051807","DOIUrl":"10.1117/1.JMI.12.5.051807","url":null,"abstract":"<p><strong>Purpose: </strong>Four-dimensional (4D) ultrasound imaging is widely used in clinics for diagnostics and therapy guidance. Accurate target tracking in 4D ultrasound is crucial for autonomous therapy guidance systems, such as radiotherapy, where precise tumor localization ensures effective treatment. Supervised deep learning approaches rely on reliable ground truth, making accurate labels essential. We investigate the reliability of expert-labeled ground truth data by evaluating intra- and inter-observer variability in landmark labeling for 4D ultrasound imaging in the liver.</p><p><strong>Approach: </strong>Eight 4D liver ultrasound sequences were labeled by eight expert observers, each labeling eight landmarks three times. Intra- and inter-observer variability was quantified, and observer survey and motion analysis were conducted to determine factors influencing labeling accuracy, such as ultrasound artifacts and motion amplitude.</p><p><strong>Results: </strong>The mean intra-observer variability ranged from <math><mrow><mn>1.58</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>0.90</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> to <math><mrow><mn>2.05</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>1.22</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> depending on the observer. The inter-observer variability for the two observer groups was <math><mrow><mn>2.68</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>1.69</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> and <math><mrow><mn>3.06</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>1.74</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> . The observer survey and motion analysis revealed that ultrasound artifacts significantly affected labeling accuracy due to limited landmark visibility, whereas motion amplitude had no measurable effect. Our measured mean landmark motion was <math><mrow><mn>11.56</mn> <mtext>  </mtext> <mi>mm</mi> <mo>±</mo> <mn>5.86</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> .</p><p><strong>Conclusions: </strong>We highlight variability in expert-labeled ground truth data for 4D ultrasound imaging and identify ultrasound artifacts as a major source of labeling inaccuracies. These findings underscore the importance of addressing observer variability and artifact-related challenges to improve the reliability of ground truth data for evaluating target tracking algorithms in 4D ultrasound applications.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"051807"},"PeriodicalIF":1.9,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12207815/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144545488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluation of algorithmic requirements for clinical application of material decomposition using a multi-layer flat panel detector. 多层平板探测器材料分解临床应用的算法要求评价。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-09-04 DOI: 10.1117/1.JMI.12.5.053501
Jamin Schaefer, Steffen Kappler, Ferdinand Lueck, Ludwig Ritschl, Thomas Weber, Georg Rose

Purpose: The combination of multi-layer flat panel detector (FPDT) X-ray imaging and physics-based material decomposition algorithms allows for the removal of anatomical structures. However, the reliability of these algorithms may be compromised by unaccounted materials or scattered radiation.

Approach: We investigated the two-material decomposition performance of a multi-layer FPDT in the context of 2D chest radiography without and with a 13:1 anti-scatter grid employed. A matrix-based material decomposition (MBMD) (equivalent to weighted logarithmic subtraction), a matrix-based material decomposition with polynomial beam hardening pre-correction (MBMD-PBC), and a projection domain decomposition were evaluated. The decomposition accuracy of simulated data was evaluated by comparing the bone and soft tissue images to the ground truth using the structural similarity index measure (SSIM). Simulation results were supported by experiments using a commercially available triple-layer FPDT retrofitted to a digital X-ray system.

Results: Independent of the selected decomposition algorithm, uncorrected scatter leads to negative bone estimates, resulting in small SSIM values and bone structures to remain visible in soft tissue images. Even with a 13:1 anti-scatter grid employed, bone images continue to show negative bone estimates, and bone structures appear in soft tissue images. Adipose tissue on the contrary has an almost negligible effect.

Conclusions: In a contact scan, scattered radiation leads to negative bone contrast estimates in the bone images and remaining bone contrast in the soft tissue images. Therefore, accurate scatter estimation and correction algorithms are essential when aiming for material decomposition using image data obtained with a multi-layer FPDT.

目的:多层平板探测器(FPDT) x射线成像和基于物理的材料分解算法相结合,可以去除解剖结构。然而,这些算法的可靠性可能会受到不明材料或散射辐射的影响。方法:我们研究了多层FPDT在不使用和使用13:1抗散射网格的二维胸片背景下的双材料分解性能。评估了基于矩阵的材料分解(MBMD)(相当于加权对数减法)、基于矩阵的材料分解与多项式光束硬化预校正(MBMD- pbc)以及投影域分解。利用结构相似指数度量(SSIM)将骨骼和软组织图像与地面真实情况进行比较,评估模拟数据的分解精度。仿真结果得到了商用三层FPDT改造成数字x射线系统的实验的支持。结果:与所选择的分解算法无关,未校正的散点会导致骨骼估计为负,导致软组织图像中SSIM值较小,骨骼结构仍然可见。即使采用13:1的反散射网格,骨骼图像仍然显示负骨估计,并且骨骼结构出现在软组织图像中。相反,脂肪组织的影响几乎可以忽略不计。结论:在接触扫描中,散射辐射导致骨骼图像中的骨对比度估计为负,软组织图像中的骨对比度估计为剩余。因此,在利用多层FPDT获得的图像数据进行材料分解时,精确的散射估计和校正算法是必不可少的。
{"title":"Evaluation of algorithmic requirements for clinical application of material decomposition using a multi-layer flat panel detector.","authors":"Jamin Schaefer, Steffen Kappler, Ferdinand Lueck, Ludwig Ritschl, Thomas Weber, Georg Rose","doi":"10.1117/1.JMI.12.5.053501","DOIUrl":"10.1117/1.JMI.12.5.053501","url":null,"abstract":"<p><strong>Purpose: </strong>The combination of multi-layer flat panel detector (FPDT) X-ray imaging and physics-based material decomposition algorithms allows for the removal of anatomical structures. However, the reliability of these algorithms may be compromised by unaccounted materials or scattered radiation.</p><p><strong>Approach: </strong>We investigated the two-material decomposition performance of a multi-layer FPDT in the context of 2D chest radiography without and with a 13:1 anti-scatter grid employed. A matrix-based material decomposition (MBMD) (equivalent to weighted logarithmic subtraction), a matrix-based material decomposition with polynomial beam hardening pre-correction (MBMD-PBC), and a projection domain decomposition were evaluated. The decomposition accuracy of simulated data was evaluated by comparing the bone and soft tissue images to the ground truth using the structural similarity index measure (SSIM). Simulation results were supported by experiments using a commercially available triple-layer FPDT retrofitted to a digital X-ray system.</p><p><strong>Results: </strong>Independent of the selected decomposition algorithm, uncorrected scatter leads to negative bone estimates, resulting in small SSIM values and bone structures to remain visible in soft tissue images. Even with a 13:1 anti-scatter grid employed, bone images continue to show negative bone estimates, and bone structures appear in soft tissue images. Adipose tissue on the contrary has an almost negligible effect.</p><p><strong>Conclusions: </strong>In a contact scan, scattered radiation leads to negative bone contrast estimates in the bone images and remaining bone contrast in the soft tissue images. Therefore, accurate scatter estimation and correction algorithms are essential when aiming for material decomposition using image data obtained with a multi-layer FPDT.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"053501"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12410756/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145015224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Statistical testing of agreement in overlap-based performance between an AI segmentation device and a multi-expert human panel without requiring a reference standard. 在不需要参考标准的情况下,对人工智能分割设备和多专家小组之间基于重叠的性能的一致性进行统计测试。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-22 DOI: 10.1117/1.JMI.12.5.055003
Tingting Hu, Berkman Sahiner, Shuyue Guan, Mike Mikailov, Kenny Cha, Frank Samuelson, Nicholas Petrick

Purpose: Artificial intelligence (AI)-based medical imaging devices often include lesion or organ segmentation capabilities. Existing methods for segmentation performance evaluation compare AI results with an aggregated reference standard using accuracy metrics such as the Dice coefficient or Hausdorff distance. However, these approaches are limited by lacking a gold standard and challenges in defining meaningful success criteria. To address this, we developed a statistical method to assess agreement between an AI device and multiple human experts without requiring a reference standard.

Approach: We propose a paired-testing method to evaluate whether an AI device's segmentation performance significantly differs from that of multiple human experts. The method compares device-to-expert dissimilarity with expert-to-expert dissimilarity, avoiding the need for a reference standard. We validated the method through (1) statistical simulations where the Dice coefficient performance is either shared ("overlap agreeable") or not shared ("overlap disagreeable") between the device and experts; (2) image-based simulations using 2D contours with shared or nonshared transformation parameters (transformation agreeable or disagreeable). We also applied the method to compare an AI segmentation algorithm with four radiologists using data from the Lung Image Database Consortium.

Results: Statistical simulations show the method controls type I error ( 0.05 ) for overlap-agreeable and type II error ( 0 ) for overlap-disagreeable scenarios. Image-based simulations show acceptable performance with a mean type I error of 0.07 (SD 0.03) for transformation-agreeable and a mean type II error of 0.07 (SD 0.18) for transformation-disagreeable cases.

Conclusions: The paired-testing method offers a new tool for assessing the agreement between an AI segmentation device and multiple human expert panelists without requiring a reference standard.

目的:基于人工智能(AI)的医学成像设备通常包括病变或器官分割功能。现有的分割性能评估方法使用精确度指标(如Dice系数或Hausdorff距离)将AI结果与汇总参考标准进行比较。然而,这些方法由于缺乏黄金标准和定义有意义的成功标准的挑战而受到限制。为了解决这个问题,我们开发了一种统计方法来评估人工智能设备和多个人类专家之间的一致性,而不需要参考标准。方法:我们提出了一种配对测试方法来评估人工智能设备的分割性能是否与多个人类专家的分割性能有显著差异。该方法比较了设备与专家之间的不相似性和专家与专家之间的不相似性,避免了对参考标准的需要。我们通过(1)统计模拟验证了该方法,其中Dice系数性能在设备和专家之间共享(“重叠一致”)或不共享(“重叠不一致”);(2)基于图像的模拟,使用具有共享或非共享变换参数的二维轮廓(变换符合或不符合)。我们还应用该方法将人工智能分割算法与四位放射科医生使用肺图像数据库联盟的数据进行比较。结果:统计模拟表明,该方法控制了重叠适宜情景的I型误差(~ 0.05)和重叠不适宜情景的II型误差(~ 0)。基于图像的模拟显示出可接受的性能,对于符合变换的情况,平均I类误差为0.07 (SD 0.03),对于不符合变换的情况,平均II类误差为0.07 (SD 0.18)。结论:配对测试方法为评估人工智能分割设备与多个人类专家小组成员之间的一致性提供了一种新工具,而无需参考标准。
{"title":"Statistical testing of agreement in overlap-based performance between an AI segmentation device and a multi-expert human panel without requiring a reference standard.","authors":"Tingting Hu, Berkman Sahiner, Shuyue Guan, Mike Mikailov, Kenny Cha, Frank Samuelson, Nicholas Petrick","doi":"10.1117/1.JMI.12.5.055003","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.055003","url":null,"abstract":"<p><strong>Purpose: </strong>Artificial intelligence (AI)-based medical imaging devices often include lesion or organ segmentation capabilities. Existing methods for segmentation performance evaluation compare AI results with an aggregated reference standard using accuracy metrics such as the Dice coefficient or Hausdorff distance. However, these approaches are limited by lacking a gold standard and challenges in defining meaningful success criteria. To address this, we developed a statistical method to assess agreement between an AI device and multiple human experts without requiring a reference standard.</p><p><strong>Approach: </strong>We propose a paired-testing method to evaluate whether an AI device's segmentation performance significantly differs from that of multiple human experts. The method compares device-to-expert dissimilarity with expert-to-expert dissimilarity, avoiding the need for a reference standard. We validated the method through (1) statistical simulations where the Dice coefficient performance is either shared (\"overlap agreeable\") or not shared (\"overlap disagreeable\") between the device and experts; (2) image-based simulations using 2D contours with shared or nonshared transformation parameters (transformation agreeable or disagreeable). We also applied the method to compare an AI segmentation algorithm with four radiologists using data from the Lung Image Database Consortium.</p><p><strong>Results: </strong>Statistical simulations show the method controls type I error ( <math><mrow><mo>∼</mo> <mn>0.05</mn></mrow> </math> ) for overlap-agreeable and type II error ( <math><mrow><mo>∼</mo> <mn>0</mn></mrow> </math> ) for overlap-disagreeable scenarios. Image-based simulations show acceptable performance with a mean type I error of 0.07 (SD 0.03) for transformation-agreeable and a mean type II error of 0.07 (SD 0.18) for transformation-disagreeable cases.</p><p><strong>Conclusions: </strong>The paired-testing method offers a new tool for assessing the agreement between an AI segmentation device and multiple human expert panelists without requiring a reference standard.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"055003"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12543030/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356392","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking 3D generative autoencoders for pseudo-healthy reconstruction of brain 18F-fluorodeoxyglucose positron emission tomography. 脑18f -氟脱氧葡萄糖正电子发射断层假健康重建的三维生成式自编码器的基准测试。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-10-16 DOI: 10.1117/1.JMI.12.5.054005
Ravi Hassanaly, Maëlys Solal, Olivier Colliot, Ninon Burgos

Purpose: Many deep generative models have been proposed to reconstruct pseudo-healthy images for anomaly detection. Among these models, the variational autoencoder (VAE) has emerged as both simple and efficient. Although significant progress has been made in refining the VAE within the field of computer vision, these advancements have not been extensively applied to medical imaging applications.

Approach: We present a benchmark that assesses the ability of multiple VAEs to reconstruct pseudo-healthy neuroimages for anomaly detection in the context of dementia. We first propose a rigorous methodology to define the optimal architecture of the vanilla VAE and select through random searches the best hyperparameters of the VAE variants. Relying on a simulation-based evaluation framework, we thoroughly assess the ability of 20 VAE models to reconstruct pseudo-healthy images for the detection of dementia-related anomalies in 3D brain F 18 -fluorodeoxyglucose (FDG) positron emission tomography (PET) and compare their performance.

Results: This benchmark demonstrated that the majority of the VAE models tested were able to reconstruct images of good quality and generate healthy-looking images from simulated images presenting anomalies.

Conclusions: Even if no model clearly outperformed all the others, the benchmark allowed identifying a few models that perform slightly better than the vanilla VAE. It further showed that many VAE-based models can generalize to the detection of anomalies of various intensities, shapes, and locations in 3D brain FDG PET.

目的:提出了许多深度生成模型来重建用于异常检测的伪健康图像。在这些模型中,变分自编码器(VAE)以其简单和高效的特点而出现。尽管在细化计算机视觉领域的VAE方面取得了重大进展,但这些进展尚未广泛应用于医学成像应用。方法:我们提出了一个基准,评估多个VAEs重建伪健康神经图像的能力,以便在痴呆症的背景下进行异常检测。我们首先提出了一种严格的方法来定义香草VAE的最优架构,并通过随机搜索选择VAE变体的最佳超参数。基于基于模拟的评估框架,我们全面评估了20种VAE模型重建伪健康图像的能力,以检测3D脑f18 -氟脱氧葡萄糖(FDG)正电子发射断层扫描(PET)中与痴呆症相关的异常,并比较了它们的性能。结果:该基准测试表明,大多数测试的VAE模型能够重建质量良好的图像,并从呈现异常的模拟图像中生成看起来健康的图像。结论:即使没有模型明显优于所有其他模型,基准测试允许识别一些比普通VAE稍微好一点的模型。进一步表明,许多基于vae的模型可以推广到3D脑FDG PET中各种强度、形状和位置的异常检测。
{"title":"Benchmarking 3D generative autoencoders for pseudo-healthy reconstruction of brain <sup>18</sup>F-fluorodeoxyglucose positron emission tomography.","authors":"Ravi Hassanaly, Maëlys Solal, Olivier Colliot, Ninon Burgos","doi":"10.1117/1.JMI.12.5.054005","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.054005","url":null,"abstract":"<p><strong>Purpose: </strong>Many deep generative models have been proposed to reconstruct pseudo-healthy images for anomaly detection. Among these models, the variational autoencoder (VAE) has emerged as both simple and efficient. Although significant progress has been made in refining the VAE within the field of computer vision, these advancements have not been extensively applied to medical imaging applications.</p><p><strong>Approach: </strong>We present a benchmark that assesses the ability of multiple VAEs to reconstruct pseudo-healthy neuroimages for anomaly detection in the context of dementia. We first propose a rigorous methodology to define the optimal architecture of the vanilla VAE and select through random searches the best hyperparameters of the VAE variants. Relying on a simulation-based evaluation framework, we thoroughly assess the ability of 20 VAE models to reconstruct pseudo-healthy images for the detection of dementia-related anomalies in 3D brain <math> <mrow> <mmultiscripts><mrow><mi>F</mi></mrow> <mprescripts></mprescripts> <none></none> <mrow><mn>18</mn></mrow> </mmultiscripts> </mrow> </math> -fluorodeoxyglucose (FDG) positron emission tomography (PET) and compare their performance.</p><p><strong>Results: </strong>This benchmark demonstrated that the majority of the VAE models tested were able to reconstruct images of good quality and generate healthy-looking images from simulated images presenting anomalies.</p><p><strong>Conclusions: </strong>Even if no model clearly outperformed all the others, the benchmark allowed identifying a few models that perform slightly better than the vanilla VAE. It further showed that many VAE-based models can generalize to the detection of anomalies of various intensities, shapes, and locations in 3D brain FDG PET.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"054005"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12543032/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145356605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DMM-UNet: dual-path multi-scale Mamba UNet for medical image segmentation. DMM-UNet:用于医学图像分割的双路径多尺度曼巴UNet。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-09-29 DOI: 10.1117/1.JMI.12.5.054003
Liquan Zhao, Mingxia Cao, Yanfei Jia

Purpose: State space models have shown promise in medical image segmentation by modeling long-range dependencies with linear complexity. However, they are limited in their ability to capture local features, which hinders their capacity to extract multiscale details and integrate global and local contextual information effectively. To address these shortcomings, we propose the dual-path multi-scale Mamba UNet (DMM-UNet) model.

Approach: This architecture facilitates deep fusion of local and global features through multi-scale modules within a U-shaped encoder-decoder framework. First, we introduce the multi-scale channel attention selective scanning block in the encoder, which combines global selective scanning with multi-scale channel attention to model both long-range and local dependencies simultaneously. Second, we design the spatial attention selective scanning block for the decoder. This block integrates global scanning with spatial attention mechanisms, enabling precise aggregation of semantic features through gated weighting. Finally, we develop the multi-dimensional collaborative attention layer to extract complementary attention weights across height, width, and channel dimensions, facilitating cross-space-channel feature interactions.

Results: Experiments were conducted on the ISIC17, ISIC18, Synapse, and ACDC datasets. One of the indicators, Dice similarity coefficient, achieved 89.88% on the ISIC17 dataset, 90.52% on the ISIC18 dataset, 83.07% on the Synapse dataset, and 92.60% on the ACDC dataset. There are also other indicators that perform well on this model.

Conclusions: The DMM-UNet model effectively addresses the shortcomings of state space models by enabling the integration of both local and global features, improving segmentation performance, and offering enhanced multiscale feature fusion for medical image segmentation tasks.

目的:状态空间模型通过对具有线性复杂性的远程依赖关系进行建模,在医学图像分割中显示出良好的前景。然而,它们捕获局部特征的能力有限,这阻碍了它们提取多尺度细节和有效整合全局和局部上下文信息的能力。为了解决这些缺点,我们提出了双路径多尺度曼巴UNet (DMM-UNet)模型。方法:该架构通过u型编码器-解码器框架内的多尺度模块促进局部和全局特征的深度融合。首先,我们在编码器中引入了多尺度通道注意选择性扫描块,将全局选择性扫描与多尺度通道注意相结合,同时对远程和局部依赖关系进行建模。其次,设计了译码器的空间注意选择扫描块。该块集成了全局扫描和空间注意机制,通过门控加权实现语义特征的精确聚合。最后,我们开发了多维协同关注层,以提取跨高度、宽度和通道维度的互补关注权重,促进跨空间通道特征交互。结果:在ISIC17、ISIC18、Synapse和ACDC数据集上进行了实验。其中Dice相似系数在ISIC17数据集上达到89.88%,在ISIC18数据集上达到90.52%,在Synapse数据集上达到83.07%,在ACDC数据集上达到92.60%。还有其他一些指标在这个模型上表现良好。结论:DMM-UNet模型有效地解决了状态空间模型的不足,实现了局部和全局特征的融合,提高了分割性能,并为医学图像分割任务提供了增强的多尺度特征融合。
{"title":"DMM-UNet: dual-path multi-scale Mamba UNet for medical image segmentation.","authors":"Liquan Zhao, Mingxia Cao, Yanfei Jia","doi":"10.1117/1.JMI.12.5.054003","DOIUrl":"https://doi.org/10.1117/1.JMI.12.5.054003","url":null,"abstract":"<p><strong>Purpose: </strong>State space models have shown promise in medical image segmentation by modeling long-range dependencies with linear complexity. However, they are limited in their ability to capture local features, which hinders their capacity to extract multiscale details and integrate global and local contextual information effectively. To address these shortcomings, we propose the dual-path multi-scale Mamba UNet (DMM-UNet) model.</p><p><strong>Approach: </strong>This architecture facilitates deep fusion of local and global features through multi-scale modules within a U-shaped encoder-decoder framework. First, we introduce the multi-scale channel attention selective scanning block in the encoder, which combines global selective scanning with multi-scale channel attention to model both long-range and local dependencies simultaneously. Second, we design the spatial attention selective scanning block for the decoder. This block integrates global scanning with spatial attention mechanisms, enabling precise aggregation of semantic features through gated weighting. Finally, we develop the multi-dimensional collaborative attention layer to extract complementary attention weights across height, width, and channel dimensions, facilitating cross-space-channel feature interactions.</p><p><strong>Results: </strong>Experiments were conducted on the ISIC17, ISIC18, Synapse, and ACDC datasets. One of the indicators, Dice similarity coefficient, achieved 89.88% on the ISIC17 dataset, 90.52% on the ISIC18 dataset, 83.07% on the Synapse dataset, and 92.60% on the ACDC dataset. There are also other indicators that perform well on this model.</p><p><strong>Conclusions: </strong>The DMM-UNet model effectively addresses the shortcomings of state space models by enabling the integration of both local and global features, improving segmentation performance, and offering enhanced multiscale feature fusion for medical image segmentation tasks.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"054003"},"PeriodicalIF":1.7,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12480969/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145208018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using a limited field of view to improve training for pulmonary nodule detection on radiographs. 利用有限视场改进x线片上肺结节检测的培训。
IF 1.9 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2025-09-01 Epub Date: 2025-04-25 DOI: 10.1117/1.JMI.12.5.051804
Samual K Zenger, Rishabh Agarwal, William F Auffermann

Purpose: Perceptual error is a significant cause of medical errors in radiology. Given the amount of information in a medical image, an image interpreter may become distracted by information unrelated to their search pattern. This may be especially challenging for novices. We aim to examine teaching medical trainees to evaluate chest radiographs (CXRs) for pulmonary nodules on limited field-of-view (LFOV) images, with the field of view (FOV) restricted to the lungs and mediastinum.

Approach: Healthcare trainees with limited exposure to interpreting images were asked to identify pulmonary nodules on CXRs, half of which contained nodules. The control and experimental groups evaluated two sets of CXRs. After the first set, the experimental group was trained to evaluate LFOV images, and both groups were again asked to assess CXRs for pulmonary nodules. Participants were given surveys after this educational session to determine their thoughts about the training and symptoms of computer vision syndrome (CVS).

Results: There was a significant improvement in performance in pulmonary nodule identification for both the experimental and control groups, but the improvement was more considerable in the experimental group ( p - value = 0.022 ). Survey responses were uniformly positive, and each question was statistically significant (all p - values < 0.001 ).

Conclusions: Our results show that using LFOV images may be helpful when teaching trainees specific high-yield perceptual tasks, such as nodule identification. The use of LFOV images was associated with reduced symptoms of CVS.

目的:感知错误是导致放射学医疗错误的重要原因。给定医学图像中的信息量,图像解释器可能会被与其搜索模式无关的信息分散注意力。这对新手来说尤其具有挑战性。我们的目的是检查医学培训生在有限视场(LFOV)图像上评估胸片(cxr)对肺结节的诊断,视场(FOV)仅限于肺和纵隔。方法:医疗保健培训生与有限的接触解释图像被要求识别肺结节的cxr,其中一半包含结节。对照组和实验组分别评价两组cxr。在第一组之后,实验组接受训练以评估LFOV图像,两组再次被要求评估肺结节的cxr。在这一教育课程结束后,参与者接受了调查,以确定他们对训练和计算机视觉综合征(CVS)症状的看法。结果:实验组与对照组肺结节识别能力均有显著提高,但实验组提高更明显(p值= 0.022)。调查结果一致是肯定的,每个问题都有统计学意义(p值均为0.001)。结论:我们的研究结果表明,在教授受训者特定的高收益感知任务(如结节识别)时,使用LFOV图像可能有所帮助。LFOV图像的使用与CVS症状的减轻有关。
{"title":"Using a limited field of view to improve training for pulmonary nodule detection on radiographs.","authors":"Samual K Zenger, Rishabh Agarwal, William F Auffermann","doi":"10.1117/1.JMI.12.5.051804","DOIUrl":"10.1117/1.JMI.12.5.051804","url":null,"abstract":"<p><strong>Purpose: </strong>Perceptual error is a significant cause of medical errors in radiology. Given the amount of information in a medical image, an image interpreter may become distracted by information unrelated to their search pattern. This may be especially challenging for novices. We aim to examine teaching medical trainees to evaluate chest radiographs (CXRs) for pulmonary nodules on limited field-of-view (LFOV) images, with the field of view (FOV) restricted to the lungs and mediastinum.</p><p><strong>Approach: </strong>Healthcare trainees with limited exposure to interpreting images were asked to identify pulmonary nodules on CXRs, half of which contained nodules. The control and experimental groups evaluated two sets of CXRs. After the first set, the experimental group was trained to evaluate LFOV images, and both groups were again asked to assess CXRs for pulmonary nodules. Participants were given surveys after this educational session to determine their thoughts about the training and symptoms of computer vision syndrome (CVS).</p><p><strong>Results: </strong>There was a significant improvement in performance in pulmonary nodule identification for both the experimental and control groups, but the improvement was more considerable in the experimental group ( <math><mrow><mi>p</mi> <mtext>-</mtext> <mtext>value</mtext> <mo>=</mo> <mn>0.022</mn></mrow> </math> ). Survey responses were uniformly positive, and each question was statistically significant (all <math><mrow><mi>p</mi> <mtext>-</mtext> <mtext>values</mtext> <mo><</mo> <mn>0.001</mn></mrow> </math> ).</p><p><strong>Conclusions: </strong>Our results show that using LFOV images may be helpful when teaching trainees specific high-yield perceptual tasks, such as nodule identification. The use of LFOV images was associated with reduced symptoms of CVS.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"12 5","pages":"051804"},"PeriodicalIF":1.9,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12023444/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144021338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Medical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1