首页 > 最新文献

Journal of Medical Imaging最新文献

英文 中文
Asynchronous federated learning for web-based OCT image analysis. 基于web的OCT图像分析异步联合学习。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-05 DOI: 10.1117/1.JMI.13.1.014501
Hasan Md Tusfiqur Alam, Tim Maurer, Abdulrahman Mohamed Selim, Matthias Eiletz, Michael Barz, Daniel Sonntag

Purpose: Centralized machine learning often struggles with limited data access and expert involvement. We investigate decentralized approaches that preserve data privacy while enabling collaborative model training for medical imaging tasks.

Approach: We explore asynchronous federated learning (FL) using the FL with buffered asynchronous aggregation (FedBuff) algorithm for classifying optical coherence tomography (OCT) retina images. Unlike synchronous algorithms such as FedAvg, which require all clients to participate simultaneously, FedBuff supports independent client updates. We compare its performance to both centralized models and FedAvg. In addition, we develop a browser-based proof-of-concept system using modern web technologies to assess the feasibility and limitations of interactive, collaborative learning in real-world settings.

Results: FedBuff performs well in binary OCT classification tasks but shows reduced accuracy in more complex, multiclass scenarios. FedAvg achieves results comparable to centralized training, consistent with previous findings. Although FedBuff underperforms compared with FedAvg and centralized models, it still delivers acceptable accuracy in less complex settings. The browser-based prototype demonstrates the potential for accessible, user-driven FL systems but also highlights technical limitations in current web standards, especially regarding local computation and communication efficiency.

Conclusion: Asynchronous FL via FedBuff offers a promising, privacy-preserving approach for medical image classification, particularly when synchronous participation is impractical. However, its scalability to complex classification tasks remains limited. Web-based implementations have the potential to broaden access to collaborative AI tools, but limitations of the current technologies need to be further investigated.

目的:集中式机器学习经常与有限的数据访问和专家参与作斗争。我们研究了分散的方法,这些方法既可以保护数据隐私,又可以为医学成像任务提供协作模型训练。方法:我们使用带缓冲异步聚合(FedBuff)算法的异步联邦学习(FL)来对光学相干断层扫描(OCT)视网膜图像进行分类。与fedag等要求所有客户端同时参与的同步算法不同,FedBuff支持独立客户端更新。我们将其性能与集中式模型和fedag进行了比较。此外,我们开发了一个基于浏览器的概念验证系统,使用现代网络技术来评估现实世界中交互式协作学习的可行性和局限性。结果:FedBuff在二元OCT分类任务中表现良好,但在更复杂的多类别场景中准确性降低。fedag达到了与集中训练相当的结果,与先前的研究结果一致。尽管与fedag和集中式模型相比,FedBuff表现不佳,但在不太复杂的设置中,它仍然提供了可接受的准确性。基于浏览器的原型展示了可访问的、用户驱动的FL系统的潜力,但也突出了当前web标准的技术限制,特别是在本地计算和通信效率方面。结论:通过FedBuff的异步FL为医学图像分类提供了一种有前途的、保护隐私的方法,特别是在同步参与不切实际的情况下。然而,它对复杂分类任务的可扩展性仍然有限。基于网络的实现有可能扩大对协作人工智能工具的访问,但当前技术的局限性需要进一步研究。
{"title":"Asynchronous federated learning for web-based OCT image analysis.","authors":"Hasan Md Tusfiqur Alam, Tim Maurer, Abdulrahman Mohamed Selim, Matthias Eiletz, Michael Barz, Daniel Sonntag","doi":"10.1117/1.JMI.13.1.014501","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014501","url":null,"abstract":"<p><strong>Purpose: </strong>Centralized machine learning often struggles with limited data access and expert involvement. We investigate decentralized approaches that preserve data privacy while enabling collaborative model training for medical imaging tasks.</p><p><strong>Approach: </strong>We explore asynchronous federated learning (FL) using the FL with buffered asynchronous aggregation (FedBuff) algorithm for classifying optical coherence tomography (OCT) retina images. Unlike synchronous algorithms such as FedAvg, which require all clients to participate simultaneously, FedBuff supports independent client updates. We compare its performance to both centralized models and FedAvg. In addition, we develop a browser-based proof-of-concept system using modern web technologies to assess the feasibility and limitations of interactive, collaborative learning in real-world settings.</p><p><strong>Results: </strong>FedBuff performs well in binary OCT classification tasks but shows reduced accuracy in more complex, multiclass scenarios. FedAvg achieves results comparable to centralized training, consistent with previous findings. Although FedBuff underperforms compared with FedAvg and centralized models, it still delivers acceptable accuracy in less complex settings. The browser-based prototype demonstrates the potential for accessible, user-driven FL systems but also highlights technical limitations in current web standards, especially regarding local computation and communication efficiency.</p><p><strong>Conclusion: </strong>Asynchronous FL via FedBuff offers a promising, privacy-preserving approach for medical image classification, particularly when synchronous participation is impractical. However, its scalability to complex classification tasks remains limited. Web-based implementations have the potential to broaden access to collaborative AI tools, but limitations of the current technologies need to be further investigated.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014501"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12767620/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145913135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LiteMIL: a computationally efficient cross-attention multiple instance learning for cancer subtyping on whole-slide images. LiteMIL:一种计算效率高的跨注意多实例学习,用于整张幻灯片图像的癌症亚型分型。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-06 DOI: 10.1117/1.JMI.13.1.017501
Haitham Kussaibi

Purpose: Accurate cancer subtyping is essential for precision medicine but challenged by the computational demands of gigapixel whole-slide images (WSIs). Although transformer-based multiple instance learning (MIL) methods achieve strong performance, their quadratic complexity limits clinical deployment. We introduce LiteMIL, a computationally efficient cross-attention MIL, optimized for WSIs classification.

Approach: LiteMIL employs a single learnable query with multi-head cross-attention for bag-level aggregation from extracted features. We evaluated LiteMIL against five baselines (mean/max pooling, ABMIL, MAD-MIL, and TransMIL) on four TCGA datasets (breast: n = 875 , kidney: n = 906 , lung: n = 958 , and TUPAC16: n = 821 ) using nested cross-validation with patient-level splitting. Systematic ablation studies evaluated multi-query variants, attention heads, dropout rates, and architectural components.

Results: LiteMIL achieved competitive accuracy (average 83.5%), matching TransMIL, while offering substantial efficiency gains: 4.8 × fewer parameters (560K versus 2.67M), 2.9 × faster inference (1.6s versus 4.6s per fold), and 6.7 × lower Graphics Processing Unit (GPU) memory usage (1.15 GB versus 7.77 GB). LiteMIL excelled on lung (86.3% versus 85.0%), TUPAC16 (72% versus 71.4%), and matched kidney performance (89.9% versus 89.7%). Ablation studies revealed task-dependent multi-query performance benefits: Q = 4 versus Q = 1 improved morphologically heterogeneous tasks (breast/lung + 1.3 % each, p < 0.05 ) but degraded on grading tasks (TUPAC16: - 1.6 % ), validating single-query optimality for focused attention scenarios.

Conclusions: LiteMIL provides a resource-efficient solution for WSI classification. The cross-attention architecture matches complex transformer performance while enabling deployment on consumer GPUs. Task-dependent design insights, single query for sparse discriminating features, multi-query for heterogeneous patterns, guide practical implementation. The architecture's efficiency, combined with compact features, makes LiteMIL suitable for clinical integration in settings with limited computational infrastructure.

目的:准确的癌症亚型对精准医学至关重要,但受到千兆像素整片图像(wsi)计算需求的挑战。尽管基于变压器的多实例学习(MIL)方法具有较强的性能,但其二次复杂度限制了临床应用。我们介绍了LiteMIL,一个计算效率高的交叉注意MIL,优化了wsi分类。方法:LiteMIL采用单个可学习的查询,具有多头交叉关注,用于从提取的特征中进行袋级聚合。我们在4个TCGA数据集(乳腺:n = 875,肾脏:n = 906,肺:n = 958, TUPAC16: n = 821)上对LiteMIL的5个基线(mean/max pooling, ABMIL, med - mil和TransMIL)进行了评估,采用嵌套交叉验证和患者水平分割。系统消融研究评估了多查询变量、注意头、辍学率和架构组件。结果:LiteMIL达到了具有竞争力的准确性(平均83.5%),与TransMIL相匹配,同时提供了可观的效率提升:参数减少4.8倍(560K对2.67M),推理速度加快2.9倍(1.6s对4.6s每倍),图形处理单元(GPU)内存使用降低6.7倍(1.15 GB对7.77 GB)。LiteMIL在肺(86.3%对85.0%)、TUPAC16(72%对71.4%)和匹配肾脏性能(89.9%对89.7%)方面表现出色。消融研究显示了任务相关的多查询性能优势:Q = 4和Q = 1改善了形态学异构任务(乳腺/肺各+ 1.3%,p 0.05),但在分级任务上有所下降(TUPAC16: - 1.6%),验证了单查询在集中注意力场景中的最佳性。结论:LiteMIL为WSI分类提供了一种资源高效的解决方案。交叉关注架构匹配复杂的变压器性能,同时支持在消费级gpu上部署。任务相关的设计见解,单个查询用于稀疏区分特征,多个查询用于异构模式,指导实际实现。该架构的效率,加上紧凑的功能,使LiteMIL适合在计算基础设施有限的情况下进行临床集成。
{"title":"LiteMIL: a computationally efficient cross-attention multiple instance learning for cancer subtyping on whole-slide images.","authors":"Haitham Kussaibi","doi":"10.1117/1.JMI.13.1.017501","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.017501","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate cancer subtyping is essential for precision medicine but challenged by the computational demands of gigapixel whole-slide images (WSIs). Although transformer-based multiple instance learning (MIL) methods achieve strong performance, their quadratic complexity limits clinical deployment. We introduce LiteMIL, a computationally efficient cross-attention MIL, optimized for WSIs classification.</p><p><strong>Approach: </strong>LiteMIL employs a single learnable query with multi-head cross-attention for bag-level aggregation from extracted features. We evaluated LiteMIL against five baselines (mean/max pooling, ABMIL, MAD-MIL, and TransMIL) on four TCGA datasets (breast: <math><mrow><mi>n</mi> <mo>=</mo> <mn>875</mn></mrow> </math> , kidney: <math><mrow><mi>n</mi> <mo>=</mo> <mn>906</mn></mrow> </math> , lung: <math><mrow><mi>n</mi> <mo>=</mo> <mn>958</mn></mrow> </math> , and TUPAC16: <math><mrow><mi>n</mi> <mo>=</mo> <mn>821</mn></mrow> </math> ) using nested cross-validation with patient-level splitting. Systematic ablation studies evaluated multi-query variants, attention heads, dropout rates, and architectural components.</p><p><strong>Results: </strong>LiteMIL achieved competitive accuracy (average 83.5%), matching TransMIL, while offering substantial efficiency gains: <math><mrow><mn>4.8</mn> <mo>×</mo></mrow> </math> fewer parameters (560K versus 2.67M), <math><mrow><mn>2.9</mn> <mo>×</mo></mrow> </math> faster inference (1.6s versus 4.6s per fold), and <math><mrow><mn>6.7</mn> <mo>×</mo></mrow> </math> lower Graphics Processing Unit (GPU) memory usage (1.15 GB versus 7.77 GB). LiteMIL excelled on lung (86.3% versus 85.0%), TUPAC16 (72% versus 71.4%), and matched kidney performance (89.9% versus 89.7%). Ablation studies revealed task-dependent multi-query performance benefits: <math><mrow><mi>Q</mi> <mo>=</mo> <mn>4</mn></mrow> </math> versus <math><mrow><mi>Q</mi> <mo>=</mo> <mn>1</mn></mrow> </math> improved morphologically heterogeneous tasks (breast/lung <math><mrow><mo>+</mo> <mn>1.3</mn> <mo>%</mo></mrow> </math> each, <math><mrow><mi>p</mi> <mo><</mo> <mn>0.05</mn></mrow> </math> ) but degraded on grading tasks (TUPAC16: <math><mrow><mo>-</mo> <mn>1.6</mn> <mo>%</mo></mrow> </math> ), validating single-query optimality for focused attention scenarios.</p><p><strong>Conclusions: </strong>LiteMIL provides a resource-efficient solution for WSI classification. The cross-attention architecture matches complex transformer performance while enabling deployment on consumer GPUs. Task-dependent design insights, single query for sparse discriminating features, multi-query for heterogeneous patterns, guide practical implementation. The architecture's efficiency, combined with compact features, makes LiteMIL suitable for clinical integration in settings with limited computational infrastructure.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"017501"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12772520/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145918838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Characterizing the effects of noncontrast head CT reconstruction kernel and slice thickness parameters on the performance of an automated AI algorithm in the evaluation of ischemic stroke. 表征非对比头部CT重建核和层厚参数对一种评估缺血性卒中的自动AI算法性能的影响。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-09 DOI: 10.1117/1.JMI.13.1.014503
Spencer H Welland, Grace Hyun J Kim, Anil Yadav, Kambiz Nael, John M Hoffman, Matthew S Brown, Michael F McNitt-Gray, William Hsu

Purpose: There are multiple commercially available, Food and Drug Administration (FDA)-cleared, artificial intelligence (AI)-based tools automating stroke evaluation in noncontrast computed tomography (NCCT). This study assessed the impact of variations in reconstruction kernel and slice thickness on two outputs of such a system: hypodense volume and Alberta Stroke Program Early CT Score (ASPECTS).

Approach: The NCCT series image data of 67 patients imaged with a CT stroke protocol were reconstructed with four kernels (H10s-smooth, H40s-medium, H60s-sharp, and H70h-very sharp) and three slice thicknesses (1.5, 3.0, and 5.0 mm) to create 1 reference condition (H40s/5.0 mm) and 11 nonreference conditions. The 12 reconstructions per patient were processed with a commercially available FDA-cleared software package that yields total hypodense volume (mL) and ASPECTS. A mixed-effect model was used to test the difference in hypodense volume, and an ordered logistic model was used to test the difference in e-ASPECTS.

Results: Hypodense volume differences from the reference condition ranged from - 14.6 to 1.1 mL and were significant for all nonreference kernels (H10s p = 0.025 , H60s p < 0.001 , and H70h p < 0.001 ) and for thinner slices (1.5 mm p < 0.001 and 3.0 mm p = 0.002 ). e-ASPECTS was invariant to the nonreference kernels and slice thicknesses, with a mean difference ranging from - 0.1 to 0.5. No significant differences were found for any kernel or slice thickness (all p > 0.05 ).

Conclusions: Automated hypodense volume measured with a commercially available, FDA-cleared software package is substantially impacted by reconstruction kernel and slice thickness. Conversely, automated ASPECTS is invariant to these reconstruction parameters.

目的:有多种市售的、美国食品和药物管理局(FDA)批准的、基于人工智能(AI)的工具,可以在非对比计算机断层扫描(NCCT)中自动评估脑卒中。本研究评估了重建核和切片厚度的变化对该系统的两个输出结果的影响:低密度体积和Alberta卒中程序早期CT评分(ASPECTS)。方法:对67例脑卒中CT成像患者的NCCT系列图像数据进行4个核(h10 -平滑、h40 -中等、h60 -尖锐、h70 -非常尖锐)和3个切片厚度(1.5、3.0、5.0 mm)重构,创建1个参考条件(h40 /5.0 mm)和11个非参考条件。每位患者的12个重建用市售的fda批准的软件包进行处理,该软件包产生总低密度体积(mL)和ASPECTS。采用混合效应模型检验低密度体积的差异,采用有序logistic模型检验e-ASPECTS的差异。结果:与参考条件相比,低密度体积差异在- 14.6至1.1 mL之间,并且在所有非参考条件下(H10s p = 0.025, h60h p = 0.001, H70h p = 0.001)和较薄切片(1.5 mm p = 0.001和3.0 mm p = 0.002)均具有显著性。e-ASPECTS对非参考核和切片厚度不变,平均差值在- 0.1 ~ 0.5之间。仁层厚度和切片厚度均无显著差异(p < 0.05)。结论:用市售的、fda批准的软件包自动测量低密度体积受到重建核和切片厚度的很大影响。相反,自动化方面对这些重建参数是不变的。
{"title":"Characterizing the effects of noncontrast head CT reconstruction kernel and slice thickness parameters on the performance of an automated AI algorithm in the evaluation of ischemic stroke.","authors":"Spencer H Welland, Grace Hyun J Kim, Anil Yadav, Kambiz Nael, John M Hoffman, Matthew S Brown, Michael F McNitt-Gray, William Hsu","doi":"10.1117/1.JMI.13.1.014503","DOIUrl":"10.1117/1.JMI.13.1.014503","url":null,"abstract":"<p><strong>Purpose: </strong>There are multiple commercially available, Food and Drug Administration (FDA)-cleared, artificial intelligence (AI)-based tools automating stroke evaluation in noncontrast computed tomography (NCCT). This study assessed the impact of variations in reconstruction kernel and slice thickness on two outputs of such a system: hypodense volume and Alberta Stroke Program Early CT Score (ASPECTS).</p><p><strong>Approach: </strong>The NCCT series image data of 67 patients imaged with a CT stroke protocol were reconstructed with four kernels (H10s-smooth, H40s-medium, H60s-sharp, and H70h-very sharp) and three slice thicknesses (1.5, 3.0, and 5.0 mm) to create 1 reference condition (H40s/5.0 mm) and 11 nonreference conditions. The 12 reconstructions per patient were processed with a commercially available FDA-cleared software package that yields total hypodense volume (mL) and ASPECTS. A mixed-effect model was used to test the difference in hypodense volume, and an ordered logistic model was used to test the difference in e-ASPECTS.</p><p><strong>Results: </strong>Hypodense volume differences from the reference condition ranged from <math><mrow><mo>-</mo> <mn>14.6</mn></mrow> </math> to 1.1 mL and were significant for all nonreference kernels (H10s <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.025</mn></mrow> </math> , H60s <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> , and H70h <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> ) and for thinner slices (1.5 mm <math><mrow><mi>p</mi> <mo><</mo> <mn>0.001</mn></mrow> </math> and 3.0 mm <math><mrow><mi>p</mi> <mo>=</mo> <mn>0.002</mn></mrow> </math> ). e-ASPECTS was invariant to the nonreference kernels and slice thicknesses, with a mean difference ranging from <math><mrow><mo>-</mo> <mn>0.1</mn></mrow> </math> to 0.5. No significant differences were found for any kernel or slice thickness (all <math><mrow><mi>p</mi> <mo>></mo> <mn>0.05</mn></mrow> </math> ).</p><p><strong>Conclusions: </strong>Automated hypodense volume measured with a commercially available, FDA-cleared software package is substantially impacted by reconstruction kernel and slice thickness. Conversely, automated ASPECTS is invariant to these reconstruction parameters.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014503"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12782953/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145953582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Attention-driven framework to segment renal ablation zone in posttreatment CT images: a step toward ablation margin evaluation. 治疗后CT图像中肾消融区分割的注意力驱动框架:消融边缘评估的一个步骤。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-05 DOI: 10.1117/1.JMI.13.1.014001
Maryam Rastegarpoor, Derek W Cool, Aaron Fenster

Purpose: Thermal ablation is a minimally invasive therapy used for the treatment of small renal cell carcinoma tumors. Treatment success is evaluated on postablation computed tomography (CT) to determine if the ablation zone covered the tumor with an adequate treatment margin (often 5 to 10 mm). Incorrect margin identification can lead to treatment misassessment, resulting in unnecessary additional ablation. Therefore, segmentation of the renal ablation zone (RAZ) is crucial for treatment evaluation. We aim to develop and assess an accurate deep learning workflow for delineating the RAZ from surrounding tissues in kidney CT images.

Approach: We present an advanced deep learning method using the attention-based U-Net architecture to segment the RAZ. The workflow leverages the strengths of U-Net, enhanced with attention mechanisms, to improve the network's focus on the most relevant parts of the images, resulting in an accurate segmentation.

Results: Our model was trained and evaluated on a dataset comprising 76 patients' annotated RAZs in CT images. Analysis demonstrated that the proposed workflow achieved an accuracy = 0.97 ± 0.02 , precision = 0.74 ± 0.23 , recall = 0.73 ± 0.25 , DSC = 0.70 ± 0.22 , Jaccard = 0.58 ± 0.22 , specificity = 0.99 ± 0.01 , Hausdorff distance = 6.70 ± 4.44    mm , and mean absolute boundary distance = 2.67 ± 2.22    mm .

Conclusions: We used 3D CT images with RAZs and, for the first time, addressed deep-learning-based RAZ segmentation using parallel CT images. Our framework can effectively segment RAZs, allowing clinicians to automatically determine the ablation margin, making our tool ready for clinical use. Prediction time is 1    s per patient, enabling clinicians to perform quick reviews, especially in time-constrained settings.

目的:热消融是一种微创治疗肾小细胞癌的方法。通过消融后的计算机断层扫描(CT)评估治疗成功与否,以确定消融区是否覆盖肿瘤并有足够的治疗范围(通常为5 - 10mm)。不正确的切缘识别可导致治疗评估错误,导致不必要的额外消融。因此,肾消融区(RAZ)的分割对治疗评价至关重要。我们的目标是开发和评估一种准确的深度学习工作流程,用于在肾脏CT图像中从周围组织中描绘RAZ。方法:我们提出了一种先进的深度学习方法,使用基于注意力的U-Net架构来分割RAZ。该工作流利用U-Net的优势,增强了注意力机制,以提高网络对图像最相关部分的关注,从而实现准确的分割。结果:我们的模型在包含76例患者CT图像注释raz的数据集上进行了训练和评估。结果表明,该工作流的准确率为0.97±0.02,精密度为0.74±0.23,召回率为0.73±0.25,DSC为0.70±0.22,Jaccard为0.58±0.22,特异性为0.99±0.01,Hausdorff距离为6.70±4.44 mm,平均绝对边界距离为2.67±2.22 mm。结论:我们将三维CT图像与RAZ结合使用,并首次使用并行CT图像解决了基于深度学习的RAZ分割问题。我们的框架可以有效地分割raz,允许临床医生自动确定消融范围,使我们的工具为临床使用做好准备。预测时间为每位患者约1秒,使临床医生能够进行快速审查,特别是在时间有限的情况下。
{"title":"Attention-driven framework to segment renal ablation zone in posttreatment CT images: a step toward ablation margin evaluation.","authors":"Maryam Rastegarpoor, Derek W Cool, Aaron Fenster","doi":"10.1117/1.JMI.13.1.014001","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014001","url":null,"abstract":"<p><strong>Purpose: </strong>Thermal ablation is a minimally invasive therapy used for the treatment of small renal cell carcinoma tumors. Treatment success is evaluated on postablation computed tomography (CT) to determine if the ablation zone covered the tumor with an adequate treatment margin (often 5 to 10 mm). Incorrect margin identification can lead to treatment misassessment, resulting in unnecessary additional ablation. Therefore, segmentation of the renal ablation zone (RAZ) is crucial for treatment evaluation. We aim to develop and assess an accurate deep learning workflow for delineating the RAZ from surrounding tissues in kidney CT images.</p><p><strong>Approach: </strong>We present an advanced deep learning method using the attention-based U-Net architecture to segment the RAZ. The workflow leverages the strengths of U-Net, enhanced with attention mechanisms, to improve the network's focus on the most relevant parts of the images, resulting in an accurate segmentation.</p><p><strong>Results: </strong>Our model was trained and evaluated on a dataset comprising 76 patients' annotated RAZs in CT images. Analysis demonstrated that the proposed workflow achieved an accuracy <math><mrow><mo>=</mo> <mn>0.97</mn> <mo>±</mo> <mn>0.02</mn></mrow> </math> , precision <math><mrow><mo>=</mo> <mn>0.74</mn> <mo>±</mo> <mn>0.23</mn></mrow> </math> , <math><mrow><mtext>recall</mtext> <mo>=</mo> <mn>0.73</mn> <mo>±</mo> <mn>0.25</mn></mrow> </math> , <math><mrow><mi>DSC</mi> <mo>=</mo> <mn>0.70</mn> <mo>±</mo> <mn>0.22</mn></mrow> </math> , Jaccard <math><mrow><mo>=</mo> <mn>0.58</mn> <mo>±</mo> <mn>0.22</mn></mrow> </math> , specificity <math><mrow><mo>=</mo> <mn>0.99</mn> <mo>±</mo> <mn>0.01</mn></mrow> </math> , Hausdorff distance <math><mrow><mo>=</mo> <mn>6.70</mn> <mo>±</mo> <mn>4.44</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> , and mean absolute boundary distance <math><mrow><mo>=</mo> <mn>2.67</mn> <mo>±</mo> <mn>2.22</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> .</p><p><strong>Conclusions: </strong>We used 3D CT images with RAZs and, for the first time, addressed deep-learning-based RAZ segmentation using parallel CT images. Our framework can effectively segment RAZs, allowing clinicians to automatically determine the ablation margin, making our tool ready for clinical use. Prediction time is <math><mrow><mo>∼</mo> <mn>1</mn> <mtext>  </mtext> <mi>s</mi></mrow> </math> per patient, enabling clinicians to perform quick reviews, especially in time-constrained settings.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014001"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12767621/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145913125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward robust modeling of breast biomechanical compression: an extended study using graph neural networks. 乳房生物力学压缩的鲁棒建模:使用图神经网络的扩展研究。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2025-12-29 DOI: 10.1117/1.JMI.13.1.015001
Hadeel Awwad, Eloy García, Robert Martí

Purpose: Accurate simulation of breast tissue deformation is essential for reliable image registration between 3D imaging modalities and 2D mammograms, where compression significantly alters tissue geometry. Although finite element analysis (FEA) provides high-fidelity modeling, it is computationally intensive and not well suited for rapid simulations. To address this, the physics-based graph neural network (PhysGNN) has been introduced as a computationally efficient approximation model trained on FEA-generated deformations. We extend prior work by evaluating the performance of PhysGNN on new digital breast phantoms and assessing the impact of training on multiple phantoms.

Approach: PhysGNN was trained on both single-phantom (per-geometry) and multiphantom (multigeometry) datasets generated from incremental FEA simulations. The digital breast phantoms represent the uncompressed state, serving as input geometries for predicting compressed configurations. A leave-one-deformation-out evaluation strategy was used to assess predictive performance under compression.

Results: Training on new digital phantoms confirmed the model's robust performance, though with some variability in prediction accuracy reflecting the diverse anatomical structures. Multiphantom training further enhanced this robustness and reduced prediction errors.

Conclusions: PhysGNN offers a computationally efficient alternative to FEA for simulating breast compression. The results showed that model performance remains robust when trained per-geometry, and further demonstrated that multigeometry training enhances predictive accuracy and robustness for the geometries included in the training set. This suggests a strong potential path toward developing reliable models for generating compressed breast volumes, which could facilitate image registration and algorithm development.

目的:准确模拟乳腺组织变形对于3D成像模式和2D乳房x线照片之间的可靠图像配准至关重要,其中压缩显着改变了组织几何形状。虽然有限元分析(FEA)提供了高保真的建模,但它的计算量很大,不适合快速模拟。为了解决这个问题,基于物理的图神经网络(PhysGNN)作为一种计算效率高的近似模型被引入到有限元生成的变形上。我们通过评估PhysGNN在新的数字乳房模型上的性能和评估训练对多个模型的影响来扩展先前的工作。方法:PhysGNN在增量有限元模拟生成的单模型(每几何)和多模型(多几何)数据集上进行训练。数字乳房幻影表示未压缩状态,作为预测压缩配置的输入几何形状。采用“留一变形”评价策略对压缩下的预测性能进行评价。结果:对新的数字模型的训练证实了该模型的鲁棒性,尽管在反映不同解剖结构的预测精度上存在一些差异。多幻影训练进一步增强了这种鲁棒性并减少了预测误差。结论:PhysGNN为模拟乳房压缩提供了一种计算效率高的替代方法。结果表明,在按几何形状训练时,模型的鲁棒性保持不变,并进一步证明了多几何形状训练提高了训练集中几何形状的预测精度和鲁棒性。这为开发可靠的模型来生成压缩乳房体积提供了一条强有力的潜在途径,这可以促进图像配准和算法开发。
{"title":"Toward robust modeling of breast biomechanical compression: an extended study using graph neural networks.","authors":"Hadeel Awwad, Eloy García, Robert Martí","doi":"10.1117/1.JMI.13.1.015001","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.015001","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate simulation of breast tissue deformation is essential for reliable image registration between 3D imaging modalities and 2D mammograms, where compression significantly alters tissue geometry. Although finite element analysis (FEA) provides high-fidelity modeling, it is computationally intensive and not well suited for rapid simulations. To address this, the physics-based graph neural network (PhysGNN) has been introduced as a computationally efficient approximation model trained on FEA-generated deformations. We extend prior work by evaluating the performance of PhysGNN on new digital breast phantoms and assessing the impact of training on multiple phantoms.</p><p><strong>Approach: </strong>PhysGNN was trained on both single-phantom (per-geometry) and multiphantom (multigeometry) datasets generated from incremental FEA simulations. The digital breast phantoms represent the uncompressed state, serving as input geometries for predicting compressed configurations. A leave-one-deformation-out evaluation strategy was used to assess predictive performance under compression.</p><p><strong>Results: </strong>Training on new digital phantoms confirmed the model's robust performance, though with some variability in prediction accuracy reflecting the diverse anatomical structures. Multiphantom training further enhanced this robustness and reduced prediction errors.</p><p><strong>Conclusions: </strong>PhysGNN offers a computationally efficient alternative to FEA for simulating breast compression. The results showed that model performance remains robust when trained per-geometry, and further demonstrated that multigeometry training enhances predictive accuracy and robustness for the geometries included in the training set. This suggests a strong potential path toward developing reliable models for generating compressed breast volumes, which could facilitate image registration and algorithm development.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"015001"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12745487/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145866013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LCSD-Net: a light-weight cross-attention-based semantic dual transformer for domain generalization in melanoma detection. LCSD-Net:用于黑色素瘤检测领域泛化的轻量级交叉注意语义双转换器。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-06 DOI: 10.1117/1.JMI.13.1.014502
Rishi Agrawal, Neeraj Gupta, Anand Singh Jalal

Purpose: Research in deep learning has shown a great advancement in the detection of melanoma. However, recent literature has emphasized a tendency of certain models to rely on disease-irrelevant visual artifacts such as dark corners, dense hair, or ruler marks. The dependence on these markers leads to biased models that do well for training but generalize poorly to heterogeneous clinical environments. To address these limitations in developing reliability in skin lesion detection, a lightweight cross-attention-based semantic dual (LCSD) transformer model was proposed.

Approach: The LCSD model extracts global-level semantic information, uses feature normalization to improve model accuracy, and employs semantic queries to improve domain generalization. Multihead attention is included with the semantic queries to refine global features. The cross-attention between feature maps and semantic query provides the model with a generalized encoding of the global context. The model improved the computational complexity from O ( n 2 d ) to O ( n m d + m 2 d ) , which makes the model suitable for the development of real-time and mobile applications.

Results: Empirical evaluation was conducted on three challenging datasets: Derm7pt-Dermoscopic, Derm7pt-Clinical, and PAD-UFES-20. The proposed model achieved classification accuracies of 82.82%, 72.95%, and 86.21%, respectively. These results demonstrate superior performance compared with conventional transformer-based models, highlighting both improved robustness and reduced computational cost.

Conclusion: The LCSD model mitigates the influence of irrelevant visual characteristics, enhances domain generalization, and ensures better adaptability across diverse clinical scenarios. Its lightweight design further supports deployment in mobile applications, making it a reliable and efficient solution for real-world melanoma detection.

目的:深度学习的研究在黑色素瘤的检测方面取得了很大的进展。然而,最近的文献强调,某些模型倾向于依赖与疾病无关的视觉人工制品,如黑暗的角落,浓密的头发,或标尺标记。对这些标记的依赖导致有偏见的模型在训练中表现良好,但在异质临床环境中泛化能力差。为了解决这些限制在开发可靠性的皮肤损伤检测,提出了一个轻量级的基于交叉注意的语义对偶(LCSD)变压器模型。方法:LCSD模型提取全局语义信息,使用特征归一化来提高模型精度,使用语义查询来提高领域泛化。语义查询中包含多头注意,以细化全局特征。特征映射和语义查询之间的交叉关注为模型提供了全局上下文的通用编码。该模型将计算复杂度从0 (n²d)提高到0 (n²d + m²d),适合实时和移动应用的开发。结果:对三个具有挑战性的数据集:Derm7pt-Dermoscopic、Derm7pt-Clinical和pad - upes -20进行了实证评估。该模型的分类准确率分别为82.82%、72.95%和86.21%。与传统的基于变压器的模型相比,这些结果显示了优越的性能,突出了增强的鲁棒性和降低的计算成本。结论:LCSD模型减轻了不相关视觉特征的影响,增强了领域泛化,确保了对不同临床场景更好的适应性。其轻量级设计进一步支持移动应用程序的部署,使其成为现实世界黑色素瘤检测的可靠和高效的解决方案。
{"title":"LCSD-Net: a light-weight cross-attention-based semantic dual transformer for domain generalization in melanoma detection.","authors":"Rishi Agrawal, Neeraj Gupta, Anand Singh Jalal","doi":"10.1117/1.JMI.13.1.014502","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014502","url":null,"abstract":"<p><strong>Purpose: </strong>Research in deep learning has shown a great advancement in the detection of melanoma. However, recent literature has emphasized a tendency of certain models to rely on disease-irrelevant visual artifacts such as dark corners, dense hair, or ruler marks. The dependence on these markers leads to biased models that do well for training but generalize poorly to heterogeneous clinical environments. To address these limitations in developing reliability in skin lesion detection, a lightweight cross-attention-based semantic dual (LCSD) transformer model was proposed.</p><p><strong>Approach: </strong>The LCSD model extracts global-level semantic information, uses feature normalization to improve model accuracy, and employs semantic queries to improve domain generalization. Multihead attention is included with the semantic queries to refine global features. The cross-attention between feature maps and semantic query provides the model with a generalized encoding of the global context. The model improved the computational complexity from <math><mrow><mi>O</mi> <mo>(</mo> <msup><mrow><mi>n</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mi>d</mi> <mo>)</mo></mrow> </math> to <math><mrow><mi>O</mi> <mo>(</mo> <mi>n</mi> <mi>m</mi> <mi>d</mi> <mo>+</mo> <msup><mrow><mi>m</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mi>d</mi> <mo>)</mo></mrow> </math> , which makes the model suitable for the development of real-time and mobile applications.</p><p><strong>Results: </strong>Empirical evaluation was conducted on three challenging datasets: Derm7pt-Dermoscopic, Derm7pt-Clinical, and PAD-UFES-20. The proposed model achieved classification accuracies of 82.82%, 72.95%, and 86.21%, respectively. These results demonstrate superior performance compared with conventional transformer-based models, highlighting both improved robustness and reduced computational cost.</p><p><strong>Conclusion: </strong>The LCSD model mitigates the influence of irrelevant visual characteristics, enhances domain generalization, and ensures better adaptability across diverse clinical scenarios. Its lightweight design further supports deployment in mobile applications, making it a reliable and efficient solution for real-world melanoma detection.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014502"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12773922/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145918799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiomic signatures from baseline CT predict chemotherapy response in unresectable colorectal liver metastases. 基线CT放射学特征预测不可切除的结直肠癌肝转移的化疗反应。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-13 DOI: 10.1117/1.JMI.13.1.014505
Mane Piliposyan, Jacob J Peoples, Mohammad Hamghalam, Ramtin Mojtahedi, Kaitlyn Kobayashi, E Claire Bunker, Natalie Gangai, Hyunseon C Kang, Yun Shin Chun, Christian Muise, Richard K G Do, Amber L Simpson

Purpose: Colorectal cancer is the third most common cancer globally, with a high mortality rate due to metastatic progression, particularly in the liver. Surgical resection remains the main curative treatment, but only a small subset of patients is eligible for surgery at diagnosis. For patients with initially unresectable colorectal liver metastases (CRLM), neoadjuvant chemotherapy can downstage tumors, potentially making surgery feasible. We investigate whether radiomic signatures-quantitative imaging biomarkers derived from baseline computed tomography (CT) scans-can noninvasively predict chemotherapy response in patients with unresectable CRLM, offering a pathway toward personalized treatment planning.

Approach: We used radiomics combined with a stacking classifier (SC) to predict treatment outcome. Baseline CT imaging data from 355 patients with initially unresectable CRLM were analyzed using two regions of interest (ROIs) separately (all tumors in the liver and the largest tumor by volume). From each ROI, 107 radiomic features were extracted. The dataset was split into training and testing sets, and multiple machine learning models were trained and integrated via stacking to enhance prediction. Logistic regression coefficients were used to derive radiomic signatures.

Results: The SC achieved strong predictive performance, with an area under the receiver operating characteristic curve of up to 0.77 for response prediction. Logistic regression identified 12 and 7 predictive features for treatment response in all tumors and the largest tumor ROIs, respectively.

Conclusion: Our findings demonstrate that radiomic features from baseline CT scans can serve as robust, interpretable biomarkers for predicting chemotherapy response, offering insights to guide personalized treatment in unresectable CRLM.

目的:结直肠癌是全球第三大常见癌症,由于转移进展,特别是在肝脏,死亡率很高。手术切除仍然是主要的治疗方法,但只有一小部分患者在诊断时符合手术条件。对于最初无法切除的结肠肝转移(CRLM)患者,新辅助化疗可以降低肿瘤的分期,可能使手术成为可能。我们研究放射学特征-来自基线计算机断层扫描(CT)扫描的定量成像生物标志物-是否可以无创地预测不可切除的CRLM患者的化疗反应,为个性化治疗计划提供途径。方法:我们使用放射组学结合堆叠分类器(SC)来预测治疗结果。355例最初不可切除的CRLM患者的基线CT成像数据分别使用两个感兴趣区域(所有肝脏肿瘤和体积最大的肿瘤)进行分析。从每个ROI中提取107个放射学特征。将数据集分为训练集和测试集,对多个机器学习模型进行训练和叠加,增强预测能力。逻辑回归系数被用来推导放射性特征。结果:SC具有较强的预测效果,受试者工作特征曲线下面积可达0.77。Logistic回归分别确定了所有肿瘤治疗反应和最大肿瘤roi的12个和7个预测特征。结论:我们的研究结果表明,基线CT扫描的放射学特征可以作为预测化疗反应的可靠、可解释的生物标志物,为指导不可切除的CRLM的个性化治疗提供见解。
{"title":"Radiomic signatures from baseline CT predict chemotherapy response in unresectable colorectal liver metastases.","authors":"Mane Piliposyan, Jacob J Peoples, Mohammad Hamghalam, Ramtin Mojtahedi, Kaitlyn Kobayashi, E Claire Bunker, Natalie Gangai, Hyunseon C Kang, Yun Shin Chun, Christian Muise, Richard K G Do, Amber L Simpson","doi":"10.1117/1.JMI.13.1.014505","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014505","url":null,"abstract":"<p><strong>Purpose: </strong>Colorectal cancer is the third most common cancer globally, with a high mortality rate due to metastatic progression, particularly in the liver. Surgical resection remains the main curative treatment, but only a small subset of patients is eligible for surgery at diagnosis. For patients with initially unresectable colorectal liver metastases (CRLM), neoadjuvant chemotherapy can downstage tumors, potentially making surgery feasible. We investigate whether radiomic signatures-quantitative imaging biomarkers derived from baseline computed tomography (CT) scans-can noninvasively predict chemotherapy response in patients with unresectable CRLM, offering a pathway toward personalized treatment planning.</p><p><strong>Approach: </strong>We used radiomics combined with a stacking classifier (SC) to predict treatment outcome. Baseline CT imaging data from 355 patients with initially unresectable CRLM were analyzed using two regions of interest (ROIs) separately (all tumors in the liver and the largest tumor by volume). From each ROI, 107 radiomic features were extracted. The dataset was split into training and testing sets, and multiple machine learning models were trained and integrated via stacking to enhance prediction. Logistic regression coefficients were used to derive radiomic signatures.</p><p><strong>Results: </strong>The SC achieved strong predictive performance, with an area under the receiver operating characteristic curve of up to 0.77 for response prediction. Logistic regression identified 12 and 7 predictive features for treatment response in all tumors and the largest tumor ROIs, respectively.</p><p><strong>Conclusion: </strong>Our findings demonstrate that radiomic features from baseline CT scans can serve as robust, interpretable biomarkers for predicting chemotherapy response, offering insights to guide personalized treatment in unresectable CRLM.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014505"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12797257/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145971472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient computed tomography-based image segmentation for predicting lateral cervical lymph node metastasis in papillary thyroid carcinoma. 基于ct的高效图像分割预测甲状腺乳头状癌侧颈淋巴结转移。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-13 DOI: 10.1117/1.JMI.13.1.014504
Lei Xu, Bin Zhang, Xingyuan Li, Guona Zheng, Yong Wang, YanHui Peng

Purpose: Papillary thyroid carcinoma (PTC) is a common thyroid cancer, and accurate preoperative assessment of lateral cervical lymph node metastasis is critical for surgical planning. Current methods are often subjective and prone to misdiagnosis. This study aims to improve the accuracy of metastasis evaluation using a deep learning-based segmentation method on enhanced computed tomography (CT) images.

Approach: We propose a YOLOv8-based deep learning model integrated with a deformable self-attention module to enhance metastatic lymph node segmentation. The model was trained on a large dataset of pathology-confirmed CT images from PTC patients.

Results: The model demonstrated diagnostic performance comparable to experienced physicians, with high precision in identifying metastatic nodes. The deformable self-attention module improved segmentation accuracy, with strong sensitivity and specificity.

Conclusion: This deep learning approach improves the accuracy of preoperative assessment for lateral cervical lymph node metastasis in PTC patients, aiding surgical planning, reducing misdiagnosis, and lowering medical costs. It shows promise for enhancing patient outcomes in PTC management.

目的:甲状腺乳头状癌(PTC)是一种常见的甲状腺癌,术前准确评估颈部外侧淋巴结转移对手术计划至关重要。目前的方法往往是主观的,容易误诊。本研究旨在利用基于深度学习的增强计算机断层扫描(CT)图像分割方法提高转移评估的准确性。方法:我们提出了一种基于yolov8的深度学习模型,并集成了一个可变形的自关注模块,以增强转移性淋巴结的分割。该模型是在PTC患者病理证实的CT图像的大型数据集上训练的。结果:该模型表现出与经验丰富的医生相当的诊断性能,在识别转移淋巴结方面具有很高的精度。可变形自关注模块提高了分割精度,具有较强的敏感性和特异性。结论:该深度学习方法提高了PTC患者颈外侧淋巴结转移术前评估的准确性,有助于手术规划,减少误诊,降低医疗费用。它显示了在PTC管理中提高患者预后的希望。
{"title":"Efficient computed tomography-based image segmentation for predicting lateral cervical lymph node metastasis in papillary thyroid carcinoma.","authors":"Lei Xu, Bin Zhang, Xingyuan Li, Guona Zheng, Yong Wang, YanHui Peng","doi":"10.1117/1.JMI.13.1.014504","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014504","url":null,"abstract":"<p><strong>Purpose: </strong>Papillary thyroid carcinoma (PTC) is a common thyroid cancer, and accurate preoperative assessment of lateral cervical lymph node metastasis is critical for surgical planning. Current methods are often subjective and prone to misdiagnosis. This study aims to improve the accuracy of metastasis evaluation using a deep learning-based segmentation method on enhanced computed tomography (CT) images.</p><p><strong>Approach: </strong>We propose a YOLOv8-based deep learning model integrated with a deformable self-attention module to enhance metastatic lymph node segmentation. The model was trained on a large dataset of pathology-confirmed CT images from PTC patients.</p><p><strong>Results: </strong>The model demonstrated diagnostic performance comparable to experienced physicians, with high precision in identifying metastatic nodes. The deformable self-attention module improved segmentation accuracy, with strong sensitivity and specificity.</p><p><strong>Conclusion: </strong>This deep learning approach improves the accuracy of preoperative assessment for lateral cervical lymph node metastasis in PTC patients, aiding surgical planning, reducing misdiagnosis, and lowering medical costs. It shows promise for enhancing patient outcomes in PTC management.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014504"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12797499/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145971502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BAF-UNet: a boundary-aware segmentation model for skin lesion segmentation. BAF-UNet:一种边界感知的皮肤损伤分割模型。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-08 DOI: 10.1117/1.JMI.13.1.014003
Menglei Zhang, Congwei Zhang, Zhibin Quan, Bing Guo, Wankou Yang

Purpose: Skin lesion segmentation plays a significant role in the diagnosis and treatment of skin cancer. Accurate skin lesion segmentation is essential for skin cancer diagnosis and treatment but is challenged by ambiguous boundaries and diverse lesion shapes and sizes. We aim to improve segmentation performance with enhanced boundary preservation.

Approach: We propose BAF-UNet, a boundary-aware segmentation network. It integrates a multiscale boundary-aware feature fusion (BFF) module to combine low-level boundary features with high-level semantic information, and a boundary-aware vision transformer (BAViT) that incorporates boundary guidance into MobileViT to capture local and global context. A boundary-focused loss function is also introduced to prioritize edge accuracy during training. The model is evaluated on ISIC2016, ISIC2017, and PH2 datasets.

Results: Experiments demonstrate that BAF-UNet improves Dice scores and boundary accuracy compared to baseline models. The BFF and BAViT modules enhance boundary delineation while maintaining robustness across lesions of varying shapes and sizes.

Conclusions: BAF-UNet effectively integrates boundary guidance into feature fusion and transformer-based context modeling, significantly improving segmentation accuracy, particularly along lesion edges, and shows potential for clinical application in automated skin cancer diagnosis.

目的:皮肤病变分割在皮肤癌的诊断和治疗中起着重要的作用。准确的皮肤病变分割对于皮肤癌的诊断和治疗至关重要,但由于皮肤病变边界模糊、形状和大小不一而受到挑战。我们的目标是通过增强边界保持来提高分割性能。方法:我们提出了一种边界感知分割网络BAF-UNet。它集成了一个多尺度边界感知特征融合(BFF)模块,将低级边界特征与高级语义信息相结合,以及一个边界感知视觉转换器(BAViT),将边界引导集成到MobileViT中,以捕获局部和全局上下文。在训练过程中引入边界聚焦损失函数对边缘精度进行优先排序。在ISIC2016、ISIC2017和PH2数据集上对模型进行了评估。结果:实验表明,与基线模型相比,BAF-UNet提高了Dice分数和边界精度。BFF和BAViT模块增强了边界划分,同时保持了不同形状和大小病变的稳健性。结论:BAF-UNet有效地将边界引导集成到特征融合和基于变压器的上下文建模中,显著提高了分割精度,特别是沿病灶边缘的分割精度,在皮肤癌自动诊断中具有临床应用潜力。
{"title":"BAF-UNet: a boundary-aware segmentation model for skin lesion segmentation.","authors":"Menglei Zhang, Congwei Zhang, Zhibin Quan, Bing Guo, Wankou Yang","doi":"10.1117/1.JMI.13.1.014003","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014003","url":null,"abstract":"<p><strong>Purpose: </strong>Skin lesion segmentation plays a significant role in the diagnosis and treatment of skin cancer. Accurate skin lesion segmentation is essential for skin cancer diagnosis and treatment but is challenged by ambiguous boundaries and diverse lesion shapes and sizes. We aim to improve segmentation performance with enhanced boundary preservation.</p><p><strong>Approach: </strong>We propose BAF-UNet, a boundary-aware segmentation network. It integrates a multiscale boundary-aware feature fusion (BFF) module to combine low-level boundary features with high-level semantic information, and a boundary-aware vision transformer (BAViT) that incorporates boundary guidance into MobileViT to capture local and global context. A boundary-focused loss function is also introduced to prioritize edge accuracy during training. The model is evaluated on ISIC2016, ISIC2017, and PH2 datasets.</p><p><strong>Results: </strong>Experiments demonstrate that BAF-UNet improves Dice scores and boundary accuracy compared to baseline models. The BFF and BAViT modules enhance boundary delineation while maintaining robustness across lesions of varying shapes and sizes.</p><p><strong>Conclusions: </strong>BAF-UNet effectively integrates boundary guidance into feature fusion and transformer-based context modeling, significantly improving segmentation accuracy, particularly along lesion edges, and shows potential for clinical application in automated skin cancer diagnosis.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014003"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12782429/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145953505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultrasound imaging using single-element biaxial beamforming. 超声成像使用单元件双轴波束成形。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2025-12-23 DOI: 10.1117/1.JMI.13.1.017001
Nathan Meulenbroek, Laura Curiel, Adam Waspe, Samuel Pichardo

Purpose: Dynamic focusing of received ultrasound signals, or beamforming, is foundational for ultrasound imaging. Conventionally, it requires arrays of ultrasound sensors to estimate where sound came from using time-of-flight (TOF) measurements. We demonstrate passive beamforming with a single biaxial sensor and accurate passive acoustic mapping with two biaxial sensors using only direction of arrival (DOA) information.

Approach: We introduce two single-element biaxial beamforming algorithms and four biaxial image reconstruction algorithms for a two-element biaxial piezoceramic transducer array. Imaging of a hemispherical acoustic source is characterized in an acoustic scanning tank within the region - 30.29    mm x 29.94 mm and 50.11 mm z 90.45 mm relative to the center of the array. Imaging performance is contrasted with delay, sum, and integrate (DSAI) and delay, multiply, sum, and integrate (DMSAI) algorithms.

Results: Single-element biaxial beamforming can identify DOA with a median error (± interquartile range) of 0.36 ± 0.63    deg and median full-width half-prominence of 7.3 ± 8.6    deg . Using both array elements, DOA-only images demonstrate overall median localization error of 6.41 mm (lateral: 1.02 mm, axial: 5.85 mm, signal-to-noise ratio (SNR): 15.37) and DOA + TOF images demonstrate overall median error of 6.91 mm (lateral: 1.69 mm, axial: 6.11 mm, SNR: 18.37).

Conclusions: To the best of our knowledge, we provide the first demonstration of single-element beamforming using a single stationary piezoceramic and the first demonstration of passive ultrasound imaging without the use of TOF information. These results enable simpler, smaller, more cost-effective arrays for passive ultrasound imaging.

目的:接收超声信号的动态聚焦或波束形成是超声成像的基础。传统上,它需要超声波传感器阵列来估计声音来自何处,使用飞行时间(TOF)测量。我们演示了使用单个双轴传感器的被动波束形成和仅使用到达方向(DOA)信息的两个双轴传感器的精确被动声学映射。方法:针对两元双轴压电换能器阵列,介绍了两种单元双轴波束形成算法和四种双轴图像重建算法。在相对于阵列中心的- 30.29 mm≤x≤29.94 mm和50.11 mm≤z≤90.45 mm区域内,对半球形声源进行成像。对比了延迟、求和和积分(DSAI)算法和延迟、乘法、求和和积分(DMSAI)算法的成像性能。结果:单单元双轴波束形成识别DOA的中位误差(±四分位间距)为0.36±0.63°,中位全宽半凸度为7.3±8.6°。使用这两种阵列元素,仅DOA图像的总体中位数定位误差为6.41 mm(侧向:1.02 mm,轴向:5.85 mm,信噪比(SNR): 15.37), DOA + TOF图像的总体中位数定位误差为6.91 mm(侧向:1.69 mm,轴向:6.11 mm,信噪比:18.37)。结论:据我们所知,我们提供了第一个使用单个静止压电陶瓷的单元件波束形成演示,以及第一个不使用TOF信息的被动超声成像演示。这些结果使被动超声成像阵列更简单、更小、更具成本效益。
{"title":"Ultrasound imaging using single-element biaxial beamforming.","authors":"Nathan Meulenbroek, Laura Curiel, Adam Waspe, Samuel Pichardo","doi":"10.1117/1.JMI.13.1.017001","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.017001","url":null,"abstract":"<p><strong>Purpose: </strong>Dynamic focusing of received ultrasound signals, or beamforming, is foundational for ultrasound imaging. Conventionally, it requires arrays of ultrasound sensors to estimate where sound came from using time-of-flight (TOF) measurements. We demonstrate passive beamforming with a single biaxial sensor and accurate passive acoustic mapping with two biaxial sensors using only direction of arrival (DOA) information.</p><p><strong>Approach: </strong>We introduce two single-element biaxial beamforming algorithms and four biaxial image reconstruction algorithms for a two-element biaxial piezoceramic transducer array. Imaging of a hemispherical acoustic source is characterized in an acoustic scanning tank within the region <math><mrow><mo>-</mo> <mn>30.29</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> <math><mrow><mo>≤</mo> <mi>x</mi> <mo>≤</mo></mrow> </math> 29.94 mm and 50.11 mm <math><mrow><mo>≤</mo> <mi>z</mi> <mo>≤</mo></mrow> </math> 90.45 mm relative to the center of the array. Imaging performance is contrasted with delay, sum, and integrate (DSAI) and delay, multiply, sum, and integrate (DMSAI) algorithms.</p><p><strong>Results: </strong>Single-element biaxial beamforming can identify DOA with a median error (± interquartile range) of <math><mrow><mn>0.36</mn> <mo>±</mo> <mn>0.63</mn> <mtext>  </mtext> <mi>deg</mi></mrow> </math> and median full-width half-prominence of <math><mrow><mn>7.3</mn> <mo>±</mo> <mn>8.6</mn> <mtext>  </mtext> <mi>deg</mi></mrow> </math> . Using both array elements, DOA-only images demonstrate overall median localization error of 6.41 mm (lateral: 1.02 mm, axial: 5.85 mm, signal-to-noise ratio (SNR): 15.37) and DOA + TOF images demonstrate overall median error of 6.91 mm (lateral: 1.69 mm, axial: 6.11 mm, SNR: 18.37).</p><p><strong>Conclusions: </strong>To the best of our knowledge, we provide the first demonstration of single-element beamforming using a single stationary piezoceramic and the first demonstration of passive ultrasound imaging without the use of TOF information. These results enable simpler, smaller, more cost-effective arrays for passive ultrasound imaging.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"017001"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12726554/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145828543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Medical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1