首页 > 最新文献

Journal of Medical Imaging最新文献

英文 中文
Airway quantifications of bronchitis patients with photon-counting and energy-integrating computed tomography. 光子计数和能量积分计算机断层扫描对支气管炎患者气道的定量研究。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-02-02 DOI: 10.1117/1.JMI.13.1.013501
Fong Chi Ho, William Paul Segars, Ehsan Samei, Ehsan Abadi

Purpose: Accurate airway measurement is critical for bronchitis quantification with computed tomography (CT), yet optimal protocols and the added value of photon-counting CT (PCCT) over energy-integrating CT (EICT) for reducing bias remain unclear. We quantified biomarker accuracy across modalities and protocols and assessed strategies to reduce bias.

Approach: A virtual imaging trial with 20 bronchitis anthropomorphic models was scanned using a validated simulator for two systems (EICT: SOMATOM Flash; PCCT: NAEOTOM Alpha) at 6.3 and 12.6 mGy. Reconstructions varied algorithm, kernel sharpness, slice thickness, and pixel size. Pi10 (square-root wall thickness at 10-mm perimeter) and WA% (wall-area percentage) were compared against ground-truth airway dimensions obtained from the 0.1-mm-precision anatomical models prior to CT simulation. External validation used clinical PCCT ( n = 22 ) and EICT ( n = 80 ).

Results: Simulated airway dimensions agreed with pathological references ( R = 0.89 - 0.93 ). PCCT had lower errors than EICT across segmented generations ( p < 0.05 ). Under optimal parameters, PCCT improved Pi10 and WA% accuracy by 26.3% and 64.9%. Across the tested PCCT and EICT imaging protocols, improvements were associated with sharper kernels (25.8% Pi10, 33.0% WA%), thinner slices (23.9% Pi10, 49.8% WA%), smaller pixels (17.0% Pi10, 23.1% WA%), and higher dose ( 3.9 % ). Clinically, PCCT achieved higher maximum airway generation ( 8.8 ± 0.5 versus 6.0 ± 1.1 ) and lower variability, mirroring trends in virtual results.

Conclusions: PCCT improves the accuracy and consistency of airway biomarker quantification relative to EICT, particularly with optimized protocols. The validated virtual platform enables modality-bias assessment and protocol optimization for accurate, reproducible bronchitis measurements.

目的:准确的气道测量对于用计算机断层扫描(CT)量化支气管炎至关重要,但最佳方案和光子计数CT (PCCT)比能量积分CT (EICT)在减少偏倚方面的附加价值尚不清楚。我们量化了各种模式和方案的生物标志物准确性,并评估了减少偏倚的策略。方法:对20个支气管炎拟人模型进行虚拟成像试验,使用两种系统(EICT: SOMATOM Flash; PCCT: NAEOTOM Alpha)在6.3和12.6 mGy下进行扫描。重建不同的算法,核清晰度,切片厚度和像素大小。Pi10 (10mm周长的平方根壁厚)和WA%(壁面积百分比)与CT模拟前从0.1 mm精度的解剖模型中获得的真实气道尺寸进行比较。外部验证采用临床PCCT (n = 22)和EICT (n = 80)。结果:模拟气道尺寸与病理对照吻合(R = 0.89 ~ 0.93)。PCCT在不同世代间的误差低于EICT (p < 0.05)。在最优参数下,PCCT的Pi10和WA%准确率分别提高了26.3%和64.9%。在测试的PCCT和EICT成像方案中,改善与更清晰的核(25.8% Pi10, 33.0% WA%),更薄的切片(23.9% Pi10, 49.8% WA%),更小的像素(17.0% Pi10, 23.1% WA%)和更高的剂量(≤3.9%)相关。在临床上,PCCT实现了更高的最大气道生成(8.8±0.5比6.0±1.1)和更低的变异性,反映了虚拟结果的趋势。结论:相对于EICT, PCCT提高了气道生物标志物定量的准确性和一致性,特别是优化的方案。经过验证的虚拟平台能够进行模式偏差评估和方案优化,以实现准确,可重复的支气管炎测量。
{"title":"Airway quantifications of bronchitis patients with photon-counting and energy-integrating computed tomography.","authors":"Fong Chi Ho, William Paul Segars, Ehsan Samei, Ehsan Abadi","doi":"10.1117/1.JMI.13.1.013501","DOIUrl":"10.1117/1.JMI.13.1.013501","url":null,"abstract":"<p><strong>Purpose: </strong>Accurate airway measurement is critical for bronchitis quantification with computed tomography (CT), yet optimal protocols and the added value of photon-counting CT (PCCT) over energy-integrating CT (EICT) for reducing bias remain unclear. We quantified biomarker accuracy across modalities and protocols and assessed strategies to reduce bias.</p><p><strong>Approach: </strong>A virtual imaging trial with 20 bronchitis anthropomorphic models was scanned using a validated simulator for two systems (EICT: SOMATOM Flash; PCCT: NAEOTOM Alpha) at 6.3 and 12.6 mGy. Reconstructions varied algorithm, kernel sharpness, slice thickness, and pixel size. Pi10 (square-root wall thickness at 10-mm perimeter) and WA% (wall-area percentage) were compared against ground-truth airway dimensions obtained from the 0.1-mm-precision anatomical models prior to CT simulation. External validation used clinical PCCT ( <math><mrow><mi>n</mi> <mo>=</mo> <mn>22</mn></mrow> </math> ) and EICT ( <math><mrow><mi>n</mi> <mo>=</mo> <mn>80</mn></mrow> </math> ).</p><p><strong>Results: </strong>Simulated airway dimensions agreed with pathological references ( <math><mrow><mi>R</mi> <mo>=</mo> <mn>0.89</mn> <mo>-</mo> <mn>0.93</mn></mrow> </math> ). PCCT had lower errors than EICT across segmented generations ( <math><mrow><mi>p</mi> <mo><</mo> <mn>0.05</mn></mrow> </math> ). Under optimal parameters, PCCT improved Pi10 and WA% accuracy by 26.3% and 64.9%. Across the tested PCCT and EICT imaging protocols, improvements were associated with sharper kernels (25.8% Pi10, 33.0% WA%), thinner slices (23.9% Pi10, 49.8% WA%), smaller pixels (17.0% Pi10, 23.1% WA%), and higher dose ( <math><mrow><mo>≤</mo> <mn>3.9</mn> <mo>%</mo></mrow> </math> ). Clinically, PCCT achieved higher maximum airway generation ( <math><mrow><mn>8.8</mn> <mo>±</mo> <mn>0.5</mn></mrow> </math> versus <math><mrow><mn>6.0</mn> <mo>±</mo> <mn>1.1</mn></mrow> </math> ) and lower variability, mirroring trends in virtual results.</p><p><strong>Conclusions: </strong>PCCT improves the accuracy and consistency of airway biomarker quantification relative to EICT, particularly with optimized protocols. The validated virtual platform enables modality-bias assessment and protocol optimization for accurate, reproducible bronchitis measurements.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"013501"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12863983/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146114479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LCSD-Net: a light-weight cross-attention-based semantic dual transformer for domain generalization in melanoma detection. LCSD-Net:用于黑色素瘤检测领域泛化的轻量级交叉注意语义双转换器。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-06 DOI: 10.1117/1.JMI.13.1.014502
Rishi Agrawal, Neeraj Gupta, Anand Singh Jalal

Purpose: Research in deep learning has shown a great advancement in the detection of melanoma. However, recent literature has emphasized a tendency of certain models to rely on disease-irrelevant visual artifacts such as dark corners, dense hair, or ruler marks. The dependence on these markers leads to biased models that do well for training but generalize poorly to heterogeneous clinical environments. To address these limitations in developing reliability in skin lesion detection, a lightweight cross-attention-based semantic dual (LCSD) transformer model was proposed.

Approach: The LCSD model extracts global-level semantic information, uses feature normalization to improve model accuracy, and employs semantic queries to improve domain generalization. Multihead attention is included with the semantic queries to refine global features. The cross-attention between feature maps and semantic query provides the model with a generalized encoding of the global context. The model improved the computational complexity from O ( n 2 d ) to O ( n m d + m 2 d ) , which makes the model suitable for the development of real-time and mobile applications.

Results: Empirical evaluation was conducted on three challenging datasets: Derm7pt-Dermoscopic, Derm7pt-Clinical, and PAD-UFES-20. The proposed model achieved classification accuracies of 82.82%, 72.95%, and 86.21%, respectively. These results demonstrate superior performance compared with conventional transformer-based models, highlighting both improved robustness and reduced computational cost.

Conclusion: The LCSD model mitigates the influence of irrelevant visual characteristics, enhances domain generalization, and ensures better adaptability across diverse clinical scenarios. Its lightweight design further supports deployment in mobile applications, making it a reliable and efficient solution for real-world melanoma detection.

目的:深度学习的研究在黑色素瘤的检测方面取得了很大的进展。然而,最近的文献强调,某些模型倾向于依赖与疾病无关的视觉人工制品,如黑暗的角落,浓密的头发,或标尺标记。对这些标记的依赖导致有偏见的模型在训练中表现良好,但在异质临床环境中泛化能力差。为了解决这些限制在开发可靠性的皮肤损伤检测,提出了一个轻量级的基于交叉注意的语义对偶(LCSD)变压器模型。方法:LCSD模型提取全局语义信息,使用特征归一化来提高模型精度,使用语义查询来提高领域泛化。语义查询中包含多头注意,以细化全局特征。特征映射和语义查询之间的交叉关注为模型提供了全局上下文的通用编码。该模型将计算复杂度从0 (n²d)提高到0 (n²d + m²d),适合实时和移动应用的开发。结果:对三个具有挑战性的数据集:Derm7pt-Dermoscopic、Derm7pt-Clinical和pad - upes -20进行了实证评估。该模型的分类准确率分别为82.82%、72.95%和86.21%。与传统的基于变压器的模型相比,这些结果显示了优越的性能,突出了增强的鲁棒性和降低的计算成本。结论:LCSD模型减轻了不相关视觉特征的影响,增强了领域泛化,确保了对不同临床场景更好的适应性。其轻量级设计进一步支持移动应用程序的部署,使其成为现实世界黑色素瘤检测的可靠和高效的解决方案。
{"title":"LCSD-Net: a light-weight cross-attention-based semantic dual transformer for domain generalization in melanoma detection.","authors":"Rishi Agrawal, Neeraj Gupta, Anand Singh Jalal","doi":"10.1117/1.JMI.13.1.014502","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014502","url":null,"abstract":"<p><strong>Purpose: </strong>Research in deep learning has shown a great advancement in the detection of melanoma. However, recent literature has emphasized a tendency of certain models to rely on disease-irrelevant visual artifacts such as dark corners, dense hair, or ruler marks. The dependence on these markers leads to biased models that do well for training but generalize poorly to heterogeneous clinical environments. To address these limitations in developing reliability in skin lesion detection, a lightweight cross-attention-based semantic dual (LCSD) transformer model was proposed.</p><p><strong>Approach: </strong>The LCSD model extracts global-level semantic information, uses feature normalization to improve model accuracy, and employs semantic queries to improve domain generalization. Multihead attention is included with the semantic queries to refine global features. The cross-attention between feature maps and semantic query provides the model with a generalized encoding of the global context. The model improved the computational complexity from <math><mrow><mi>O</mi> <mo>(</mo> <msup><mrow><mi>n</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mi>d</mi> <mo>)</mo></mrow> </math> to <math><mrow><mi>O</mi> <mo>(</mo> <mi>n</mi> <mi>m</mi> <mi>d</mi> <mo>+</mo> <msup><mrow><mi>m</mi></mrow> <mrow><mn>2</mn></mrow> </msup> <mi>d</mi> <mo>)</mo></mrow> </math> , which makes the model suitable for the development of real-time and mobile applications.</p><p><strong>Results: </strong>Empirical evaluation was conducted on three challenging datasets: Derm7pt-Dermoscopic, Derm7pt-Clinical, and PAD-UFES-20. The proposed model achieved classification accuracies of 82.82%, 72.95%, and 86.21%, respectively. These results demonstrate superior performance compared with conventional transformer-based models, highlighting both improved robustness and reduced computational cost.</p><p><strong>Conclusion: </strong>The LCSD model mitigates the influence of irrelevant visual characteristics, enhances domain generalization, and ensures better adaptability across diverse clinical scenarios. Its lightweight design further supports deployment in mobile applications, making it a reliable and efficient solution for real-world melanoma detection.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014502"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12773922/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145918799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiomic signatures from baseline CT predict chemotherapy response in unresectable colorectal liver metastases. 基线CT放射学特征预测不可切除的结直肠癌肝转移的化疗反应。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-13 DOI: 10.1117/1.JMI.13.1.014505
Mane Piliposyan, Jacob J Peoples, Mohammad Hamghalam, Ramtin Mojtahedi, Kaitlyn Kobayashi, E Claire Bunker, Natalie Gangai, Hyunseon C Kang, Yun Shin Chun, Christian Muise, Richard K G Do, Amber L Simpson

Purpose: Colorectal cancer is the third most common cancer globally, with a high mortality rate due to metastatic progression, particularly in the liver. Surgical resection remains the main curative treatment, but only a small subset of patients is eligible for surgery at diagnosis. For patients with initially unresectable colorectal liver metastases (CRLM), neoadjuvant chemotherapy can downstage tumors, potentially making surgery feasible. We investigate whether radiomic signatures-quantitative imaging biomarkers derived from baseline computed tomography (CT) scans-can noninvasively predict chemotherapy response in patients with unresectable CRLM, offering a pathway toward personalized treatment planning.

Approach: We used radiomics combined with a stacking classifier (SC) to predict treatment outcome. Baseline CT imaging data from 355 patients with initially unresectable CRLM were analyzed using two regions of interest (ROIs) separately (all tumors in the liver and the largest tumor by volume). From each ROI, 107 radiomic features were extracted. The dataset was split into training and testing sets, and multiple machine learning models were trained and integrated via stacking to enhance prediction. Logistic regression coefficients were used to derive radiomic signatures.

Results: The SC achieved strong predictive performance, with an area under the receiver operating characteristic curve of up to 0.77 for response prediction. Logistic regression identified 12 and 7 predictive features for treatment response in all tumors and the largest tumor ROIs, respectively.

Conclusion: Our findings demonstrate that radiomic features from baseline CT scans can serve as robust, interpretable biomarkers for predicting chemotherapy response, offering insights to guide personalized treatment in unresectable CRLM.

目的:结直肠癌是全球第三大常见癌症,由于转移进展,特别是在肝脏,死亡率很高。手术切除仍然是主要的治疗方法,但只有一小部分患者在诊断时符合手术条件。对于最初无法切除的结肠肝转移(CRLM)患者,新辅助化疗可以降低肿瘤的分期,可能使手术成为可能。我们研究放射学特征-来自基线计算机断层扫描(CT)扫描的定量成像生物标志物-是否可以无创地预测不可切除的CRLM患者的化疗反应,为个性化治疗计划提供途径。方法:我们使用放射组学结合堆叠分类器(SC)来预测治疗结果。355例最初不可切除的CRLM患者的基线CT成像数据分别使用两个感兴趣区域(所有肝脏肿瘤和体积最大的肿瘤)进行分析。从每个ROI中提取107个放射学特征。将数据集分为训练集和测试集,对多个机器学习模型进行训练和叠加,增强预测能力。逻辑回归系数被用来推导放射性特征。结果:SC具有较强的预测效果,受试者工作特征曲线下面积可达0.77。Logistic回归分别确定了所有肿瘤治疗反应和最大肿瘤roi的12个和7个预测特征。结论:我们的研究结果表明,基线CT扫描的放射学特征可以作为预测化疗反应的可靠、可解释的生物标志物,为指导不可切除的CRLM的个性化治疗提供见解。
{"title":"Radiomic signatures from baseline CT predict chemotherapy response in unresectable colorectal liver metastases.","authors":"Mane Piliposyan, Jacob J Peoples, Mohammad Hamghalam, Ramtin Mojtahedi, Kaitlyn Kobayashi, E Claire Bunker, Natalie Gangai, Hyunseon C Kang, Yun Shin Chun, Christian Muise, Richard K G Do, Amber L Simpson","doi":"10.1117/1.JMI.13.1.014505","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014505","url":null,"abstract":"<p><strong>Purpose: </strong>Colorectal cancer is the third most common cancer globally, with a high mortality rate due to metastatic progression, particularly in the liver. Surgical resection remains the main curative treatment, but only a small subset of patients is eligible for surgery at diagnosis. For patients with initially unresectable colorectal liver metastases (CRLM), neoadjuvant chemotherapy can downstage tumors, potentially making surgery feasible. We investigate whether radiomic signatures-quantitative imaging biomarkers derived from baseline computed tomography (CT) scans-can noninvasively predict chemotherapy response in patients with unresectable CRLM, offering a pathway toward personalized treatment planning.</p><p><strong>Approach: </strong>We used radiomics combined with a stacking classifier (SC) to predict treatment outcome. Baseline CT imaging data from 355 patients with initially unresectable CRLM were analyzed using two regions of interest (ROIs) separately (all tumors in the liver and the largest tumor by volume). From each ROI, 107 radiomic features were extracted. The dataset was split into training and testing sets, and multiple machine learning models were trained and integrated via stacking to enhance prediction. Logistic regression coefficients were used to derive radiomic signatures.</p><p><strong>Results: </strong>The SC achieved strong predictive performance, with an area under the receiver operating characteristic curve of up to 0.77 for response prediction. Logistic regression identified 12 and 7 predictive features for treatment response in all tumors and the largest tumor ROIs, respectively.</p><p><strong>Conclusion: </strong>Our findings demonstrate that radiomic features from baseline CT scans can serve as robust, interpretable biomarkers for predicting chemotherapy response, offering insights to guide personalized treatment in unresectable CRLM.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014505"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12797257/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145971472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ER2Net: an evidential reasoning rule-enabled neural network for reliable triple-negative breast cancer tumor segmentation in magnetic resonance imaging. ER2Net:一个支持证据推理规则的神经网络,用于磁共振成像中可靠的三阴性乳腺癌肿瘤分割。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-29 DOI: 10.1117/1.JMI.13.1.014005
Kazi Md Farhad Mahmud, Ahmad Qasem, Joshua M Staley, Rachel Yoder, Allison Aripoli, Shane R Stecklein, Priyanka Sharma, Zhiguo Zhou

Purpose: Triple-negative breast cancer (TNBC) is an aggressive subtype with limited treatment options and high recurrence rates. Magnetic resonance imaging (MRI) is widely used for tumor assessment, but manual segmentation is labor-intensive and variable. Existing deep learning methods often lack generalizability, calibrated confidence, and robust uncertainty quantification.

Approach: We propose ER2Net, an evidential reasoning-enabled neural network for reliable TNBC tumor segmentation on MRI. ER2Net trains multiple U-Net variants with dropouts to generate diverse predictions and introduces pixel-wise reliability to quantify model agreement. We then introduce two ensemble fusion techniques: weighted reliability (WR) segmentation, which leverages pixel-wise reliability to enhance sensitivity, and Bayesian fusion (BF), which integrates predictions probabilistically for robust consensus. Confidence calibration is achieved using evidential reasoning, and we further propose pixel-wise reliable confidence entropy (PWRE) as a uncertainty measure.

Results: ER2Net improved performance compared with individual models. WR achieved IoU = 0.886, sensitivity = 0.928, precision = 0.952, and Hausdorff distance = 5.429 mm, whereas BF achieved IoU = 0.885 and sensitivity = 0.929. Reliable fusion provided the best calibration [expected calibration error = 0.00003; maximum calibration error = 0.017]. PWRE produced lower variance than conventional entropy, yielding more stable uncertainty estimates.

Conclusion: ER2Net introduces WR segmentation and BF as enhanced fusion techniques and PWRE as a uncertainty metric. Together, these advances improve segmentation accuracy, sensitivity, confidence calibration, and uncertainty estimation, paving the way for reliable MRI-based tools to support personalized treatment planning and response assessment in TNBC.

目的:三阴性乳腺癌(TNBC)是一种侵袭性亚型,治疗选择有限,复发率高。磁共振成像(MRI)被广泛应用于肿瘤评估,但人工分割是劳动密集型和可变的。现有的深度学习方法往往缺乏通用性、校准置信度和稳健的不确定性量化。方法:我们提出了ER2Net,一个基于证据推理的神经网络,用于可靠的MRI TNBC肿瘤分割。ER2Net训练多个带有dropout的U-Net变体,以生成不同的预测,并引入像素级可靠性来量化模型一致性。然后,我们介绍了两种集成融合技术:加权可靠性(WR)分割,它利用像素级可靠性来提高灵敏度,以及贝叶斯融合(BF),它以概率方式集成预测以获得稳健的共识。利用证据推理实现置信度校准,并进一步提出像素级可靠置信度熵(PWRE)作为不确定性度量。结果:与单个模型相比,ER2Net提高了性能。WR实现IoU = 0.886,灵敏度= 0.928,精度= 0.952,Hausdorff距离= 5.429 mm, BF实现IoU = 0.885,灵敏度= 0.929。可靠融合提供最佳校准[期望校准误差= 0.00003;最大校准误差= 0.017]。PWRE产生比常规熵更低的方差,产生更稳定的不确定性估计。结论:ER2Net引入了WR分割和BF作为增强融合技术,并将PWRE作为不确定度度量。总之,这些进步提高了分割的准确性、灵敏度、置信度校准和不确定性估计,为可靠的基于mri的工具铺平了道路,以支持TNBC的个性化治疗计划和反应评估。
{"title":"ER<sup>2</sup>Net: an evidential reasoning rule-enabled neural network for reliable triple-negative breast cancer tumor segmentation in magnetic resonance imaging.","authors":"Kazi Md Farhad Mahmud, Ahmad Qasem, Joshua M Staley, Rachel Yoder, Allison Aripoli, Shane R Stecklein, Priyanka Sharma, Zhiguo Zhou","doi":"10.1117/1.JMI.13.1.014005","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014005","url":null,"abstract":"<p><strong>Purpose: </strong>Triple-negative breast cancer (TNBC) is an aggressive subtype with limited treatment options and high recurrence rates. Magnetic resonance imaging (MRI) is widely used for tumor assessment, but manual segmentation is labor-intensive and variable. Existing deep learning methods often lack generalizability, calibrated confidence, and robust uncertainty quantification.</p><p><strong>Approach: </strong>We propose ER<sup>2</sup>Net, an evidential reasoning-enabled neural network for reliable TNBC tumor segmentation on MRI. ER<sup>2</sup>Net trains multiple U-Net variants with dropouts to generate diverse predictions and introduces pixel-wise reliability to quantify model agreement. We then introduce two ensemble fusion techniques: weighted reliability (WR) segmentation, which leverages pixel-wise reliability to enhance sensitivity, and Bayesian fusion (BF), which integrates predictions probabilistically for robust consensus. Confidence calibration is achieved using evidential reasoning, and we further propose pixel-wise reliable confidence entropy (PWRE) as a uncertainty measure.</p><p><strong>Results: </strong>ER<sup>2</sup>Net improved performance compared with individual models. WR achieved IoU = 0.886, sensitivity = 0.928, precision = 0.952, and Hausdorff distance = 5.429 mm, whereas BF achieved IoU = 0.885 and sensitivity = 0.929. Reliable fusion provided the best calibration [expected calibration error = 0.00003; maximum calibration error = 0.017]. PWRE produced lower variance than conventional entropy, yielding more stable uncertainty estimates.</p><p><strong>Conclusion: </strong>ER<sup>2</sup>Net introduces WR segmentation and BF as enhanced fusion techniques and PWRE as a uncertainty metric. Together, these advances improve segmentation accuracy, sensitivity, confidence calibration, and uncertainty estimation, paving the way for reliable MRI-based tools to support personalized treatment planning and response assessment in TNBC.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014005"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12853374/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146107856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient computed tomography-based image segmentation for predicting lateral cervical lymph node metastasis in papillary thyroid carcinoma. 基于ct的高效图像分割预测甲状腺乳头状癌侧颈淋巴结转移。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-13 DOI: 10.1117/1.JMI.13.1.014504
Lei Xu, Bin Zhang, Xingyuan Li, Guona Zheng, Yong Wang, YanHui Peng

Purpose: Papillary thyroid carcinoma (PTC) is a common thyroid cancer, and accurate preoperative assessment of lateral cervical lymph node metastasis is critical for surgical planning. Current methods are often subjective and prone to misdiagnosis. This study aims to improve the accuracy of metastasis evaluation using a deep learning-based segmentation method on enhanced computed tomography (CT) images.

Approach: We propose a YOLOv8-based deep learning model integrated with a deformable self-attention module to enhance metastatic lymph node segmentation. The model was trained on a large dataset of pathology-confirmed CT images from PTC patients.

Results: The model demonstrated diagnostic performance comparable to experienced physicians, with high precision in identifying metastatic nodes. The deformable self-attention module improved segmentation accuracy, with strong sensitivity and specificity.

Conclusion: This deep learning approach improves the accuracy of preoperative assessment for lateral cervical lymph node metastasis in PTC patients, aiding surgical planning, reducing misdiagnosis, and lowering medical costs. It shows promise for enhancing patient outcomes in PTC management.

目的:甲状腺乳头状癌(PTC)是一种常见的甲状腺癌,术前准确评估颈部外侧淋巴结转移对手术计划至关重要。目前的方法往往是主观的,容易误诊。本研究旨在利用基于深度学习的增强计算机断层扫描(CT)图像分割方法提高转移评估的准确性。方法:我们提出了一种基于yolov8的深度学习模型,并集成了一个可变形的自关注模块,以增强转移性淋巴结的分割。该模型是在PTC患者病理证实的CT图像的大型数据集上训练的。结果:该模型表现出与经验丰富的医生相当的诊断性能,在识别转移淋巴结方面具有很高的精度。可变形自关注模块提高了分割精度,具有较强的敏感性和特异性。结论:该深度学习方法提高了PTC患者颈外侧淋巴结转移术前评估的准确性,有助于手术规划,减少误诊,降低医疗费用。它显示了在PTC管理中提高患者预后的希望。
{"title":"Efficient computed tomography-based image segmentation for predicting lateral cervical lymph node metastasis in papillary thyroid carcinoma.","authors":"Lei Xu, Bin Zhang, Xingyuan Li, Guona Zheng, Yong Wang, YanHui Peng","doi":"10.1117/1.JMI.13.1.014504","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014504","url":null,"abstract":"<p><strong>Purpose: </strong>Papillary thyroid carcinoma (PTC) is a common thyroid cancer, and accurate preoperative assessment of lateral cervical lymph node metastasis is critical for surgical planning. Current methods are often subjective and prone to misdiagnosis. This study aims to improve the accuracy of metastasis evaluation using a deep learning-based segmentation method on enhanced computed tomography (CT) images.</p><p><strong>Approach: </strong>We propose a YOLOv8-based deep learning model integrated with a deformable self-attention module to enhance metastatic lymph node segmentation. The model was trained on a large dataset of pathology-confirmed CT images from PTC patients.</p><p><strong>Results: </strong>The model demonstrated diagnostic performance comparable to experienced physicians, with high precision in identifying metastatic nodes. The deformable self-attention module improved segmentation accuracy, with strong sensitivity and specificity.</p><p><strong>Conclusion: </strong>This deep learning approach improves the accuracy of preoperative assessment for lateral cervical lymph node metastasis in PTC patients, aiding surgical planning, reducing misdiagnosis, and lowering medical costs. It shows promise for enhancing patient outcomes in PTC management.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014504"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12797499/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145971502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing deep learning interpretability for hand-crafted feature-guided histologic image classification via weak-to-strong generalization. 通过弱到强的泛化,增强手工制作的特征引导组织学图像分类的深度学习可解释性。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-20 DOI: 10.1117/1.JMI.13.1.017502
Zong Fan, Changjie Lu, Jialin Yue, Mark Anastasio, Lulu Sun, Xiaowei Wang, Hua Li

Purpose: Deep learning (DL) models have achieved promising performance in histologic whole-slide image analysis for various clinical applications. However, their black-box nature hinders the interpretability of DL features for clinical adoption. By contrast, hand-crafted features (HCFs) directly calculated from images offer strong interpretability but with reduced predictive power of DL models. The relationship between DL features and HCFs remains insufficiently explored. We aim to enhance the interpretability and performance of DL models using a weak-to-strong generalization (WSG) framework that integrates HCFs into the learning process.

Approach: The proposed WSG framework leverages an interpretable and HCF-based "weak" teacher model that supervises a "strong" DL student model to learn and improve itself by generalizing from weaker forms of reasoning to stronger ones, for classification tasks. An adaptive bootstrap WSG loss function is designed to optimize the transfer of knowledge from hand-crafted to deep-learned features, enabling systematic analysis of feature interactions. Innovatively, mutual information (MI) between HCFs and DL features learned by student models is analyzed to gain insights into their correlations and the interpretability of DL features. The framework is evaluated using extensive experiments on three public datasets with diverse combinations of teacher and student models for tumor classification.

Results: The WSG framework achieves consistent improvements in classification performance across all evaluated models. Qualitative saliency-map analysis indicates that WSG supervision enables student models to concentrate on diagnostically relevant regions, thereby improving interpretability. Furthermore, quantitative analysis reveals a notable increase in MI between hand-crafted and deep-learned features following WSG training compared with that without WSG training, indicating more effective integration of expert knowledge into the learned representations.

Conclusions: Our study elucidates the key HCFs that drive DL model predictions in histologic image classification. The findings demonstrate that integrating HCFs into DL model training via the WSG framework can enhance both the interpretability and the model's predictive performance, supporting their broader clinical adoption.

目的:深度学习(DL)模型在各种临床应用的组织学全切片图像分析中取得了很好的表现。然而,它们的黑箱性质阻碍了临床采用DL特征的可解释性。相比之下,直接从图像中计算的手工特征(hcf)具有很强的可解释性,但DL模型的预测能力较低。DL特征与hcf之间的关系尚未得到充分探讨。我们的目标是使用将hcf集成到学习过程中的弱到强泛化(WSG)框架来增强DL模型的可解释性和性能。方法:提出的WSG框架利用了一个可解释的、基于hcf的“弱”教师模型,该模型监督一个“强”DL学生模型,通过从较弱的推理形式推广到较强的推理形式来学习和提高自己,以完成分类任务。设计了自适应自举WSG损失函数,优化了从手工特征到深度学习特征的知识转移,实现了特征交互的系统分析。创新性地分析了hcf和学生模型学习的DL特征之间的互信息(MI),以深入了解它们的相关性和DL特征的可解释性。该框架使用三个公共数据集的广泛实验进行评估,这些数据集具有不同的教师和学生模型组合用于肿瘤分类。结果:WSG框架在所有被评估的模型中实现了分类性能的一致改进。定性显著性图分析表明,WSG监督使学生模型能够专注于诊断相关区域,从而提高可解释性。此外,定量分析显示,与未经WSG训练相比,经过WSG训练的手工特征和深度学习特征之间的MI显著增加,表明专家知识更有效地集成到学习表征中。结论:我们的研究阐明了在组织学图像分类中驱动DL模型预测的关键hcf。研究结果表明,通过WSG框架将hcf整合到DL模型训练中可以提高可解释性和模型的预测性能,从而支持其更广泛的临床应用。
{"title":"Enhancing deep learning interpretability for hand-crafted feature-guided histologic image classification via weak-to-strong generalization.","authors":"Zong Fan, Changjie Lu, Jialin Yue, Mark Anastasio, Lulu Sun, Xiaowei Wang, Hua Li","doi":"10.1117/1.JMI.13.1.017502","DOIUrl":"10.1117/1.JMI.13.1.017502","url":null,"abstract":"<p><strong>Purpose: </strong>Deep learning (DL) models have achieved promising performance in histologic whole-slide image analysis for various clinical applications. However, their black-box nature hinders the interpretability of DL features for clinical adoption. By contrast, hand-crafted features (HCFs) directly calculated from images offer strong interpretability but with reduced predictive power of DL models. The relationship between DL features and HCFs remains insufficiently explored. We aim to enhance the interpretability and performance of DL models using a weak-to-strong generalization (WSG) framework that integrates HCFs into the learning process.</p><p><strong>Approach: </strong>The proposed WSG framework leverages an interpretable and HCF-based \"weak\" teacher model that supervises a \"strong\" DL student model to learn and improve itself by generalizing from weaker forms of reasoning to stronger ones, for classification tasks. An adaptive bootstrap WSG loss function is designed to optimize the transfer of knowledge from hand-crafted to deep-learned features, enabling systematic analysis of feature interactions. Innovatively, mutual information (MI) between HCFs and DL features learned by student models is analyzed to gain insights into their correlations and the interpretability of DL features. The framework is evaluated using extensive experiments on three public datasets with diverse combinations of teacher and student models for tumor classification.</p><p><strong>Results: </strong>The WSG framework achieves consistent improvements in classification performance across all evaluated models. Qualitative saliency-map analysis indicates that WSG supervision enables student models to concentrate on diagnostically relevant regions, thereby improving interpretability. Furthermore, quantitative analysis reveals a notable increase in MI between hand-crafted and deep-learned features following WSG training compared with that without WSG training, indicating more effective integration of expert knowledge into the learned representations.</p><p><strong>Conclusions: </strong>Our study elucidates the key HCFs that drive DL model predictions in histologic image classification. The findings demonstrate that integrating HCFs into DL model training via the WSG framework can enhance both the interpretability and the model's predictive performance, supporting their broader clinical adoption.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"017502"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12818695/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146020152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MedPTQ: a practical pipeline for real post-training quantization in 3D medical image segmentation. MedPTQ:一个实用的管道,真正的训练后量化在三维医学图像分割。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-02-17 DOI: 10.1117/1.JMI.13.1.014006
Chongyu Qu, Ritchie Zhao, Ye Yu, Bin Liu, Tianyuan Yao, Junchao Zhu, Bennett A Landman, Yucheng Tang, Yuankai Huo
<p><strong>Purpose: </strong>Quantizing deep neural networks, reducing the precision (bit-width) of their computations, can remarkably decrease memory usage and accelerate processing, making these models more suitable for large-scale medical imaging applications with limited computational resources. However, many existing methods studied "simulated quantization," which simulates lower precision operations during inference but does not actually reduce model size or improve real-world inference speed. Moreover, the potential of deploying real three-dimensional (3D) low-bit quantization on modern graphics processing units (GPUs) is still unexplored.</p><p><strong>Approach: </strong>We introduce MedPTQ, an open-source pipeline for real post-training quantization that implements true 8-bit (INT8) inference on state-of-the-art (SOTA) 3D medical segmentation models, i.e., U-Net, SegResNet, SwinUNETR, nnU-Net, UNesT, TransUNet, ST-UNet, and VISTA3D. MedPTQ involves two main steps. First, we use TensorRT to perform simulated quantization for both weights and activations with an unlabeled calibration dataset. Second, we convert this simulated quantization into real quantization via the TensorRT engine on real GPUs, resulting in real-world reductions in model size and inference latency.</p><p><strong>Results: </strong>Extensive experiments benchmark MedPTQ across seven models and three datasets and demonstrate that it effectively performs INT8 quantization on GPUs, reducing model size by up to 3.83× and latency by up to 2.74×, while maintaining nearly identical Dice similarity coefficient (mDSC) performance to FP32 models. This advancement enables the deployment of efficient deep learning models in medical imaging applications where computational resources are constrained. The MedPTQ code and models have been released, including U-Net, TransUNet pretrained on the BTCV dataset for abdominal (13-label) segmentation, UNesT pretrained on the Whole Brain Dataset for whole brain (133-label) segmentation, and nnU-Net, SegResNet, SwinUNETR, and VISTA3D pretrained on TotalSegmentator V2 for full body (104-label) segmentation.</p><p><strong>Conclusions: </strong>We have introduced MedPTQ, a real post-training quantization pipeline that delivers INT8 inference for SOTA 3D artificial intelligence (AI) models in medical imaging segmentation. MedPTQ effectively reduces real-world model size, computational requirements, and inference latency without compromising segmentation accuracy on modern GPUs, as evidenced by mDSC comparable to full-precision baselines. We validate MedPTQ across a diverse set of AI architectures, ranging from convolutional-neural-network-based to transformer-based models, and a wide variety of medical imaging datasets. These datasets are collected from multiple hospitals with distinct imaging protocols, cover different body regions (such as the brain, abdomen, or full body), and include multiple imaging modalities [computed tomography (CT) and magne
目的:量化深度神经网络,降低其计算精度(位宽),可以显著减少内存使用并加速处理,使这些模型更适合计算资源有限的大规模医学成像应用。然而,许多现有的方法研究的是“模拟量化”,它在推理过程中模拟较低精度的操作,但实际上并没有减小模型大小或提高现实世界的推理速度。此外,在现代图形处理单元(gpu)上部署真正的三维(3D)低比特量化的潜力仍未得到探索。方法:我们引入MedPTQ,一个真正的训练后量化的开源管道,在最先进的(SOTA) 3D医学分割模型上实现真正的8位(INT8)推理,即U-Net, SegResNet, SwinUNETR, nnU-Net, UNesT, TransUNet, ST-UNet和VISTA3D。MedPTQ包括两个主要步骤。首先,我们使用TensorRT对未标记的校准数据集的权重和激活执行模拟量化。其次,我们通过真实gpu上的TensorRT引擎将这种模拟量化转换为真实量化,从而减少了模型大小和推理延迟。结果:在七个模型和三个数据集上对MedPTQ进行了广泛的实验基准测试,并证明它有效地在gpu上执行INT8量化,将模型大小减少了3.83倍,延迟减少了2.74倍,同时保持与FP32模型几乎相同的骰子相似系数(mDSC)性能。这一进步使得在计算资源受限的医学成像应用中部署高效的深度学习模型成为可能。MedPTQ代码和模型已经发布,包括U-Net、TransUNet在BTCV数据集上进行腹部(13个标签)分割的预训练,UNesT在全脑数据集上进行全脑(133个标签)分割的预训练,以及nnU-Net、SegResNet、SwinUNETR和VISTA3D在TotalSegmentator V2上进行全身(104个标签)分割的预训练。结论:我们引入了MedPTQ,一个真正的训练后量化管道,为SOTA 3D人工智能(AI)模型在医学成像分割中提供INT8推理。MedPTQ有效地减少了现实世界的模型大小、计算需求和推理延迟,而不会影响现代gpu上的分割精度,mDSC可以与全精度基线相媲美。我们在不同的人工智能架构中验证MedPTQ,从基于卷积神经网络的模型到基于变压器的模型,以及各种各样的医学成像数据集。这些数据集来自多家具有不同成像方案的医院,涵盖不同的身体区域(如大脑、腹部或全身),并包括多种成像方式[计算机断层扫描(CT)和磁共振成像(MRI)]。总的来说,这些结果突出了我们的MedPTQ在广泛的医学成像任务中具有很强的通用性和适应性。
{"title":"MedPTQ: a practical pipeline for real post-training quantization in 3D medical image segmentation.","authors":"Chongyu Qu, Ritchie Zhao, Ye Yu, Bin Liu, Tianyuan Yao, Junchao Zhu, Bennett A Landman, Yucheng Tang, Yuankai Huo","doi":"10.1117/1.JMI.13.1.014006","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014006","url":null,"abstract":"&lt;p&gt;&lt;strong&gt;Purpose: &lt;/strong&gt;Quantizing deep neural networks, reducing the precision (bit-width) of their computations, can remarkably decrease memory usage and accelerate processing, making these models more suitable for large-scale medical imaging applications with limited computational resources. However, many existing methods studied \"simulated quantization,\" which simulates lower precision operations during inference but does not actually reduce model size or improve real-world inference speed. Moreover, the potential of deploying real three-dimensional (3D) low-bit quantization on modern graphics processing units (GPUs) is still unexplored.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Approach: &lt;/strong&gt;We introduce MedPTQ, an open-source pipeline for real post-training quantization that implements true 8-bit (INT8) inference on state-of-the-art (SOTA) 3D medical segmentation models, i.e., U-Net, SegResNet, SwinUNETR, nnU-Net, UNesT, TransUNet, ST-UNet, and VISTA3D. MedPTQ involves two main steps. First, we use TensorRT to perform simulated quantization for both weights and activations with an unlabeled calibration dataset. Second, we convert this simulated quantization into real quantization via the TensorRT engine on real GPUs, resulting in real-world reductions in model size and inference latency.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Results: &lt;/strong&gt;Extensive experiments benchmark MedPTQ across seven models and three datasets and demonstrate that it effectively performs INT8 quantization on GPUs, reducing model size by up to 3.83× and latency by up to 2.74×, while maintaining nearly identical Dice similarity coefficient (mDSC) performance to FP32 models. This advancement enables the deployment of efficient deep learning models in medical imaging applications where computational resources are constrained. The MedPTQ code and models have been released, including U-Net, TransUNet pretrained on the BTCV dataset for abdominal (13-label) segmentation, UNesT pretrained on the Whole Brain Dataset for whole brain (133-label) segmentation, and nnU-Net, SegResNet, SwinUNETR, and VISTA3D pretrained on TotalSegmentator V2 for full body (104-label) segmentation.&lt;/p&gt;&lt;p&gt;&lt;strong&gt;Conclusions: &lt;/strong&gt;We have introduced MedPTQ, a real post-training quantization pipeline that delivers INT8 inference for SOTA 3D artificial intelligence (AI) models in medical imaging segmentation. MedPTQ effectively reduces real-world model size, computational requirements, and inference latency without compromising segmentation accuracy on modern GPUs, as evidenced by mDSC comparable to full-precision baselines. We validate MedPTQ across a diverse set of AI architectures, ranging from convolutional-neural-network-based to transformer-based models, and a wide variety of medical imaging datasets. These datasets are collected from multiple hospitals with distinct imaging protocols, cover different body regions (such as the brain, abdomen, or full body), and include multiple imaging modalities [computed tomography (CT) and magne","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014006"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12912285/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146221578","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BAF-UNet: a boundary-aware segmentation model for skin lesion segmentation. BAF-UNet:一种边界感知的皮肤损伤分割模型。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-08 DOI: 10.1117/1.JMI.13.1.014003
Menglei Zhang, Congwei Zhang, Zhibin Quan, Bing Guo, Wankou Yang

Purpose: Skin lesion segmentation plays a significant role in the diagnosis and treatment of skin cancer. Accurate skin lesion segmentation is essential for skin cancer diagnosis and treatment but is challenged by ambiguous boundaries and diverse lesion shapes and sizes. We aim to improve segmentation performance with enhanced boundary preservation.

Approach: We propose BAF-UNet, a boundary-aware segmentation network. It integrates a multiscale boundary-aware feature fusion (BFF) module to combine low-level boundary features with high-level semantic information, and a boundary-aware vision transformer (BAViT) that incorporates boundary guidance into MobileViT to capture local and global context. A boundary-focused loss function is also introduced to prioritize edge accuracy during training. The model is evaluated on ISIC2016, ISIC2017, and PH2 datasets.

Results: Experiments demonstrate that BAF-UNet improves Dice scores and boundary accuracy compared to baseline models. The BFF and BAViT modules enhance boundary delineation while maintaining robustness across lesions of varying shapes and sizes.

Conclusions: BAF-UNet effectively integrates boundary guidance into feature fusion and transformer-based context modeling, significantly improving segmentation accuracy, particularly along lesion edges, and shows potential for clinical application in automated skin cancer diagnosis.

目的:皮肤病变分割在皮肤癌的诊断和治疗中起着重要的作用。准确的皮肤病变分割对于皮肤癌的诊断和治疗至关重要,但由于皮肤病变边界模糊、形状和大小不一而受到挑战。我们的目标是通过增强边界保持来提高分割性能。方法:我们提出了一种边界感知分割网络BAF-UNet。它集成了一个多尺度边界感知特征融合(BFF)模块,将低级边界特征与高级语义信息相结合,以及一个边界感知视觉转换器(BAViT),将边界引导集成到MobileViT中,以捕获局部和全局上下文。在训练过程中引入边界聚焦损失函数对边缘精度进行优先排序。在ISIC2016、ISIC2017和PH2数据集上对模型进行了评估。结果:实验表明,与基线模型相比,BAF-UNet提高了Dice分数和边界精度。BFF和BAViT模块增强了边界划分,同时保持了不同形状和大小病变的稳健性。结论:BAF-UNet有效地将边界引导集成到特征融合和基于变压器的上下文建模中,显著提高了分割精度,特别是沿病灶边缘的分割精度,在皮肤癌自动诊断中具有临床应用潜力。
{"title":"BAF-UNet: a boundary-aware segmentation model for skin lesion segmentation.","authors":"Menglei Zhang, Congwei Zhang, Zhibin Quan, Bing Guo, Wankou Yang","doi":"10.1117/1.JMI.13.1.014003","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.014003","url":null,"abstract":"<p><strong>Purpose: </strong>Skin lesion segmentation plays a significant role in the diagnosis and treatment of skin cancer. Accurate skin lesion segmentation is essential for skin cancer diagnosis and treatment but is challenged by ambiguous boundaries and diverse lesion shapes and sizes. We aim to improve segmentation performance with enhanced boundary preservation.</p><p><strong>Approach: </strong>We propose BAF-UNet, a boundary-aware segmentation network. It integrates a multiscale boundary-aware feature fusion (BFF) module to combine low-level boundary features with high-level semantic information, and a boundary-aware vision transformer (BAViT) that incorporates boundary guidance into MobileViT to capture local and global context. A boundary-focused loss function is also introduced to prioritize edge accuracy during training. The model is evaluated on ISIC2016, ISIC2017, and PH2 datasets.</p><p><strong>Results: </strong>Experiments demonstrate that BAF-UNet improves Dice scores and boundary accuracy compared to baseline models. The BFF and BAViT modules enhance boundary delineation while maintaining robustness across lesions of varying shapes and sizes.</p><p><strong>Conclusions: </strong>BAF-UNet effectively integrates boundary guidance into feature fusion and transformer-based context modeling, significantly improving segmentation accuracy, particularly along lesion edges, and shows potential for clinical application in automated skin cancer diagnosis.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014003"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12782429/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145953505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultrasound imaging using single-element biaxial beamforming. 超声成像使用单元件双轴波束成形。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2025-12-23 DOI: 10.1117/1.JMI.13.1.017001
Nathan Meulenbroek, Laura Curiel, Adam Waspe, Samuel Pichardo

Purpose: Dynamic focusing of received ultrasound signals, or beamforming, is foundational for ultrasound imaging. Conventionally, it requires arrays of ultrasound sensors to estimate where sound came from using time-of-flight (TOF) measurements. We demonstrate passive beamforming with a single biaxial sensor and accurate passive acoustic mapping with two biaxial sensors using only direction of arrival (DOA) information.

Approach: We introduce two single-element biaxial beamforming algorithms and four biaxial image reconstruction algorithms for a two-element biaxial piezoceramic transducer array. Imaging of a hemispherical acoustic source is characterized in an acoustic scanning tank within the region - 30.29    mm x 29.94 mm and 50.11 mm z 90.45 mm relative to the center of the array. Imaging performance is contrasted with delay, sum, and integrate (DSAI) and delay, multiply, sum, and integrate (DMSAI) algorithms.

Results: Single-element biaxial beamforming can identify DOA with a median error (± interquartile range) of 0.36 ± 0.63    deg and median full-width half-prominence of 7.3 ± 8.6    deg . Using both array elements, DOA-only images demonstrate overall median localization error of 6.41 mm (lateral: 1.02 mm, axial: 5.85 mm, signal-to-noise ratio (SNR): 15.37) and DOA + TOF images demonstrate overall median error of 6.91 mm (lateral: 1.69 mm, axial: 6.11 mm, SNR: 18.37).

Conclusions: To the best of our knowledge, we provide the first demonstration of single-element beamforming using a single stationary piezoceramic and the first demonstration of passive ultrasound imaging without the use of TOF information. These results enable simpler, smaller, more cost-effective arrays for passive ultrasound imaging.

目的:接收超声信号的动态聚焦或波束形成是超声成像的基础。传统上,它需要超声波传感器阵列来估计声音来自何处,使用飞行时间(TOF)测量。我们演示了使用单个双轴传感器的被动波束形成和仅使用到达方向(DOA)信息的两个双轴传感器的精确被动声学映射。方法:针对两元双轴压电换能器阵列,介绍了两种单元双轴波束形成算法和四种双轴图像重建算法。在相对于阵列中心的- 30.29 mm≤x≤29.94 mm和50.11 mm≤z≤90.45 mm区域内,对半球形声源进行成像。对比了延迟、求和和积分(DSAI)算法和延迟、乘法、求和和积分(DMSAI)算法的成像性能。结果:单单元双轴波束形成识别DOA的中位误差(±四分位间距)为0.36±0.63°,中位全宽半凸度为7.3±8.6°。使用这两种阵列元素,仅DOA图像的总体中位数定位误差为6.41 mm(侧向:1.02 mm,轴向:5.85 mm,信噪比(SNR): 15.37), DOA + TOF图像的总体中位数定位误差为6.91 mm(侧向:1.69 mm,轴向:6.11 mm,信噪比:18.37)。结论:据我们所知,我们提供了第一个使用单个静止压电陶瓷的单元件波束形成演示,以及第一个不使用TOF信息的被动超声成像演示。这些结果使被动超声成像阵列更简单、更小、更具成本效益。
{"title":"Ultrasound imaging using single-element biaxial beamforming.","authors":"Nathan Meulenbroek, Laura Curiel, Adam Waspe, Samuel Pichardo","doi":"10.1117/1.JMI.13.1.017001","DOIUrl":"https://doi.org/10.1117/1.JMI.13.1.017001","url":null,"abstract":"<p><strong>Purpose: </strong>Dynamic focusing of received ultrasound signals, or beamforming, is foundational for ultrasound imaging. Conventionally, it requires arrays of ultrasound sensors to estimate where sound came from using time-of-flight (TOF) measurements. We demonstrate passive beamforming with a single biaxial sensor and accurate passive acoustic mapping with two biaxial sensors using only direction of arrival (DOA) information.</p><p><strong>Approach: </strong>We introduce two single-element biaxial beamforming algorithms and four biaxial image reconstruction algorithms for a two-element biaxial piezoceramic transducer array. Imaging of a hemispherical acoustic source is characterized in an acoustic scanning tank within the region <math><mrow><mo>-</mo> <mn>30.29</mn> <mtext>  </mtext> <mi>mm</mi></mrow> </math> <math><mrow><mo>≤</mo> <mi>x</mi> <mo>≤</mo></mrow> </math> 29.94 mm and 50.11 mm <math><mrow><mo>≤</mo> <mi>z</mi> <mo>≤</mo></mrow> </math> 90.45 mm relative to the center of the array. Imaging performance is contrasted with delay, sum, and integrate (DSAI) and delay, multiply, sum, and integrate (DMSAI) algorithms.</p><p><strong>Results: </strong>Single-element biaxial beamforming can identify DOA with a median error (± interquartile range) of <math><mrow><mn>0.36</mn> <mo>±</mo> <mn>0.63</mn> <mtext>  </mtext> <mi>deg</mi></mrow> </math> and median full-width half-prominence of <math><mrow><mn>7.3</mn> <mo>±</mo> <mn>8.6</mn> <mtext>  </mtext> <mi>deg</mi></mrow> </math> . Using both array elements, DOA-only images demonstrate overall median localization error of 6.41 mm (lateral: 1.02 mm, axial: 5.85 mm, signal-to-noise ratio (SNR): 15.37) and DOA + TOF images demonstrate overall median error of 6.91 mm (lateral: 1.69 mm, axial: 6.11 mm, SNR: 18.37).</p><p><strong>Conclusions: </strong>To the best of our knowledge, we provide the first demonstration of single-element beamforming using a single stationary piezoceramic and the first demonstration of passive ultrasound imaging without the use of TOF information. These results enable simpler, smaller, more cost-effective arrays for passive ultrasound imaging.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"017001"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12726554/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145828543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synthetic multi-inversion time magnetic resonance images for visualization of subcortical structures. 用于皮质下结构可视化的合成多逆时间磁共振图像。
IF 1.7 Q3 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING Pub Date : 2026-01-01 Epub Date: 2026-01-06 DOI: 10.1117/1.JMI.13.1.014002
Savannah P Hays, Lianrui Zuo, Anqi Feng, Yihao Liu, Blake E Dewey, Jiachen Zhuo, Ellen M Mowry, Scott D Newsome, Jerry L Prince, Aaron Carass

Purpose: Visualization of subcortical gray matter is essential in neuroscience and clinical practice, particularly for disease understanding and surgical planning. Although multi-inversion time (multi-TI) T 1 -weighted ( T 1 -w) magnetic resonance (MR) imaging improves visualization, it is only acquired in specific clinical settings and not available in common public MR datasets.

Approach: We present SyMTIC (synthetic multi-TI contrasts), a deep learning method that generates synthetic multi-TI images using routinely acquired T 1 -w, T 2 -weighted ( T 2 -w), and fluid-attenuated inversion recovery (FLAIR) images. Our approach combines image translation via deep neural networks with imaging physics to estimate longitudinal relaxation time ( T 1 ) and proton density ( ρ ) maps. These maps are then used to compute multi-TI images with arbitrary inversion times.

Results: SyMTIC was trained using paired magnetization prepared rapid acquisition with gradient echo (MPRAGE) and fast gray matter acquisition T1 inversion recovery (FGATIR) images along with T 2 -w and FLAIR images. It accurately synthesized multi-TI images from standard clinical inputs, achieving image quality comparable to that from explicitly acquired multi-TI data. The synthetic images, especially for TI values between 400 to 800 ms, enhanced visualization of subcortical structures and improved segmentation of thalamic nuclei.

Conclusion: SyMTIC enables robust generation of high-quality multi-TI images from routine MR contrasts. When paired with the HACA3 algorithm, it generalizes well to varied clinical datasets, including those without FLAIR or T 2 -w images and unknown parameters, offering a practical solution for improving brain MR image visualization and analysis.

目的:皮层下灰质的可视化在神经科学和临床实践中是必不可少的,特别是对于疾病的理解和手术计划。虽然多次反转时间(multi-TI) t1加权(t1 -w)磁共振(MR)成像改善了可视化,但它仅在特定的临床环境中获得,而在公共MR数据集中不可用。方法:我们提出了SyMTIC(合成多ti对比),这是一种深度学习方法,使用常规获取的t1 -w、t2加权(t2 -w)和流体衰减反演恢复(FLAIR)图像生成合成多ti图像。我们的方法将通过深度神经网络的图像平移与成像物理相结合,以估计纵向松弛时间(t1)和质子密度(ρ)图。然后使用这些映射来计算具有任意反转时间的多ti图像。结果:SyMTIC使用配对磁化制备的梯度回波快速采集(MPRAGE)和快速灰质采集T1反演恢复(FGATIR)图像以及t2 -w和FLAIR图像进行训练。它准确地合成了来自标准临床输入的多ti图像,实现了与明确获取的多ti数据相当的图像质量。合成图像增强了皮层下结构的可视化,改善了丘脑核的分割,特别是在400 ~ 800 ms之间。结论:SyMTIC能够从常规MR对比中生成高质量的多ti图像。当与HACA3算法配对时,它可以很好地推广到各种临床数据集,包括那些没有FLAIR或t2 -w图像和未知参数的数据集,为提高脑MR图像的可视化和分析提供了一个实用的解决方案。
{"title":"Synthetic multi-inversion time magnetic resonance images for visualization of subcortical structures.","authors":"Savannah P Hays, Lianrui Zuo, Anqi Feng, Yihao Liu, Blake E Dewey, Jiachen Zhuo, Ellen M Mowry, Scott D Newsome, Jerry L Prince, Aaron Carass","doi":"10.1117/1.JMI.13.1.014002","DOIUrl":"10.1117/1.JMI.13.1.014002","url":null,"abstract":"<p><strong>Purpose: </strong>Visualization of subcortical gray matter is essential in neuroscience and clinical practice, particularly for disease understanding and surgical planning. Although multi-inversion time (multi-TI) <math> <mrow><msub><mi>T</mi> <mn>1</mn></msub> </mrow> </math> -weighted ( <math> <mrow><msub><mi>T</mi> <mn>1</mn></msub> </mrow> </math> -w) magnetic resonance (MR) imaging improves visualization, it is only acquired in specific clinical settings and not available in common public MR datasets.</p><p><strong>Approach: </strong>We present SyMTIC (synthetic multi-TI contrasts), a deep learning method that generates synthetic multi-TI images using routinely acquired <math> <mrow><msub><mi>T</mi> <mn>1</mn></msub> </mrow> </math> -w, <math> <mrow><msub><mi>T</mi> <mn>2</mn></msub> </mrow> </math> -weighted ( <math> <mrow><msub><mi>T</mi> <mn>2</mn></msub> </mrow> </math> -w), and fluid-attenuated inversion recovery (FLAIR) images. Our approach combines image translation via deep neural networks with imaging physics to estimate longitudinal relaxation time ( <math><mrow><mi>T</mi> <mn>1</mn></mrow> </math> ) and proton density ( <math><mrow><mi>ρ</mi></mrow> </math> ) maps. These maps are then used to compute multi-TI images with arbitrary inversion times.</p><p><strong>Results: </strong>SyMTIC was trained using paired magnetization prepared rapid acquisition with gradient echo (MPRAGE) and fast gray matter acquisition T1 inversion recovery (FGATIR) images along with <math> <mrow><msub><mi>T</mi> <mn>2</mn></msub> </mrow> </math> -w and FLAIR images. It accurately synthesized multi-TI images from standard clinical inputs, achieving image quality comparable to that from explicitly acquired multi-TI data. The synthetic images, especially for TI values between 400 to 800 ms, enhanced visualization of subcortical structures and improved segmentation of thalamic nuclei.</p><p><strong>Conclusion: </strong>SyMTIC enables robust generation of high-quality multi-TI images from routine MR contrasts. When paired with the HACA3 algorithm, it generalizes well to varied clinical datasets, including those without FLAIR or <math> <mrow><msub><mi>T</mi> <mn>2</mn></msub> </mrow> </math> -w images and unknown parameters, offering a practical solution for improving brain MR image visualization and analysis.</p>","PeriodicalId":47707,"journal":{"name":"Journal of Medical Imaging","volume":"13 1","pages":"014002"},"PeriodicalIF":1.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12770912/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145918841","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Medical Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1