首页 > 最新文献

Journal of imaging informatics in medicine最新文献

英文 中文
Comparison of Image Quality Reconstructed Using Iterative Reconstruction and Deep Learning Algorithms Under Varying Dose Reductions in Dual-Energy Carotid CT Angiography. 双能颈动脉CT血管造影不同剂量下迭代重建与深度学习算法重建图像质量的比较
Pub Date : 2026-02-04 DOI: 10.1007/s10278-026-01848-9
Chenzi Wang, Juan Long, Dapeng Zhang, Lulu Fan, Zhen Wang, Xiaohan Liu, He Zhang, Chong Wang, Yang Wu, Aiyun Sun, Kai Xu, Yankai Meng

Carotid CT angiography (CTA) is valuable for diagnosing carotid artery disease but involves radiation and contrast agent risks. Deep Learning Image Reconstruction (DLIR-H) shows potential for maintaining image quality in low-dose protocols. In this prospective study, 180 patients undergoing dual-energy CTA were divided into three groups: a control group (ASIR-V 50%, NI = 4, contrast = 0.5 mL/kg), a low-dose group (DLIR-H, NI = 11, contrast = 0.5 mL/kg), and an ultra-low-dose group (DLIR-H, NI = 13, contrast = 0.4 mL/kg). Objective (CTV[CT values], noise, SNR, CNR) and subjective (5-point Likert scale) image quality were evaluated. The ultra-low-dose group achieved a 20.3% reduction in contrast volume and a 53.3% reduction in effective dose compared to the control group (P < 0.001). Both experimental groups showed lower noise and higher CNR/SNR (except at aortic arch) than controls. However, the ultra-low-dose group had significantly lower CNR/SNR than the low-dose group (P < 0.05). Subjective image quality was superior in both experimental groups (P < 0.001), with high inter-rater agreement. DLIR-H outperformed ASIR-V in low and ultra-low-dose protocols but could not fully compensate for image quality degradation when radiation and contrast were further reduced.

颈动脉CT血管造影(CTA)对诊断颈动脉疾病很有价值,但涉及辐射和造影剂风险。深度学习图像重建(DLIR-H)显示了在低剂量协议下保持图像质量的潜力。本前瞻性研究将180例接受双能CTA治疗的患者分为3组:对照组(ASIR-V 50%, NI = 4,反差= 0.5 mL/kg)、低剂量组(DLIR-H, NI = 11,反差= 0.5 mL/kg)和超低剂量组(DLIR-H, NI = 13,反差= 0.4 mL/kg)。评价客观图像质量(CTV[CT值]、噪声、信噪比、CNR)和主观图像质量(5点李克特量表)。与对照组相比,超低剂量组造影剂体积减少20.3%,有效剂量减少53.3% (P
{"title":"Comparison of Image Quality Reconstructed Using Iterative Reconstruction and Deep Learning Algorithms Under Varying Dose Reductions in Dual-Energy Carotid CT Angiography.","authors":"Chenzi Wang, Juan Long, Dapeng Zhang, Lulu Fan, Zhen Wang, Xiaohan Liu, He Zhang, Chong Wang, Yang Wu, Aiyun Sun, Kai Xu, Yankai Meng","doi":"10.1007/s10278-026-01848-9","DOIUrl":"https://doi.org/10.1007/s10278-026-01848-9","url":null,"abstract":"<p><p>Carotid CT angiography (CTA) is valuable for diagnosing carotid artery disease but involves radiation and contrast agent risks. Deep Learning Image Reconstruction (DLIR-H) shows potential for maintaining image quality in low-dose protocols. In this prospective study, 180 patients undergoing dual-energy CTA were divided into three groups: a control group (ASIR-V 50%, NI = 4, contrast = 0.5 mL/kg), a low-dose group (DLIR-H, NI = 11, contrast = 0.5 mL/kg), and an ultra-low-dose group (DLIR-H, NI = 13, contrast = 0.4 mL/kg). Objective (CTV[CT values], noise, SNR, CNR) and subjective (5-point Likert scale) image quality were evaluated. The ultra-low-dose group achieved a 20.3% reduction in contrast volume and a 53.3% reduction in effective dose compared to the control group (P < 0.001). Both experimental groups showed lower noise and higher CNR/SNR (except at aortic arch) than controls. However, the ultra-low-dose group had significantly lower CNR/SNR than the low-dose group (P < 0.05). Subjective image quality was superior in both experimental groups (P < 0.001), with high inter-rater agreement. DLIR-H outperformed ASIR-V in low and ultra-low-dose protocols but could not fully compensate for image quality degradation when radiation and contrast were further reduced.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146121775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection of Swallowing Abnormalities in Pediatric FEES Recordings Using Rule-Based and Model-Based Methods. 使用基于规则和基于模型的方法检测儿科收费记录中的吞咽异常。
Pub Date : 2026-02-03 DOI: 10.1007/s10278-026-01845-y
Soolmaz Abbasi, Hisham Al-Kassem, Hamdy El-Hakim, Jacob Jaremko, Abhilash Hareendranathan

Pediatric swallowing dysfunction (SwD) poses serious health risks, including aspiration, malnutrition, and recurrent respiratory infections, making early and accurate diagnosis essential for preventing long-term sequelae such as chronic lung disease and growth failure. Fiberoptic endoscopic evaluation of swallowing (FEES) is widely used for direct visualization of the swallowing mechanism in children, offering advantages over fluoroscopy such as bedside accessibility and radiation-free imaging. During FEES, patients swallow green-dyed liquid with an endoscope positioned in the throat. Interpreting FEES recordings is a subjective, time-consuming process that requires specialized expertise. Automated, objective analysis tools would be useful to support clinical decision-making. In this study, we propose a hybrid framework for classifying pediatric FEES recordings as normal or abnormal. The approach combines a rule-based analysis which detects the green-tinted swallowed liquid, with a transformer-based deep learning model. Frames are first filtered using a Siamese network to exclude irrelevant or low-quality frames, followed by quantification of the green frame ratio based on frames containing green patches. A confidence-guided decision strategy classifies clear-cut cases via thresholding, while delegating uncertain cases to the deep learning model for further evaluation. Evaluation on 142 pediatric FEES videos (45 normal and 97 with abnormalities) showed that the hybrid approach outperformed both the deep learning and rule-based methods individually, achieving 89.4% accuracy, 96.6% precision, and 93.3% specificity for aspiration. Our results indicate that by combining rule-based and deep learning strategies, we could reliably detect swallowing abnormalities from pediatric FEES videos with accuracies comparable to experts.

小儿吞咽功能障碍(SwD)造成严重的健康风险,包括误吸、营养不良和反复呼吸道感染,因此早期准确诊断对于预防慢性肺部疾病和生长衰竭等长期后遗症至关重要。光纤内镜下吞咽评估(FEES)被广泛用于儿童吞咽机制的直接可视化,比透视检查具有床边可及性和无辐射成像等优点。在收费期间,患者在喉部的内窥镜下吞下绿色的液体。口译费录音是一个主观的、耗时的过程,需要专业知识。自动化、客观的分析工具将有助于支持临床决策。在这项研究中,我们提出了一个混合框架分类儿科收费记录正常或异常。该方法结合了基于规则的分析,检测绿色的吞咽液体,以及基于转换器的深度学习模型。首先使用Siamese网络对帧进行过滤,以排除不相关或低质量的帧,然后根据包含绿色补丁的帧量化绿色帧率。一种置信度引导的决策策略通过阈值对明确的案例进行分类,同时将不确定的案例委托给深度学习模型进行进一步评估。对142个儿科FEES视频(45个正常,97个异常)的评估表明,混合方法分别优于深度学习和基于规则的方法,准确率为89.4%,精密度为96.6%,吸入性为93.3%。我们的研究结果表明,通过结合基于规则和深度学习策略,我们可以可靠地从儿科FEES视频中检测出吞咽异常,其准确性与专家相当。
{"title":"Detection of Swallowing Abnormalities in Pediatric FEES Recordings Using Rule-Based and Model-Based Methods.","authors":"Soolmaz Abbasi, Hisham Al-Kassem, Hamdy El-Hakim, Jacob Jaremko, Abhilash Hareendranathan","doi":"10.1007/s10278-026-01845-y","DOIUrl":"https://doi.org/10.1007/s10278-026-01845-y","url":null,"abstract":"<p><p>Pediatric swallowing dysfunction (SwD) poses serious health risks, including aspiration, malnutrition, and recurrent respiratory infections, making early and accurate diagnosis essential for preventing long-term sequelae such as chronic lung disease and growth failure. Fiberoptic endoscopic evaluation of swallowing (FEES) is widely used for direct visualization of the swallowing mechanism in children, offering advantages over fluoroscopy such as bedside accessibility and radiation-free imaging. During FEES, patients swallow green-dyed liquid with an endoscope positioned in the throat. Interpreting FEES recordings is a subjective, time-consuming process that requires specialized expertise. Automated, objective analysis tools would be useful to support clinical decision-making. In this study, we propose a hybrid framework for classifying pediatric FEES recordings as normal or abnormal. The approach combines a rule-based analysis which detects the green-tinted swallowed liquid, with a transformer-based deep learning model. Frames are first filtered using a Siamese network to exclude irrelevant or low-quality frames, followed by quantification of the green frame ratio based on frames containing green patches. A confidence-guided decision strategy classifies clear-cut cases via thresholding, while delegating uncertain cases to the deep learning model for further evaluation. Evaluation on 142 pediatric FEES videos (45 normal and 97 with abnormalities) showed that the hybrid approach outperformed both the deep learning and rule-based methods individually, achieving 89.4% accuracy, 96.6% precision, and 93.3% specificity for aspiration. Our results indicate that by combining rule-based and deep learning strategies, we could reliably detect swallowing abnormalities from pediatric FEES videos with accuracies comparable to experts.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146115407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Performance Evaluation of a Commercial Deep Learning Software for Detecting Intracranial Hemorrhage in a Pediatric Population. 商业深度学习软件在儿科颅内出血检测中的性能评估。
Pub Date : 2026-02-02 DOI: 10.1007/s10278-026-01857-8
Hadiseh Kavandi, Kyle Costenbader, Sandrine Yazbek, Peter Kamel, Noushin Yahyavi-Firouz-Abadi, Jean Jeudy

This study evaluates a commercially available AI tool (Aidoc) for intracranial hemorrhage (ICH) detection-originally trained on adults-in pediatric patients, addressing the critical need for timely diagnosis and current research gaps in pediatric AI applications. This single-center, retrospective study included pediatric patients aged 6-17 who underwent head CT between January 2017 and November 2022. Radiological reports (unaided by AI) and CT images were analyzed by natural language processing (NLP) and image-based algorithms, respectively, to classify ICH presence or absence. Ground truth was assumed for concordant cases. Three radiologists independently reviewed discrepant cases using majority vote. Among 2502 pediatric patients undergoing head CT, the AI algorithm flagged 292 cases as suspected ICH-positive. A total of 174 discordant cases between NLP and AI were independently reviewed to create the reference standard. Results showed 144 true positives, 6 false negatives, 148 false positives, and 2204 true negatives, yielding sensitivity of 96.0% (91.5-98.5%) and specificity of 93.7% (92.6-94.7%). Overall algorithm accuracy was 93.8% (92.8-94.8%). The most frequent false positives were choroid plexus calcifications and hyperdense venous sinuses, while subdural hemorrhages accounted for most false negatives. This deep learning AI algorithm trained on adult data performs well in detecting pediatric ICH, with 96.0% sensitivity and 93.7% specificity. However, common false positives, choroid plexus calcifications and hyperdense venous sinuses, reflect pediatric-specific features, while missed subdural hemorrhages mirror known adult limitations. Results highlight the need for pediatric-focused AI training to improve diagnostic accuracy in this underserved population.

本研究评估了一种用于儿科患者颅内出血(ICH)检测的市售人工智能工具(Aidoc),该工具最初针对成人进行培训,解决了儿科人工智能应用中及时诊断的迫切需求和当前的研究空白。这项单中心回顾性研究纳入了2017年1月至2022年11月期间接受头部CT检查的6-17岁儿童患者。通过自然语言处理(NLP)和基于图像的算法分析放射报告(人工智能辅助下)和CT图像,分别对ICH的存在或不存在进行分类。在一致的情况下假定基本事实。三名放射科医生使用多数投票方式独立审查了差异病例。在2502例接受头部CT的儿童患者中,人工智能算法标记了292例疑似ich阳性。NLP和AI之间共有174个不一致的案例被独立审查以创建参考标准。结果真阳性144例,假阴性6例,假阳性148例,真阴性2204例,敏感性为96.0%(91.5 ~ 98.5%),特异性为93.7%(92.6 ~ 94.7%)。总体算法准确率为93.8%(92.8 ~ 94.8%)。最常见的假阳性是脉络膜丛钙化和高密度静脉窦,而最常见的假阴性是硬膜下出血。这种基于成人数据训练的深度学习人工智能算法在检测儿童脑出血方面表现良好,灵敏度为96.0%,特异性为93.7%。然而,常见的假阳性,脉络膜丛钙化和高密度静脉窦,反映了儿科的特异性特征,而遗漏的硬膜下出血反映了已知的成人局限性。结果强调需要以儿科为重点的人工智能培训,以提高这一服务不足人群的诊断准确性。
{"title":"Performance Evaluation of a Commercial Deep Learning Software for Detecting Intracranial Hemorrhage in a Pediatric Population.","authors":"Hadiseh Kavandi, Kyle Costenbader, Sandrine Yazbek, Peter Kamel, Noushin Yahyavi-Firouz-Abadi, Jean Jeudy","doi":"10.1007/s10278-026-01857-8","DOIUrl":"https://doi.org/10.1007/s10278-026-01857-8","url":null,"abstract":"<p><p>This study evaluates a commercially available AI tool (Aidoc) for intracranial hemorrhage (ICH) detection-originally trained on adults-in pediatric patients, addressing the critical need for timely diagnosis and current research gaps in pediatric AI applications. This single-center, retrospective study included pediatric patients aged 6-17 who underwent head CT between January 2017 and November 2022. Radiological reports (unaided by AI) and CT images were analyzed by natural language processing (NLP) and image-based algorithms, respectively, to classify ICH presence or absence. Ground truth was assumed for concordant cases. Three radiologists independently reviewed discrepant cases using majority vote. Among 2502 pediatric patients undergoing head CT, the AI algorithm flagged 292 cases as suspected ICH-positive. A total of 174 discordant cases between NLP and AI were independently reviewed to create the reference standard. Results showed 144 true positives, 6 false negatives, 148 false positives, and 2204 true negatives, yielding sensitivity of 96.0% (91.5-98.5%) and specificity of 93.7% (92.6-94.7%). Overall algorithm accuracy was 93.8% (92.8-94.8%). The most frequent false positives were choroid plexus calcifications and hyperdense venous sinuses, while subdural hemorrhages accounted for most false negatives. This deep learning AI algorithm trained on adult data performs well in detecting pediatric ICH, with 96.0% sensitivity and 93.7% specificity. However, common false positives, choroid plexus calcifications and hyperdense venous sinuses, reflect pediatric-specific features, while missed subdural hemorrhages mirror known adult limitations. Results highlight the need for pediatric-focused AI training to improve diagnostic accuracy in this underserved population.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146109458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Systematic Review: Agentic AI in Neuroradiology: Technical Promise with Limited Clinical Evidence. 系统评价:神经放射学中的人工智能:有限临床证据的技术前景。
Pub Date : 2026-02-02 DOI: 10.1007/s10278-025-01839-2
Sara Salehi, Varekan Keishing, Yashbir Singh, David Wei, Amirali Khosravi, Parnian Habibi, Jaidip Jagtap, Bradley J Erickson

Agentic artificial intelligence systems featuring iterative reasoning, autonomous tool use, or multi-agent collaboration have been proposed as solutions to the limitations of large language models (LLMs) in neuroradiology. However, the extent of their implementation and clinical validation remains unclear. We systematically searched PubMed, Web of Science, and Scopus (January 2022-August 2025) for studies implementing agentic AI in neuroradiology. Six independent reviewers (three medical doctors and three AI specialists) assessed full texts. Agentic AI was defined as requiring mandatory iterative reasoning plus either autonomous tool use or multi-agent collaboration. Study quality was evaluated using adapted QUADAS-AI criteria. From 230 records, 9 studies (3.90%) met inclusion criteria. Of these, five (55.60%) implemented true multi-agent architecture, two (22.20%) used hybrid or conceptual frameworks, and two (22.20%) relied on single-model LLMs without genuine agentic behavior. All nine studies were single center with no external validation. Sample sizes were small (median 142 cases; range 16-302). The only randomized controlled trial-INSPIRE (neurophysiology with imaging correlation)-demonstrated high technical performance (≈92% accuracy; AIGERS 0.94 for AI-assisted vs. 0.70 for AI-only, p < 0.001) but showed no measurable clinical benefit when physicians used AI assistance compared with independent reporting. Safety assessments were absent from all studies. Agentic AI in neuroradiology remains technically promising but clinically unproven. Severe evidence scarcity (3.90% inclusion rate), frequent overextension of the "agentic" label (30% of studies lacked genuine autonomy), and the persistent gap between technical performance and clinical utility indicate that the field remains in its early research phase. Current evidence is insufficient to support clinical deployment. Rigorous, multi-center prospective trials with patient-centered and safety outcomes are essential before clinical implementation can be responsibly considered.

具有迭代推理、自主工具使用或多智能体协作的代理人工智能系统已被提出作为神经放射学中大型语言模型(llm)局限性的解决方案。然而,它们的实施程度和临床验证仍不清楚。我们系统地检索了PubMed、Web of Science和Scopus(2022年1月- 2025年8月),寻找在神经放射学中实施代理人工智能的研究。6名独立审稿人(3名医生和3名人工智能专家)评估了全文。人工智能被定义为需要强制迭代推理加上自主工具使用或多智能体协作。采用适应性QUADAS-AI标准评估研究质量。230条记录中,9项研究(3.90%)符合纳入标准。其中,5个(55.60%)实现了真正的多智能体架构,2个(22.20%)使用混合或概念框架,2个(22.20%)依赖于没有真正代理行为的单模型llm。所有9项研究均为单中心,没有外部验证。样本量较小(中位数142例,范围16-302例)。唯一的随机对照试验- inspire(神经生理学与成像相关性)-显示出高技术性能(≈92%的准确性;人工智能辅助的AIGERS为0.94,而人工智能单独的AIGERS为0.70,p
{"title":"Systematic Review: Agentic AI in Neuroradiology: Technical Promise with Limited Clinical Evidence.","authors":"Sara Salehi, Varekan Keishing, Yashbir Singh, David Wei, Amirali Khosravi, Parnian Habibi, Jaidip Jagtap, Bradley J Erickson","doi":"10.1007/s10278-025-01839-2","DOIUrl":"https://doi.org/10.1007/s10278-025-01839-2","url":null,"abstract":"<p><p>Agentic artificial intelligence systems featuring iterative reasoning, autonomous tool use, or multi-agent collaboration have been proposed as solutions to the limitations of large language models (LLMs) in neuroradiology. However, the extent of their implementation and clinical validation remains unclear. We systematically searched PubMed, Web of Science, and Scopus (January 2022-August 2025) for studies implementing agentic AI in neuroradiology. Six independent reviewers (three medical doctors and three AI specialists) assessed full texts. Agentic AI was defined as requiring mandatory iterative reasoning plus either autonomous tool use or multi-agent collaboration. Study quality was evaluated using adapted QUADAS-AI criteria. From 230 records, 9 studies (3.90%) met inclusion criteria. Of these, five (55.60%) implemented true multi-agent architecture, two (22.20%) used hybrid or conceptual frameworks, and two (22.20%) relied on single-model LLMs without genuine agentic behavior. All nine studies were single center with no external validation. Sample sizes were small (median 142 cases; range 16-302). The only randomized controlled trial-INSPIRE (neurophysiology with imaging correlation)-demonstrated high technical performance (≈92% accuracy; AIGERS 0.94 for AI-assisted vs. 0.70 for AI-only, p < 0.001) but showed no measurable clinical benefit when physicians used AI assistance compared with independent reporting. Safety assessments were absent from all studies. Agentic AI in neuroradiology remains technically promising but clinically unproven. Severe evidence scarcity (3.90% inclusion rate), frequent overextension of the \"agentic\" label (30% of studies lacked genuine autonomy), and the persistent gap between technical performance and clinical utility indicate that the field remains in its early research phase. Current evidence is insufficient to support clinical deployment. Rigorous, multi-center prospective trials with patient-centered and safety outcomes are essential before clinical implementation can be responsibly considered.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146109463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure-Aware DeformConv and Hierarchical Multi-scale Transformer for Medical Image Registration. 基于结构感知的医学图像配准DeformConv和分层多尺度变压器。
Pub Date : 2026-02-02 DOI: 10.1007/s10278-025-01814-x
Xuehu Wang, Yanru Qin, Zhihao Pan, Han Qiu, Zhiyuan Zhang, Yongchang Zheng, Lihong Xing, Xiaoping Yin, Shuyang Zhao

Medical image multimodal registration is not only an indispensable processing step in medical image analysis but also plays a crucial role in disease diagnosis and treatment planning. However, the complex and unknown spatial deformation relationships between different organs and different modalities pose significant challenges to multimodal image registration. To address this problem, this study proposes an unsupervised and discriminator-free multimodal registration method based on a dual loss function-SA-HMT. Specifically, to address the challenge of cross-modal feature matching, the multi-scale skip Transformer module proposed in this study employs a hierarchical architecture to capture multi-scale deformation features. In the shallow network, the multi-scale skip pyramid module extracts modality-independent local structural features through parallel multi-branch convolution, effectively overcoming the differential expression of edges and textures across different modalities. In the deep network, the Transformer module establishes long-range dependencies via self-attention mechanism, enabling adaptive fusion of local deformation features with global semantics and effectively alleviating the matching difficulty of cross-modal structural features. In addition, this study further proposes a structure-aware deformable convolution module. The two-stage joint mechanism of "feature perception-offset generation" enhances the accuracy of feature matching through their progressive collaboration. The effectiveness of SA-HMT has been fully verified in five public data sets (covering chest and abdomen CT-MR, lung CT, brain CT-MR, cardiac MRI) and clinical abdominal data. Compared with the advanced method R2Net, our model achieves improvements in core indicators such as DSC, and the registration accuracy is generally comparable or better.

医学图像多模态配准不仅是医学图像分析中必不可少的处理步骤,而且在疾病诊断和治疗计划中起着至关重要的作用。然而,不同器官和不同模态之间复杂且未知的空间变形关系给多模态图像配准带来了重大挑战。为了解决这一问题,本研究提出了一种基于对偶损失函数sa - hmt的无监督无判别器多模态配准方法。具体来说,为了解决跨模态特征匹配的挑战,本研究提出的多尺度跳变模块采用分层结构来捕获多尺度变形特征。在浅层网络中,多尺度跳跃金字塔模块通过并行多分支卷积提取与模态无关的局部结构特征,有效克服了边缘和纹理在不同模态上的差异表达。在深度网络中,Transformer模块通过自关注机制建立远程依赖关系,实现局部变形特征与全局语义的自适应融合,有效缓解跨模态结构特征的匹配困难。此外,本研究进一步提出了一种结构感知的可变形卷积模块。“特征感知-偏移量生成”两阶段联合机制通过它们之间的递进协作来提高特征匹配的准确性。SA-HMT的有效性已在5个公开数据集(包括胸腹CT- mr、肺部CT、脑部CT- mr、心脏MRI)和临床腹部数据中得到充分验证。与先进的R2Net方法相比,我们的模型在DSC等核心指标上取得了改进,配准精度基本相当或更好。
{"title":"Structure-Aware DeformConv and Hierarchical Multi-scale Transformer for Medical Image Registration.","authors":"Xuehu Wang, Yanru Qin, Zhihao Pan, Han Qiu, Zhiyuan Zhang, Yongchang Zheng, Lihong Xing, Xiaoping Yin, Shuyang Zhao","doi":"10.1007/s10278-025-01814-x","DOIUrl":"https://doi.org/10.1007/s10278-025-01814-x","url":null,"abstract":"<p><p>Medical image multimodal registration is not only an indispensable processing step in medical image analysis but also plays a crucial role in disease diagnosis and treatment planning. However, the complex and unknown spatial deformation relationships between different organs and different modalities pose significant challenges to multimodal image registration. To address this problem, this study proposes an unsupervised and discriminator-free multimodal registration method based on a dual loss function-SA-HMT. Specifically, to address the challenge of cross-modal feature matching, the multi-scale skip Transformer module proposed in this study employs a hierarchical architecture to capture multi-scale deformation features. In the shallow network, the multi-scale skip pyramid module extracts modality-independent local structural features through parallel multi-branch convolution, effectively overcoming the differential expression of edges and textures across different modalities. In the deep network, the Transformer module establishes long-range dependencies via self-attention mechanism, enabling adaptive fusion of local deformation features with global semantics and effectively alleviating the matching difficulty of cross-modal structural features. In addition, this study further proposes a structure-aware deformable convolution module. The two-stage joint mechanism of \"feature perception-offset generation\" enhances the accuracy of feature matching through their progressive collaboration. The effectiveness of SA-HMT has been fully verified in five public data sets (covering chest and abdomen CT-MR, lung CT, brain CT-MR, cardiac MRI) and clinical abdominal data. Compared with the advanced method R2Net, our model achieves improvements in core indicators such as DSC, and the registration accuracy is generally comparable or better.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146109495","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Radiomics to Differentiate Renal Oncocytoma from Clear Cell Renal Cell Carcinoma on Contrast-Enhanced CT: A Preliminary Study. 增强CT放射组学鉴别肾癌与透明细胞癌的初步研究。
Pub Date : 2026-02-02 DOI: 10.1007/s10278-026-01851-0
Fang Liu, Longwei Jia, Xiaoming Zhou, Lan Yu

This study assessed the value of radiomics analysis in differentiating clear cell renal cell carcinoma (ccRCC) from renal oncocytoma (RO) using multi-phase contrast-enhanced CT. A retrospective analysis included 43 ccRCC and 43 RO cases (2013-2024). Preoperative three-phase CT scans (corticomedullary [CP], nephrographic [NP], excretory [EP]) were analyzed. Tumor regions of interest (ROIs) were semi-automatically segmented in 3D-Slicer, with texture features extracted via IBEX software. Receiver operating characteristic (ROC) curves and area under the curve (AUC) values were calculated for selected parameters in each phase. A support vector machine (SVM) classifier trained on texture parameters underwent diagnostic evaluation via ROC analysis. All phases showed high diagnostic accuracy (AUC > 0.9), with NP demonstrating the highest performance (AUC = 0.952; accuracy, 0.88; sensitivity, 0.91; specificity, 0.87). Intensity histogram IH_Skewness differed significantly between ccRCC and RO in CP and NP (P < 0.01 for both), with AUC values of 0.75 (CP) and 0.79 (NP). Combining LASSO dimension reduction with SVM using multi-phase CT radiomics features enabled the effective differentiation between ccRCC and RO, highlighting texture analysis as a promising clinical tool.

本研究评估了放射组学分析在多期增强CT鉴别透明细胞肾细胞癌(ccRCC)和肾癌细胞瘤(RO)中的价值。回顾性分析43例ccRCC和43例RO(2013-2024)。分析术前三相CT扫描(皮质髓质[CP]、肾造影[NP]、排泄[EP])。在3D-Slicer中对肿瘤感兴趣区域(roi)进行半自动分割,并通过IBEX软件提取纹理特征。计算各阶段选定参数的受试者工作特征(ROC)曲线和曲线下面积(AUC)值。基于纹理参数训练的支持向量机分类器通过ROC分析进行诊断性评价。各阶段的诊断准确率均较高(AUC为0.9),其中NP的诊断准确率最高(AUC = 0.952,准确率0.88,敏感性0.91,特异性0.87)。ccRCC和RO在CP和NP的强度直方图IH_Skewness差异有统计学意义(P
{"title":"Radiomics to Differentiate Renal Oncocytoma from Clear Cell Renal Cell Carcinoma on Contrast-Enhanced CT: A Preliminary Study.","authors":"Fang Liu, Longwei Jia, Xiaoming Zhou, Lan Yu","doi":"10.1007/s10278-026-01851-0","DOIUrl":"https://doi.org/10.1007/s10278-026-01851-0","url":null,"abstract":"<p><p>This study assessed the value of radiomics analysis in differentiating clear cell renal cell carcinoma (ccRCC) from renal oncocytoma (RO) using multi-phase contrast-enhanced CT. A retrospective analysis included 43 ccRCC and 43 RO cases (2013-2024). Preoperative three-phase CT scans (corticomedullary [CP], nephrographic [NP], excretory [EP]) were analyzed. Tumor regions of interest (ROIs) were semi-automatically segmented in 3D-Slicer, with texture features extracted via IBEX software. Receiver operating characteristic (ROC) curves and area under the curve (AUC) values were calculated for selected parameters in each phase. A support vector machine (SVM) classifier trained on texture parameters underwent diagnostic evaluation via ROC analysis. All phases showed high diagnostic accuracy (AUC > 0.9), with NP demonstrating the highest performance (AUC = 0.952; accuracy, 0.88; sensitivity, 0.91; specificity, 0.87). Intensity histogram IH_Skewness differed significantly between ccRCC and RO in CP and NP (P < 0.01 for both), with AUC values of 0.75 (CP) and 0.79 (NP). Combining LASSO dimension reduction with SVM using multi-phase CT radiomics features enabled the effective differentiation between ccRCC and RO, highlighting texture analysis as a promising clinical tool.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146109443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A New Insight into Imaging Diagnosis of Otosclerosis Enhanced by Machine Learning and Radiomics. 基于机器学习和放射组学的耳硬化影像学诊断新思路
Pub Date : 2026-01-30 DOI: 10.1007/s10278-026-01843-0
Marta Álvarez de Linera-Alperi, Juan Miranda Bautista, David Corral Fontecha, Josué Fernández Carnero, José A Vega, Pablo Menéndez Fernández-Miranda

Otosclerosis is a disease affecting the middle and inner ear, characterized by abnormal bone remodeling that leads to stapes fixation and progressive hearing loss. Although high-resolution computed tomography (HRCT) is the standard imaging modality for diagnosis, its sensitivity is limited, with a high false-negative rate (FNR). This study investigates the use of radiomics and machine learning (ML) to improve diagnostic accuracy. HRCT scans from 99 subjects (48 otosclerosis, 51 controls) were analyzed, focusing on the stapes, antefenestral region (AF), and oval window (OW). From each scan, 6048 radiomic features were extracted and reduced to 1317 through feature selection. Statistical analyses and ML modeling were performed using the selected features. Sixty-seven biomarkers showed significant differences between cases and controls, primarily in the AF (56) and stapes (11); none were found in the OW. Both the AF and stapes exhibited increased heterogeneity in otosclerosis, reflecting the bone remodeling process. A reduction in the stapes' major axis was also observed, possibly related to torsional deformation. Image transformation filters enhanced disease visibility. Among several ML classifiers tested, L2-regularized logistic regression performed best, achieving an AUC of 0.90 ± 0.06, thereby enhancing the diagnostic accuracy reported in some studies for radiologists. Hierarchical clustering of the most predictive features further confirmed their strong discriminative power. Our findings highlight the potential of radiomics and ML to standardize otosclerosis diagnosis, reduce FNR, and support surgical decision-making. Future studies should validate these results using larger cohorts and advanced imaging technologies such as Photon-Counting CT.

耳硬化是一种影响中耳和内耳的疾病,其特征是骨重塑异常,导致镫骨固定和进行性听力丧失。虽然高分辨率计算机断层扫描(HRCT)是诊断的标准成像方式,但其灵敏度有限,假阴性率(FNR)高。本研究探讨使用放射组学和机器学习(ML)来提高诊断准确性。对99名受试者(48名耳硬化,51名对照)的HRCT扫描结果进行分析,重点关注镫骨、前门骨区(AF)和卵圆窗(OW)。从每次扫描中提取6048个放射学特征,通过特征选择将其减少到1317个。使用选择的特征进行统计分析和ML建模。67项生物标志物在病例和对照组之间显示出显著差异,主要是房颤(56)和镫骨(11);在OW中没有发现。房颤和镫骨在耳硬化中均表现出增加的异质性,反映了骨重塑过程。镫骨长轴复位也被观察到,可能与扭转变形有关。图像变换过滤器增强了疾病的可见性。在测试的几个ML分类器中,l2正则化逻辑回归表现最好,达到了0.90±0.06的AUC,从而提高了一些研究中报道的放射科医生的诊断准确性。对最具预测性的特征进行层次聚类,进一步证实了其较强的判别能力。我们的研究结果强调了放射组学和ML在标准化耳硬化诊断、减少FNR和支持手术决策方面的潜力。未来的研究应该使用更大的队列和先进的成像技术(如光子计数CT)来验证这些结果。
{"title":"A New Insight into Imaging Diagnosis of Otosclerosis Enhanced by Machine Learning and Radiomics.","authors":"Marta Álvarez de Linera-Alperi, Juan Miranda Bautista, David Corral Fontecha, Josué Fernández Carnero, José A Vega, Pablo Menéndez Fernández-Miranda","doi":"10.1007/s10278-026-01843-0","DOIUrl":"https://doi.org/10.1007/s10278-026-01843-0","url":null,"abstract":"<p><p>Otosclerosis is a disease affecting the middle and inner ear, characterized by abnormal bone remodeling that leads to stapes fixation and progressive hearing loss. Although high-resolution computed tomography (HRCT) is the standard imaging modality for diagnosis, its sensitivity is limited, with a high false-negative rate (FNR). This study investigates the use of radiomics and machine learning (ML) to improve diagnostic accuracy. HRCT scans from 99 subjects (48 otosclerosis, 51 controls) were analyzed, focusing on the stapes, antefenestral region (AF), and oval window (OW). From each scan, 6048 radiomic features were extracted and reduced to 1317 through feature selection. Statistical analyses and ML modeling were performed using the selected features. Sixty-seven biomarkers showed significant differences between cases and controls, primarily in the AF (56) and stapes (11); none were found in the OW. Both the AF and stapes exhibited increased heterogeneity in otosclerosis, reflecting the bone remodeling process. A reduction in the stapes' major axis was also observed, possibly related to torsional deformation. Image transformation filters enhanced disease visibility. Among several ML classifiers tested, L2-regularized logistic regression performed best, achieving an AUC of 0.90 ± 0.06, thereby enhancing the diagnostic accuracy reported in some studies for radiologists. Hierarchical clustering of the most predictive features further confirmed their strong discriminative power. Our findings highlight the potential of radiomics and ML to standardize otosclerosis diagnosis, reduce FNR, and support surgical decision-making. Future studies should validate these results using larger cohorts and advanced imaging technologies such as Photon-Counting CT.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146088708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Handcrafted vs. Deep Radiomics vs. Fusion vs. Deep Learning: A Comprehensive Review of Machine Learning -Based Cancer Outcome Prediction in PET and SPECT Imaging. 手工制作vs.深度放射组学vs.融合vs.深度学习:PET和SPECT成像中基于机器学习的癌症预后预测的综合综述。
Pub Date : 2026-01-29 DOI: 10.1007/s10278-025-01804-z
Mohammad R Salmanpour, Somayeh Sadat Mehrnia, Sajad Jabarzadeh Ghandilu, Zhino Safahi, Sonya Falahati, Shahram Taeb, Ghazal Mousavi, Mehdi Maghsudi, Ahmad Shariftabrizi, Ilker Hacihaliloglu, Arman Rahmim

Machine learning (ML), particularly deep learning (DL) and radiomics-based approaches, has emerged as a powerful tool for cancer outcome prediction using PET and SPECT imaging. However, the comparative performance of different techniques-handcrafted radiomics features (HRF), deep radiomics features (DRF), DL models, and hybrid fusion models (combinations of DRF, HRF, and clinical features)-remains inconsistent across clinical applications. This systematic review analyzed 226 studies published between 2020 and 2025 that applied ML to PET or SPECT imaging for cancer outcome prediction tasks. Each study was evaluated using a 59-item framework addressing dataset construction, feature extraction methods, validation strategies, interpretability, and risk of bias. We extracted key data, including model type, cancer site, imaging modality, and performance metrics such as accuracy and area under the curve (AUC). PET-based models (95%) generally outperformed SPECT, likely due to superior spatial resolution and sensitivity. DRF models achieved the highest mean accuracy (0.862 ± 0.051), while fusion models attained the highest AUC (0.861 ± 0.088). ANOVA revealed significant differences in accuracy (p = 0.0006) and AUC (p = 0.0027). Despite these promising findings, key limitations remain, including poor management of class imbalance (59%), missing data (29%), and low population diversity (19%). Only 48% adhered to IBSI (Image Biomarker Standardization Initiative) standards. This systematic review shows that DL and DRF-based models, especially in fusion with HRFs, outperform HRF-only methods for cancer outcome prediction using PET/SPECT, particularly in data-limited settings. Despite strong performance, challenges remain in interpretability and standardization, highlighting the need for unified DRF extraction frameworks across modalities.

机器学习(ML),特别是深度学习(DL)和基于放射学的方法,已经成为使用PET和SPECT成像预测癌症结果的有力工具。然而,不同技术的比较性能——手工制作的放射组学特征(HRF)、深度放射组学特征(DRF)、DL模型和混合融合模型(DRF、HRF和临床特征的组合)——在临床应用中仍然不一致。本系统综述分析了2020年至2025年间发表的226项研究,这些研究将ML应用于PET或SPECT成像用于癌症预后预测任务。每个研究使用59项框架进行评估,包括数据集构建、特征提取方法、验证策略、可解释性和偏倚风险。我们提取了关键数据,包括模型类型、癌症部位、成像方式和性能指标,如准确性和曲线下面积(AUC)。基于pet的模型(95%)通常优于SPECT,可能是由于优越的空间分辨率和灵敏度。DRF模型的平均精度最高(0.862±0.051),而融合模型的AUC最高(0.861±0.088)。方差分析显示准确率(p = 0.0006)和AUC (p = 0.0027)有显著差异。尽管有这些有希望的发现,但主要的局限性仍然存在,包括阶级不平衡管理不善(59%),数据缺失(29%)和人口多样性低(19%)。只有48%的人遵守了IBSI(图像生物标志物标准化倡议)标准。该系统综述显示,特别是在数据有限的情况下,基于DL和drf的模型,特别是与hrf融合的模型,在使用PET/SPECT进行癌症预后预测方面优于仅hrf的方法。尽管表现强劲,但在可解释性和标准化方面仍然存在挑战,突出了跨模式统一DRF提取框架的必要性。
{"title":"Handcrafted vs. Deep Radiomics vs. Fusion vs. Deep Learning: A Comprehensive Review of Machine Learning -Based Cancer Outcome Prediction in PET and SPECT Imaging.","authors":"Mohammad R Salmanpour, Somayeh Sadat Mehrnia, Sajad Jabarzadeh Ghandilu, Zhino Safahi, Sonya Falahati, Shahram Taeb, Ghazal Mousavi, Mehdi Maghsudi, Ahmad Shariftabrizi, Ilker Hacihaliloglu, Arman Rahmim","doi":"10.1007/s10278-025-01804-z","DOIUrl":"https://doi.org/10.1007/s10278-025-01804-z","url":null,"abstract":"<p><p>Machine learning (ML), particularly deep learning (DL) and radiomics-based approaches, has emerged as a powerful tool for cancer outcome prediction using PET and SPECT imaging. However, the comparative performance of different techniques-handcrafted radiomics features (HRF), deep radiomics features (DRF), DL models, and hybrid fusion models (combinations of DRF, HRF, and clinical features)-remains inconsistent across clinical applications. This systematic review analyzed 226 studies published between 2020 and 2025 that applied ML to PET or SPECT imaging for cancer outcome prediction tasks. Each study was evaluated using a 59-item framework addressing dataset construction, feature extraction methods, validation strategies, interpretability, and risk of bias. We extracted key data, including model type, cancer site, imaging modality, and performance metrics such as accuracy and area under the curve (AUC). PET-based models (95%) generally outperformed SPECT, likely due to superior spatial resolution and sensitivity. DRF models achieved the highest mean accuracy (0.862 ± 0.051), while fusion models attained the highest AUC (0.861 ± 0.088). ANOVA revealed significant differences in accuracy (p = 0.0006) and AUC (p = 0.0027). Despite these promising findings, key limitations remain, including poor management of class imbalance (59%), missing data (29%), and low population diversity (19%). Only 48% adhered to IBSI (Image Biomarker Standardization Initiative) standards. This systematic review shows that DL and DRF-based models, especially in fusion with HRFs, outperform HRF-only methods for cancer outcome prediction using PET/SPECT, particularly in data-limited settings. Despite strong performance, challenges remain in interpretability and standardization, highlighting the need for unified DRF extraction frameworks across modalities.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146088779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmark White Matter Hyperintensity Segmentation Methods Fail on Heterogeneous Clinical MRI: A New Dataset and Deep Learning-Based Solutions. 基准白质高强度分割方法在异质临床MRI上失败:一个新的数据集和基于深度学习的解决方案。
Pub Date : 2026-01-29 DOI: 10.1007/s10278-025-01808-9
Junjie Wu, Joshua D Brown, Ranliang Hu, Paula J Edwards, Allan I Levey, James J Lah, Deqiang Qiu

Existing automated methods for white matter hyperintensity (WMH) segmentation often generalize poorly to heterogeneous clinical MRI due to variability in scanner types, field strengths, and protocols. To address this challenge, we introduce a diverse clinical WMH dataset and evaluate two deep learning-based solutions: an nnU-Net model trained directly on the data and a foundation model adapted through fine-tuning. This retrospective study included 195 routine brain MRI scans acquired from 71 scanners between June 2006 and October 2022. Participants ranged in age from 46 to 87 years (median, 70 years; 94 females). WMHs were manually annotated by an experienced rater and reviewed under neuroradiologist supervision. Several benchmark segmentation methods were evaluated against these annotations. We then developed Robust-WMH-UNet by training nnU-Net on the dataset and Robust-WMH-SAM by fine-tuning MedSAM, a vision foundation model. Benchmark methods demonstrated poor generalization, frequently missing small lesions and producing false positives in anatomically complex regions such as the septum pellucidum. Robust-WMH-UNet achieved superior accuracy (median Dice similarity coefficient [DSC], 0.768) with improved specificity, while Robust-WMH-SAM attained competitive performance (median DSC up to 0.750) after only limited training, reaching acceptable accuracy within a single epoch. This new clinically representative dataset provides a strong foundation for developing robust WMH algorithms, enabling fair cross-method comparisons, and supporting the translation of segmentation models into routine clinical practice.

由于扫描仪类型、场强和方案的差异,现有的白质高强度(WMH)分割自动化方法通常不适用于异质性临床MRI。为了应对这一挑战,我们引入了一个多样化的临床WMH数据集,并评估了两种基于深度学习的解决方案:直接在数据上训练的nnU-Net模型和通过微调调整的基础模型。这项回顾性研究包括2006年6月至2022年10月期间从71台扫描仪获得的195次常规脑MRI扫描。参与者的年龄从46岁到87岁不等(中位数为70岁;94名女性)。wmh由经验丰富的评分员手工注释,并在神经放射学家的监督下进行审查。针对这些注释评估了几种基准分割方法。然后,我们通过在数据集上训练nnU-Net开发了Robust-WMH-UNet,并通过微调MedSAM(一个视觉基础模型)开发了Robust-WMH-SAM。基准方法的通用性较差,经常遗漏小病变,并在解剖结构复杂的区域(如透明隔)产生假阳性。鲁棒- wmh - unet获得了更高的准确度(中位数骰子相似系数[DSC], 0.768),特异性提高,而鲁棒- wmh - sam仅经过有限的训练就获得了竞争性能(中位数DSC高达0.750),在单个历元内达到可接受的准确度。这个新的临床代表性数据集为开发稳健的WMH算法提供了坚实的基础,实现了公平的跨方法比较,并支持将分割模型转化为常规临床实践。
{"title":"Benchmark White Matter Hyperintensity Segmentation Methods Fail on Heterogeneous Clinical MRI: A New Dataset and Deep Learning-Based Solutions.","authors":"Junjie Wu, Joshua D Brown, Ranliang Hu, Paula J Edwards, Allan I Levey, James J Lah, Deqiang Qiu","doi":"10.1007/s10278-025-01808-9","DOIUrl":"https://doi.org/10.1007/s10278-025-01808-9","url":null,"abstract":"<p><p>Existing automated methods for white matter hyperintensity (WMH) segmentation often generalize poorly to heterogeneous clinical MRI due to variability in scanner types, field strengths, and protocols. To address this challenge, we introduce a diverse clinical WMH dataset and evaluate two deep learning-based solutions: an nnU-Net model trained directly on the data and a foundation model adapted through fine-tuning. This retrospective study included 195 routine brain MRI scans acquired from 71 scanners between June 2006 and October 2022. Participants ranged in age from 46 to 87 years (median, 70 years; 94 females). WMHs were manually annotated by an experienced rater and reviewed under neuroradiologist supervision. Several benchmark segmentation methods were evaluated against these annotations. We then developed Robust-WMH-UNet by training nnU-Net on the dataset and Robust-WMH-SAM by fine-tuning MedSAM, a vision foundation model. Benchmark methods demonstrated poor generalization, frequently missing small lesions and producing false positives in anatomically complex regions such as the septum pellucidum. Robust-WMH-UNet achieved superior accuracy (median Dice similarity coefficient [DSC], 0.768) with improved specificity, while Robust-WMH-SAM attained competitive performance (median DSC up to 0.750) after only limited training, reaching acceptable accuracy within a single epoch. This new clinically representative dataset provides a strong foundation for developing robust WMH algorithms, enabling fair cross-method comparisons, and supporting the translation of segmentation models into routine clinical practice.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146088746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Edge-Aware Dual-Branch CNN Architecture for Alzheimer's Disease Diagnosis. 边缘感知双分支CNN架构在阿尔茨海默病诊断中的应用。
Pub Date : 2026-01-27 DOI: 10.1007/s10278-025-01836-5
Man Li, Mei Choo Ang, Musatafa Abbas Abbood Albadr, Jun Kit Chaw, JianBang Liu, Kok Weng Ng, Wei Hong

The rapid development of machine learning (ML) and deep learning (DL) has greatly advanced Alzheimer's disease (AD) diagnosis. However, existing models struggle to capture weak structural features in the marginal regions of brain MRI images, leading to limited diagnostic accuracy. To address this challenge, we introduce a Dual-Branch Convolutional Neural Network (DBCNN) equipped with a Learnable Edge Detection Module designed to jointly learn global semantic representations and fine-grained edge cues within a unified framework. Experimental results on two public datasets demonstrate that DBCNN significantly improves classification accuracy, surpassing 98%. Notably, on the OASIS dataset, it achieved an average accuracy of 99.71%, demonstrating strong generalization and robustness. This high diagnostic performance indicates that the model can assist clinicians in the early detection of Alzheimer's disease, reduce subjectivity in manual image interpretation, and enhance diagnostic consistency. Overall, the proposed approach provides a promising pathway toward intelligent, interpretable, and computationally efficient solutions for MRI-based diagnosis, offering strong potential to support early clinical decision-making.

机器学习(ML)和深度学习(DL)的快速发展极大地推动了阿尔茨海默病(AD)的诊断。然而,现有的模型难以捕捉大脑MRI图像边缘区域的弱结构特征,导致诊断准确性有限。为了解决这一挑战,我们引入了一个双分支卷积神经网络(DBCNN),该网络配备了一个可学习的边缘检测模块,旨在在统一的框架内联合学习全局语义表示和细粒度边缘线索。在两个公开数据集上的实验结果表明,DBCNN显著提高了分类准确率,达到98%以上。值得注意的是,在OASIS数据集上,平均准确率达到99.71%,显示出较强的泛化和鲁棒性。这种高诊断性能表明,该模型可以帮助临床医生早期发现阿尔茨海默病,减少人工图像解释的主观性,提高诊断的一致性。总的来说,该方法为基于mri的诊断提供了智能、可解释和计算效率高的解决方案,为支持早期临床决策提供了强大的潜力。
{"title":"Edge-Aware Dual-Branch CNN Architecture for Alzheimer's Disease Diagnosis.","authors":"Man Li, Mei Choo Ang, Musatafa Abbas Abbood Albadr, Jun Kit Chaw, JianBang Liu, Kok Weng Ng, Wei Hong","doi":"10.1007/s10278-025-01836-5","DOIUrl":"https://doi.org/10.1007/s10278-025-01836-5","url":null,"abstract":"<p><p>The rapid development of machine learning (ML) and deep learning (DL) has greatly advanced Alzheimer's disease (AD) diagnosis. However, existing models struggle to capture weak structural features in the marginal regions of brain MRI images, leading to limited diagnostic accuracy. To address this challenge, we introduce a Dual-Branch Convolutional Neural Network (DBCNN) equipped with a Learnable Edge Detection Module designed to jointly learn global semantic representations and fine-grained edge cues within a unified framework. Experimental results on two public datasets demonstrate that DBCNN significantly improves classification accuracy, surpassing 98%. Notably, on the OASIS dataset, it achieved an average accuracy of 99.71%, demonstrating strong generalization and robustness. This high diagnostic performance indicates that the model can assist clinicians in the early detection of Alzheimer's disease, reduce subjectivity in manual image interpretation, and enhance diagnostic consistency. Overall, the proposed approach provides a promising pathway toward intelligent, interpretable, and computationally efficient solutions for MRI-based diagnosis, offering strong potential to support early clinical decision-making.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146055759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of imaging informatics in medicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1