首页 > 最新文献

Journal of imaging informatics in medicine最新文献

英文 中文
Prediction of Future Risk of Moderate to Severe Kidney Function Loss Using a Deep Learning Model-Enabled Chest Radiography. 使用支持深度学习模型的胸部x线摄影预测中度至重度肾功能丧失的未来风险。
Pub Date : 2026-02-01 Epub Date: 2025-04-02 DOI: 10.1007/s10278-025-01489-4
Kai-Chieh Chen, Shang-Yang Lee, Dung-Jang Tsai, Kai-Hsiung Ko, Yi-Chih Hsu, Wei-Chou Chang, Wen-Hui Fang, Chin Lin, Yu-Juei Hsu

Chronic kidney disease (CKD) remains a major public health concern, requiring better predictive models for early intervention. This study evaluates a deep learning model (DLM) that utilizes raw chest X-ray (CXR) data to predict moderate to severe kidney function decline. We analyzed data from 79,219 patients with an estimated Glomerular Filtration Rate (eGFR) between 65 and 120, segmented into development (n = 37,983), tuning (n = 15,346), internal validation (n = 14,113), and external validation (n = 11,777) sets. Our DLM, pretrained on CXR-report pairs, was fine-tuned with the development set. We retrospectively examined data spanning April 2011 to February 2022, with a 5-year maximum follow-up. Primary and secondary endpoints included CKD stage 3b progression, ESRD/dialysis, and mortality. The overall concordance index (C-index) values for the internal and external validation sets were 0.903 (95% CI, 0.885-0.922) and 0.851 (95% CI, 0.819-0.883), respectively. In these sets, the incidences of progression to CKD stage 3b at 5 years were 19.2% and 13.4% in the high-risk group, significantly higher than those in the median-risk (5.9% and 5.1%) and low-risk groups (0.9% and 0.9%), respectively. The sex, age, and eGFR-adjusted hazard ratios (HR) for the high-risk group compared to the low-risk group were 16.88 (95% CI, 10.84-26.28) and 7.77 (95% CI, 4.77-12.64), respectively. The high-risk group also exhibited higher probabilities of progressing to ESRD/dialysis or experiencing mortality compared to the low-risk group. Further analysis revealed that the high-risk group compared to the low/median-risk group had a higher prevalence of complications and abnormal blood/urine markers. Our findings demonstrate that a DLM utilizing CXR can effectively predict CKD stage 3b progression, offering a potential tool for early intervention in high-risk populations.

慢性肾脏疾病(CKD)仍然是一个主要的公共卫生问题,需要更好的早期干预预测模型。本研究评估了一种深度学习模型(DLM),该模型利用原始胸部x射线(CXR)数据预测中度至重度肾功能下降。我们分析了来自79,219例患者的数据,这些患者的肾小球滤过率(eGFR)估计在65 - 120之间,分为发展组(n = 37,983)、调整组(n = 15,346)、内部验证组(n = 14,113)和外部验证组(n = 11,777)。我们的DLM在cxr -报告对上进行了预训练,并根据开发集进行了微调。我们回顾性研究了2011年4月至2022年2月的数据,最长随访时间为5年。主要和次要终点包括CKD 3b期进展、ESRD/透析和死亡率。内部和外部验证集的总体一致性指数(C-index)值分别为0.903 (95% CI, 0.885-0.922)和0.851 (95% CI, 0.819-0.883)。在这些组中,高风险组5年进展为CKD 3b期的发生率分别为19.2%和13.4%,显著高于中危组(5.9%和5.1%)和低危组(0.9%和0.9%)。与低危组相比,高危组的性别、年龄和egfr调整后的危险比(HR)分别为16.88 (95% CI, 10.84-26.28)和7.77 (95% CI, 4.77-12.64)。与低风险组相比,高风险组也表现出更高的进展为ESRD/透析或经历死亡的可能性。进一步分析显示,与低/中危组相比,高危组并发症和异常血液/尿液标志物的患病率更高。我们的研究结果表明,利用CXR的DLM可以有效地预测CKD 3b期进展,为高危人群的早期干预提供了潜在的工具。
{"title":"Prediction of Future Risk of Moderate to Severe Kidney Function Loss Using a Deep Learning Model-Enabled Chest Radiography.","authors":"Kai-Chieh Chen, Shang-Yang Lee, Dung-Jang Tsai, Kai-Hsiung Ko, Yi-Chih Hsu, Wei-Chou Chang, Wen-Hui Fang, Chin Lin, Yu-Juei Hsu","doi":"10.1007/s10278-025-01489-4","DOIUrl":"10.1007/s10278-025-01489-4","url":null,"abstract":"<p><p>Chronic kidney disease (CKD) remains a major public health concern, requiring better predictive models for early intervention. This study evaluates a deep learning model (DLM) that utilizes raw chest X-ray (CXR) data to predict moderate to severe kidney function decline. We analyzed data from 79,219 patients with an estimated Glomerular Filtration Rate (eGFR) between 65 and 120, segmented into development (n = 37,983), tuning (n = 15,346), internal validation (n = 14,113), and external validation (n = 11,777) sets. Our DLM, pretrained on CXR-report pairs, was fine-tuned with the development set. We retrospectively examined data spanning April 2011 to February 2022, with a 5-year maximum follow-up. Primary and secondary endpoints included CKD stage 3b progression, ESRD/dialysis, and mortality. The overall concordance index (C-index) values for the internal and external validation sets were 0.903 (95% CI, 0.885-0.922) and 0.851 (95% CI, 0.819-0.883), respectively. In these sets, the incidences of progression to CKD stage 3b at 5 years were 19.2% and 13.4% in the high-risk group, significantly higher than those in the median-risk (5.9% and 5.1%) and low-risk groups (0.9% and 0.9%), respectively. The sex, age, and eGFR-adjusted hazard ratios (HR) for the high-risk group compared to the low-risk group were 16.88 (95% CI, 10.84-26.28) and 7.77 (95% CI, 4.77-12.64), respectively. The high-risk group also exhibited higher probabilities of progressing to ESRD/dialysis or experiencing mortality compared to the low-risk group. Further analysis revealed that the high-risk group compared to the low/median-risk group had a higher prevalence of complications and abnormal blood/urine markers. Our findings demonstrate that a DLM utilizing CXR can effectively predict CKD stage 3b progression, offering a potential tool for early intervention in high-risk populations.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"454-467"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920974/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143775244","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Burn Diagnosis through SE-ResNet18 and Confidence Filtering. 利用SE-ResNet18和置信度滤波增强烧伤诊断。
Pub Date : 2026-02-01 Epub Date: 2025-04-08 DOI: 10.1007/s10278-025-01495-6
Hanyue Mo, Ziwen Kuang, Haoxuan Wang, Xinyi Cai, Kun Cheng

Accurate classification of burn severity is crucial for effective clinical treatment; however, existing methods often fail to balance precision and real-time performance. To address this challenge, we propose a deep learning-based approach utilizing an enhanced ResNet18 architecture with integrated attention mechanisms to improve classification accuracy. The system consists of data preprocessing, classification, optimization, and post-processing modules. The optimization strategy employs an adaptive learning rate combining cosine annealing and class-specific gradient adaptation, alongside targeted adjustments for class imbalance, while an improved Adam optimizer enhances convergence stability. Post-processing incorporates confidence filtering (threshold 0.3) and selective evaluation, with weighted aggregation-integrating dynamic accuracy calculation and moving average to refine predictions and enhance diagnostic reliability. Experimental results on a burn skin test dataset demonstrate that the proposed model achieves a classification accuracy of 99.19% ± 0.12 and a mean average precision (mAP) of 98.72% ± 0.10, highlighting its potential for real-time clinical burn assessment.

烧伤严重程度的准确分类对于有效的临床治疗至关重要;然而,现有的方法往往无法平衡精度和实时性。为了解决这一挑战,我们提出了一种基于深度学习的方法,利用增强的ResNet18架构和集成的注意力机制来提高分类精度。该系统由数据预处理、分类、优化和后处理四个模块组成。优化策略采用结合余弦退火和类别梯度自适应的自适应学习率,并对类别不平衡进行有针对性的调整,同时改进的Adam优化器增强了收敛稳定性。后处理采用置信度过滤(阈值0.3)和选择性评估,加权聚合集成动态精度计算和移动平均,以改进预测和提高诊断可靠性。在烧伤皮肤测试数据集上的实验结果表明,该模型的分类准确率为99.19%±0.12,平均精度(mAP)为98.72%±0.10,显示了其在临床烧伤实时评估中的潜力。
{"title":"Enhancing Burn Diagnosis through SE-ResNet18 and Confidence Filtering.","authors":"Hanyue Mo, Ziwen Kuang, Haoxuan Wang, Xinyi Cai, Kun Cheng","doi":"10.1007/s10278-025-01495-6","DOIUrl":"10.1007/s10278-025-01495-6","url":null,"abstract":"<p><p>Accurate classification of burn severity is crucial for effective clinical treatment; however, existing methods often fail to balance precision and real-time performance. To address this challenge, we propose a deep learning-based approach utilizing an enhanced ResNet18 architecture with integrated attention mechanisms to improve classification accuracy. The system consists of data preprocessing, classification, optimization, and post-processing modules. The optimization strategy employs an adaptive learning rate combining cosine annealing and class-specific gradient adaptation, alongside targeted adjustments for class imbalance, while an improved Adam optimizer enhances convergence stability. Post-processing incorporates confidence filtering (threshold 0.3) and selective evaluation, with weighted aggregation-integrating dynamic accuracy calculation and moving average to refine predictions and enhance diagnostic reliability. Experimental results on a burn skin test dataset demonstrate that the proposed model achieves a classification accuracy of 99.19% ± 0.12 and a mean average precision (mAP) of 98.72% ± 0.10, highlighting its potential for real-time clinical burn assessment.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"639-654"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920881/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143813386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Dirichlet Distribution-Based Complex Ensemble Approach for Breast Cancer Classification from Ultrasound Images with Transfer Learning and Multiphase Spaced Repetition Method. 基于Dirichlet分布的复杂集成方法在乳腺癌超声图像分类中的迁移学习和多相间隔重复方法。
Pub Date : 2026-02-01 Epub Date: 2025-04-29 DOI: 10.1007/s10278-025-01515-5
Osman Güler

Breast ultrasound is a useful and rapid diagnostic tool for the early detection of breast cancer. Artificial intelligence-supported computer-aided decision systems, which assist expert radiologists and clinicians, provide reliable and rapid results. Deep learning methods and techniques are widely used in the field of health for early diagnosis, abnormality detection, and disease diagnosis. Therefore, in this study, a deep ensemble learning model based on Dirichlet distribution using pre-trained transfer learning models for breast cancer classification from ultrasound images is proposed. In the study, experiments were conducted using the Breast Ultrasound Images Dataset (BUSI). The dataset, which had an imbalanced class structure, was balanced using data augmentation techniques. DenseNet201, InceptionV3, VGG16, and ResNet152 models were used for transfer learning with fivefold cross-validation. Statistical analyses, including the ANOVA test and Tukey HSD test, were applied to evaluate the model's performance and ensure the reliability of the results. Additionally, Grad-CAM (Gradient-weighted Class Activation Mapping) was used for explainable AI (XAI), providing visual explanations of the deep learning model's decision-making process. The spaced repetition method, commonly used to improve the success of learners in educational sciences, was adapted to artificial intelligence in this study. The results of training with transfer learning models were used as input for further training, and spaced repetition was applied using previously learned information. The use of the spaced repetition method led to increased model success and reduced learning times. The weights obtained from the trained models were input into an ensemble learning system based on Dirichlet distribution with different variations. The proposed model achieved 99.60% validation accuracy on the dataset, demonstrating its effectiveness in breast cancer classification.

乳腺超声是早期发现乳腺癌的一种有用的快速诊断工具。人工智能支持的计算机辅助决策系统,可以帮助放射科专家和临床医生,提供可靠和快速的结果。深度学习方法和技术被广泛应用于健康领域的早期诊断、异常检测和疾病诊断。因此,本研究提出了一种基于Dirichlet分布的深度集成学习模型,利用预训练迁移学习模型对超声图像进行乳腺癌分类。在本研究中,实验使用乳腺超声图像数据集(BUSI)进行。类结构不平衡的数据集使用数据增强技术进行了平衡。使用DenseNet201、InceptionV3、VGG16和ResNet152模型进行迁移学习,并进行五重交叉验证。采用统计分析,包括ANOVA检验和Tukey HSD检验来评价模型的性能,确保结果的可靠性。此外,Grad-CAM(梯度加权类激活映射)用于可解释人工智能(XAI),为深度学习模型的决策过程提供可视化解释。通常用于提高教育科学学习者成功的间隔重复方法在本研究中被应用于人工智能。使用迁移学习模型的训练结果作为进一步训练的输入,并使用先前学习的信息应用间隔重复。使用间隔重复法提高了模型的成功率,减少了学习时间。将训练模型得到的权值输入到基于Dirichlet分布的集成学习系统中。该模型在数据集上的验证准确率达到99.60%,证明了其在乳腺癌分类中的有效性。
{"title":"A Dirichlet Distribution-Based Complex Ensemble Approach for Breast Cancer Classification from Ultrasound Images with Transfer Learning and Multiphase Spaced Repetition Method.","authors":"Osman Güler","doi":"10.1007/s10278-025-01515-5","DOIUrl":"10.1007/s10278-025-01515-5","url":null,"abstract":"<p><p>Breast ultrasound is a useful and rapid diagnostic tool for the early detection of breast cancer. Artificial intelligence-supported computer-aided decision systems, which assist expert radiologists and clinicians, provide reliable and rapid results. Deep learning methods and techniques are widely used in the field of health for early diagnosis, abnormality detection, and disease diagnosis. Therefore, in this study, a deep ensemble learning model based on Dirichlet distribution using pre-trained transfer learning models for breast cancer classification from ultrasound images is proposed. In the study, experiments were conducted using the Breast Ultrasound Images Dataset (BUSI). The dataset, which had an imbalanced class structure, was balanced using data augmentation techniques. DenseNet201, InceptionV3, VGG16, and ResNet152 models were used for transfer learning with fivefold cross-validation. Statistical analyses, including the ANOVA test and Tukey HSD test, were applied to evaluate the model's performance and ensure the reliability of the results. Additionally, Grad-CAM (Gradient-weighted Class Activation Mapping) was used for explainable AI (XAI), providing visual explanations of the deep learning model's decision-making process. The spaced repetition method, commonly used to improve the success of learners in educational sciences, was adapted to artificial intelligence in this study. The results of training with transfer learning models were used as input for further training, and spaced repetition was applied using previously learned information. The use of the spaced repetition method led to increased model success and reduced learning times. The weights obtained from the trained models were input into an ensemble learning system based on Dirichlet distribution with different variations. The proposed model achieved 99.60% validation accuracy on the dataset, demonstrating its effectiveness in breast cancer classification.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"202-228"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920884/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144057056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction: Deep Learning-Based Estimation of Radiographic Position to Automatically Set Up the X-Ray Prime Factors. 修正:基于深度学习的x射线位置估计,自动设置x射线质因数。
Pub Date : 2026-02-01 DOI: 10.1007/s10278-025-01476-9
C F Del Cerro, R C Gimenez, J Garcia-Blas, K Sosenko, J M Ortega, M Desco, M Abella
{"title":"Correction: Deep Learning-Based Estimation of Radiographic Position to Automatically Set Up the X-Ray Prime Factors.","authors":"C F Del Cerro, R C Gimenez, J Garcia-Blas, K Sosenko, J M Ortega, M Desco, M Abella","doi":"10.1007/s10278-025-01476-9","DOIUrl":"10.1007/s10278-025-01476-9","url":null,"abstract":"","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"1040"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920969/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144060070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy-Preserving Large Language Model for Matching Findings and Tracking Interval Changes in Longitudinal Radiology Reports. 纵向放射学报告中用于匹配结果和跟踪间隔变化的隐私保护大语言模型。
Pub Date : 2026-02-01 Epub Date: 2025-04-11 DOI: 10.1007/s10278-025-01478-7
Tejas Sudharshan Mathai, Boah Kim, Oana M Stroie, Ronald M Summers

In current radiology practice, radiologists identify a finding in the current imaging exam, manually match it against the description from the prior exam report and assess interval changes. Large Language Models (LLMs) can identify report findings, but their ability to track interval changes has not been tested. The goal of this study was to determine the utility of a privacy-preserving LLM for matching findings between two reports (prior and follow-up) and tracking interval changes in size. In this retrospective study, body MRI reports from NIH (internal) were collected. A two-stage framework was employed for matching findings and tracking interval changes. In Stage 1, the LLM took a sentence from the follow-up report and discovered a matched finding in the prior report. In Stage 2, the LLM predicted the interval change status (increase, decrease, or stable) of the matched findings. Seven LLMs were locally evaluated and the best LLM was validated on an external non-contrast chest CT dataset. Agreement with the reference (radiologist) was measured using Cohen's Kappa (κ). The internal body MRI dataset had 240 studies (120 patients, mean age, 47 ± 16 years; 65 men) and the external non-contrast chest CT dataset contained 134 studies (67 patients, mean age, 58 ± 18 years; 44 men). On the internal dataset, TenyxChat-7B LLM fared the best for matching findings with an F1-score of 85.4% (95% CI: 80.8, 89.9) over the other LLMs (p < 0.05). For interval change detection, the same LLM achieved a 62.7% F1-score and showed a moderate agreement (κ = 0.46, 95% CI: 0.37, 0.55). For the external dataset, the same LLM attained F1-scores of 81.8% (95% CI: 74.4, 89.1) for matching findings and 77.4% for interval change detection respectively, with a substantial agreement (κ = 0.64, 95% CI: 0.49, 0.80). The TenyxChat-7B LLM used for matching longitudinal report findings and tracking interval changes showed moderate to substantial agreement with the reference standard. For structured reporting, the LLM can pre-fill the "Findings" section of the next follow-up exam report with a summary of longitudinal changes to important findings. It can also enhance the communication between the referring physician and radiologist.

在当前的放射学实践中,放射科医生识别当前影像学检查中的发现,手动将其与先前检查报告中的描述进行匹配,并评估间隔变化。大型语言模型(llm)可以识别报告结果,但是它们跟踪间隔变化的能力还没有经过测试。本研究的目的是确定隐私保护LLM在匹配两个报告(先前和后续)之间的发现以及跟踪大小的间隔变化方面的效用。在这项回顾性研究中,收集了来自NIH(内部)的身体MRI报告。采用两阶段框架来匹配结果和跟踪区间变化。在阶段1中,LLM从后续报告中提取了一个句子,并在先前的报告中发现了一个匹配的发现。在第二阶段,LLM预测匹配结果的层间变化状态(增加、减少或稳定)。对7个LLM进行局部评估,并在外部非对比胸部CT数据集上验证最佳LLM。采用Cohen’s Kappa (κ)测量与参考文献(放射科医师)的一致性。体内MRI数据集有240项研究(120例患者,平均年龄47±16岁;65名男性),外部非对比胸部CT数据集包含134项研究(67例患者,平均年龄58±18岁;44岁男性)。在内部数据集上,TenyxChat-7B LLM表现最好,与其他LLM相比,f1得分为85.4% (95% CI: 80.8, 89.9)
{"title":"Privacy-Preserving Large Language Model for Matching Findings and Tracking Interval Changes in Longitudinal Radiology Reports.","authors":"Tejas Sudharshan Mathai, Boah Kim, Oana M Stroie, Ronald M Summers","doi":"10.1007/s10278-025-01478-7","DOIUrl":"10.1007/s10278-025-01478-7","url":null,"abstract":"<p><p>In current radiology practice, radiologists identify a finding in the current imaging exam, manually match it against the description from the prior exam report and assess interval changes. Large Language Models (LLMs) can identify report findings, but their ability to track interval changes has not been tested. The goal of this study was to determine the utility of a privacy-preserving LLM for matching findings between two reports (prior and follow-up) and tracking interval changes in size. In this retrospective study, body MRI reports from NIH (internal) were collected. A two-stage framework was employed for matching findings and tracking interval changes. In Stage 1, the LLM took a sentence from the follow-up report and discovered a matched finding in the prior report. In Stage 2, the LLM predicted the interval change status (increase, decrease, or stable) of the matched findings. Seven LLMs were locally evaluated and the best LLM was validated on an external non-contrast chest CT dataset. Agreement with the reference (radiologist) was measured using Cohen's Kappa (κ). The internal body MRI dataset had 240 studies (120 patients, mean age, 47 ± 16 years; 65 men) and the external non-contrast chest CT dataset contained 134 studies (67 patients, mean age, 58 ± 18 years; 44 men). On the internal dataset, TenyxChat-7B LLM fared the best for matching findings with an F1-score of 85.4% (95% CI: 80.8, 89.9) over the other LLMs (p < 0.05). For interval change detection, the same LLM achieved a 62.7% F1-score and showed a moderate agreement (κ = 0.46, 95% CI: 0.37, 0.55). For the external dataset, the same LLM attained F1-scores of 81.8% (95% CI: 74.4, 89.1) for matching findings and 77.4% for interval change detection respectively, with a substantial agreement (κ = 0.64, 95% CI: 0.49, 0.80). The TenyxChat-7B LLM used for matching longitudinal report findings and tracking interval changes showed moderate to substantial agreement with the reference standard. For structured reporting, the LLM can pre-fill the \"Findings\" section of the next follow-up exam report with a summary of longitudinal changes to important findings. It can also enhance the communication between the referring physician and radiologist.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"1017-1030"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920831/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144061042","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-Performance Prompting for LLM Extraction of Compression Fracture Findings from Radiology Reports. 从放射学报告中提取压缩性骨折发现的高性能LLM提示。
Pub Date : 2026-02-01 Epub Date: 2025-05-16 DOI: 10.1007/s10278-025-01530-6
Mohammed M Kanani, Arezu Monawer, Lauryn Brown, William E King, Zachary D Miller, Nitin Venugopal, Patrick J Heagerty, Jeffrey G Jarvik, Trevor Cohen, Nathan M Cross

Extracting information from radiology reports can provide critical data to empower many radiology workflows. For spinal compression fractures, these data can facilitate evidence-based care for at-risk populations. Manual extraction from free-text reports is laborious, and error-prone. Large language models (LLMs) have shown promise; however, fine-tuning strategies to optimize performance in specific tasks can be resource intensive. A variety of prompting strategies have achieved similar results with fewer demands. Our study pioneers the use of Meta's Llama 3.1, together with prompt-based strategies, for automated extraction of compression fractures from free-text radiology reports, outputting structured data without model training. We tested performance on a time-based sample of CT exams covering the spine from 2/20/2024 to 2/22/2024 acquired across our healthcare enterprise (637 anonymized reports, age 18-102, 47% Female). Ground truth annotations were manually generated and compared against the performance of three models (Llama 3.1 70B, Llama 3.1 8B, and Vicuna 13B) with nine different prompting configurations for a total of 27 model/prompt experiments. The highest F1 score (0.91) was achieved by the 70B Llama 3.1 model when provided with a radiologist-written background, with similar results when the background was written by a separate LLM (0.86). The addition of few-shot examples to these prompts had variable impact on F1 measurements (0.89, 0.84 respectively). Comparable ROC-AUC and PR-AUC performance was observed. Our work demonstrated that an open-weights LLM excelled at extracting compression fractures findings from free-text radiology reports using prompt-based techniques without requiring extensive manually labeled examples for model training.

从放射学报告中提取信息可以为许多放射学工作流程提供关键数据。对于脊柱压缩性骨折,这些数据可以促进对高危人群的循证护理。从自由文本报告中手动提取是费力的,而且容易出错。大型语言模型(llm)已经显示出前景;然而,在特定任务中优化性能的微调策略可能是资源密集型的。各种提示策略都以较少的需求取得了相似的效果。我们的研究率先使用Meta的Llama 3.1,结合基于提示的策略,从自由文本放射学报告中自动提取压缩性骨折,无需模型训练即可输出结构化数据。我们对从2024年2月20日至2024年2月22日在我们的医疗保健企业收集的基于时间的脊柱CT检查样本进行了性能测试(637份匿名报告,年龄18-102岁,47%为女性)。人工生成地面真相标注,并与三种模型(Llama 3.1 70B、Llama 3.1 8B和Vicuna 13B)在9种不同提示配置下的性能进行比较,共进行27次模型/提示实验。当提供放射科医生编写的背景时,70B Llama 3.1模型获得了最高的F1分数(0.91),当由单独的LLM编写背景时,结果相似(0.86)。在这些提示中添加少量示例对F1测量值有不同的影响(分别为0.89和0.84)。对比ROC-AUC和PR-AUC性能。我们的工作表明,开放权重LLM擅长使用基于提示的技术从自由文本放射学报告中提取压缩性骨折结果,而不需要大量的手动标记示例进行模型训练。
{"title":"High-Performance Prompting for LLM Extraction of Compression Fracture Findings from Radiology Reports.","authors":"Mohammed M Kanani, Arezu Monawer, Lauryn Brown, William E King, Zachary D Miller, Nitin Venugopal, Patrick J Heagerty, Jeffrey G Jarvik, Trevor Cohen, Nathan M Cross","doi":"10.1007/s10278-025-01530-6","DOIUrl":"10.1007/s10278-025-01530-6","url":null,"abstract":"<p><p>Extracting information from radiology reports can provide critical data to empower many radiology workflows. For spinal compression fractures, these data can facilitate evidence-based care for at-risk populations. Manual extraction from free-text reports is laborious, and error-prone. Large language models (LLMs) have shown promise; however, fine-tuning strategies to optimize performance in specific tasks can be resource intensive. A variety of prompting strategies have achieved similar results with fewer demands. Our study pioneers the use of Meta's Llama 3.1, together with prompt-based strategies, for automated extraction of compression fractures from free-text radiology reports, outputting structured data without model training. We tested performance on a time-based sample of CT exams covering the spine from 2/20/2024 to 2/22/2024 acquired across our healthcare enterprise (637 anonymized reports, age 18-102, 47% Female). Ground truth annotations were manually generated and compared against the performance of three models (Llama 3.1 70B, Llama 3.1 8B, and Vicuna 13B) with nine different prompting configurations for a total of 27 model/prompt experiments. The highest F1 score (0.91) was achieved by the 70B Llama 3.1 model when provided with a radiologist-written background, with similar results when the background was written by a separate LLM (0.86). The addition of few-shot examples to these prompts had variable impact on F1 measurements (0.89, 0.84 respectively). Comparable ROC-AUC and PR-AUC performance was observed. Our work demonstrated that an open-weights LLM excelled at extracting compression fractures findings from free-text radiology reports using prompt-based techniques without requiring extensive manually labeled examples for model training.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"973-988"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921121/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144087436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of a No-Reference CT Image Quality Assessment Method Using RadImageNet Pre-trained Deep Learning Models. 基于RadImageNet预训练深度学习模型的无参考CT图像质量评估方法的开发。
Pub Date : 2026-02-01 Epub Date: 2025-05-27 DOI: 10.1007/s10278-025-01542-2
Kohei Ohashi, Yukihiro Nagatani, Asumi Yamazaki, Makoto Yoshigoe, Kyohei Iwai, Ryo Uemura, Masayuki Shimomura, Kenta Tanimura, Takayuki Ishida

Accurate assessment of computed tomography (CT) image quality is crucial for ensuring diagnostic accuracy, optimizing imaging protocols, and preventing excessive radiation exposure. In clinical settings, where high-quality reference images are often unavailable, developing no-reference image quality assessment (NR-IQA) methods is essential. Recently, CT-NR-IQA methods using deep learning have been widely studied; however, significant challenges remain in handling multiple degradation factors and accurately reflecting real-world degradations. To address these issues, we propose a novel CT-NR-IQA method. Our approach utilizes a dataset that combines two degradation factors (noise and blur) to train convolutional neural network (CNN) models capable of handling multiple degradation factors. Additionally, we leveraged RadImageNet pre-trained models (ResNet50, DenseNet121, InceptionV3, and InceptionResNetV2), allowing the models to learn deep features from large-scale real clinical images, thus enhancing adaptability to real-world degradations without relying on artificially degraded images. The models' performances were evaluated by measuring the correlation between the subjective scores and predicted image quality scores for both artificially degraded and real clinical image datasets. The results demonstrated positive correlations between the subjective and predicted scores for both datasets. In particular, ResNet50 showed the best performance, with a correlation coefficient of 0.910 for the artificially degraded images and 0.831 for the real clinical images. These findings indicate that the proposed method could serve as a potential surrogate for subjective assessment in CT-NR-IQA.

准确评估计算机断层扫描(CT)图像质量对于确保诊断准确性、优化成像方案和防止过度辐射暴露至关重要。在临床环境中,通常无法获得高质量的参考图像,因此开发无参考图像质量评估(NR-IQA)方法至关重要。近年来,利用深度学习的CT-NR-IQA方法得到了广泛的研究;然而,在处理多种退化因素和准确反映实际退化方面仍然存在重大挑战。为了解决这些问题,我们提出了一种新的CT-NR-IQA方法。我们的方法利用结合了两种退化因素(噪声和模糊)的数据集来训练能够处理多种退化因素的卷积神经网络(CNN)模型。此外,我们利用RadImageNet预训练模型(ResNet50, DenseNet121, InceptionV3和InceptionResNetV2),允许模型从大规模真实临床图像中学习深度特征,从而增强对现实世界退化的适应性,而不依赖于人工退化的图像。通过测量人工降级和真实临床图像数据集的主观评分和预测图像质量评分之间的相关性来评估模型的性能。结果表明,两个数据集的主观得分和预测得分之间存在正相关。其中,ResNet50表现出最好的性能,人工降级图像的相关系数为0.910,真实临床图像的相关系数为0.831。这些结果表明,该方法可作为CT-NR-IQA主观评价的潜在替代方法。
{"title":"Development of a No-Reference CT Image Quality Assessment Method Using RadImageNet Pre-trained Deep Learning Models.","authors":"Kohei Ohashi, Yukihiro Nagatani, Asumi Yamazaki, Makoto Yoshigoe, Kyohei Iwai, Ryo Uemura, Masayuki Shimomura, Kenta Tanimura, Takayuki Ishida","doi":"10.1007/s10278-025-01542-2","DOIUrl":"10.1007/s10278-025-01542-2","url":null,"abstract":"<p><p>Accurate assessment of computed tomography (CT) image quality is crucial for ensuring diagnostic accuracy, optimizing imaging protocols, and preventing excessive radiation exposure. In clinical settings, where high-quality reference images are often unavailable, developing no-reference image quality assessment (NR-IQA) methods is essential. Recently, CT-NR-IQA methods using deep learning have been widely studied; however, significant challenges remain in handling multiple degradation factors and accurately reflecting real-world degradations. To address these issues, we propose a novel CT-NR-IQA method. Our approach utilizes a dataset that combines two degradation factors (noise and blur) to train convolutional neural network (CNN) models capable of handling multiple degradation factors. Additionally, we leveraged RadImageNet pre-trained models (ResNet50, DenseNet121, InceptionV3, and InceptionResNetV2), allowing the models to learn deep features from large-scale real clinical images, thus enhancing adaptability to real-world degradations without relying on artificially degraded images. The models' performances were evaluated by measuring the correlation between the subjective scores and predicted image quality scores for both artificially degraded and real clinical image datasets. The results demonstrated positive correlations between the subjective and predicted scores for both datasets. In particular, ResNet50 showed the best performance, with a correlation coefficient of 0.910 for the artificially degraded images and 0.831 for the real clinical images. These findings indicate that the proposed method could serve as a potential surrogate for subjective assessment in CT-NR-IQA.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"46-58"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921086/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144164187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
New Machine Learning Method for Medical Image and Microarray Data Analysis for Heart Disease Classification. 心脏疾病分类医学图像和微阵列数据分析的新机器学习方法。
Pub Date : 2026-02-01 Epub Date: 2025-04-01 DOI: 10.1007/s10278-025-01492-9
Jinglan Guo, Jue Liao, Yuanlian Chen, Lisha Wen, Song Cheng

Microarray technology has become a vital tool in cardiovascular research, enabling the simultaneous analysis of thousands of gene expressions. This capability provides a robust foundation for heart disease classification and biomarker discovery. However, the high dimensionality, noise, and sparsity of microarray data present significant challenges for effective analysis. Gene selection, which aims to identify the most relevant subset of genes, is a crucial preprocessing step for improving classification accuracy, reducing computational complexity, and enhancing biological interpretability. Traditional gene selection methods often fall short in capturing complex, nonlinear interactions among genes, limiting their effectiveness in heart disease classification tasks. In this study, we propose a novel framework that leverages deep neural networks (DNNs) for optimizing gene selection and heart disease classification using microarray data. DNNs, known for their ability to model complex, nonlinear patterns, are integrated with feature selection techniques to address the challenges of high-dimensional data. The proposed method, DeepGeneNet (DGN), combines gene selection and DNN-based classification into a unified framework, ensuring robust performance and meaningful insights into the underlying biological mechanisms. Additionally, the framework incorporates hyperparameter optimization and innovative U-Net segmentation techniques to further enhance computational performance and classification accuracy. These optimizations enable DGN to deliver robust and scalable results, outperforming traditional methods in both predictive accuracy and interpretability. Experimental results demonstrate that the proposed approach significantly improves heart disease classification accuracy compared to other methods. By focusing on the interplay between gene selection and deep learning, this work advances the field of cardiovascular genomics, providing a scalable and interpretable framework for future applications.

微阵列技术已经成为心血管研究的重要工具,可以同时分析数千种基因表达。这种能力为心脏病分类和生物标志物的发现提供了坚实的基础。然而,微阵列数据的高维、噪声和稀疏性对有效分析提出了重大挑战。基因选择旨在识别最相关的基因子集,是提高分类准确性、降低计算复杂度和增强生物可解释性的关键预处理步骤。传统的基因选择方法往往无法捕捉到基因之间复杂的非线性相互作用,限制了它们在心脏病分类任务中的有效性。在这项研究中,我们提出了一个新的框架,利用深度神经网络(dnn)来优化基因选择和使用微阵列数据的心脏病分类。深度神经网络以其模拟复杂非线性模式的能力而闻名,它与特征选择技术相结合,以解决高维数据的挑战。所提出的方法DeepGeneNet (DGN)将基因选择和基于dnn的分类结合到一个统一的框架中,确保了强大的性能和对潜在生物学机制的有意义的见解。此外,该框架还结合了超参数优化和创新的U-Net分割技术,进一步提高了计算性能和分类精度。这些优化使DGN能够提供强大且可扩展的结果,在预测准确性和可解释性方面优于传统方法。实验结果表明,与其他方法相比,该方法显著提高了心脏病分类的准确率。通过关注基因选择和深度学习之间的相互作用,这项工作推动了心血管基因组学领域的发展,为未来的应用提供了一个可扩展和可解释的框架。
{"title":"New Machine Learning Method for Medical Image and Microarray Data Analysis for Heart Disease Classification.","authors":"Jinglan Guo, Jue Liao, Yuanlian Chen, Lisha Wen, Song Cheng","doi":"10.1007/s10278-025-01492-9","DOIUrl":"10.1007/s10278-025-01492-9","url":null,"abstract":"<p><p>Microarray technology has become a vital tool in cardiovascular research, enabling the simultaneous analysis of thousands of gene expressions. This capability provides a robust foundation for heart disease classification and biomarker discovery. However, the high dimensionality, noise, and sparsity of microarray data present significant challenges for effective analysis. Gene selection, which aims to identify the most relevant subset of genes, is a crucial preprocessing step for improving classification accuracy, reducing computational complexity, and enhancing biological interpretability. Traditional gene selection methods often fall short in capturing complex, nonlinear interactions among genes, limiting their effectiveness in heart disease classification tasks. In this study, we propose a novel framework that leverages deep neural networks (DNNs) for optimizing gene selection and heart disease classification using microarray data. DNNs, known for their ability to model complex, nonlinear patterns, are integrated with feature selection techniques to address the challenges of high-dimensional data. The proposed method, DeepGeneNet (DGN), combines gene selection and DNN-based classification into a unified framework, ensuring robust performance and meaningful insights into the underlying biological mechanisms. Additionally, the framework incorporates hyperparameter optimization and innovative U-Net segmentation techniques to further enhance computational performance and classification accuracy. These optimizations enable DGN to deliver robust and scalable results, outperforming traditional methods in both predictive accuracy and interpretability. Experimental results demonstrate that the proposed approach significantly improves heart disease classification accuracy compared to other methods. By focusing on the interplay between gene selection and deep learning, this work advances the field of cardiovascular genomics, providing a scalable and interpretable framework for future applications.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"884-907"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921063/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143766305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How Do Radiologists Currently Monitor AI in Radiology and What Challenges Do They Face? An Interview Study and Qualitative Analysis. 放射科医生目前如何监控放射学中的人工智能?他们面临哪些挑战?访谈研究与定性分析。
Pub Date : 2026-02-01 Epub Date: 2025-04-08 DOI: 10.1007/s10278-025-01493-8
Jamie Chow, Ryan Lee, Honghan Wu

Artificial intelligence (AI) in radiology is becoming increasingly prevalent; however, there is not a clear picture of how AI is being monitored today and how this should practically be done given the inherent risk of AI model performance degradation over time. This research investigates current practices and what difficulties radiologists face in monitoring AI. Semi-structured virtual interviews were conducted with 6 USA and 10 Europe-based radiologists. The interviews were automatically transcribed and underwent thematic analysis. The findings suggest that AI monitoring in radiology is still relatively nascent as most of the AI projects had not yet progressed into a fully live clinical deployment. The most common method of monitoring involved a manual process of retrospectively comparing the AI results against the radiology report. Automated and statistical methods of monitoring were much less common. The biggest challenges are a lack of resources to support AI monitoring and uncertainty about how to create a robust and scalable process of monitoring the breadth and variety of radiology AI applications available. There is currently a lack of practical guidelines on how to monitor AI which has led to a variety of approaches being proposed from both healthcare providers and vendors. An ensemble of mixed methods is recommended to monitor AI across multiple domains and metrics. This will be enabled by appropriate allocation of resources and the formation of robust and diverse multidisciplinary AI governance groups.

人工智能(AI)在放射学中的应用越来越普遍;然而,鉴于人工智能模型性能随着时间的推移而下降的固有风险,目前尚不清楚人工智能是如何被监控的,也不清楚应该如何实际做到这一点。本研究调查了目前的做法以及放射科医生在监测人工智能方面面临的困难。对6名美国和10名欧洲的放射科医生进行了半结构化的虚拟访谈。访谈内容被自动抄录并进行专题分析。研究结果表明,放射学中的人工智能监测仍处于相对初级阶段,因为大多数人工智能项目尚未发展到完全现场临床部署。最常见的监测方法是手动将人工智能结果与放射报告进行回顾性比较。自动化和统计监测方法则不太常见。最大的挑战是缺乏支持人工智能监测的资源,以及如何创建一个强大且可扩展的流程来监测可用放射学人工智能应用的广度和多样性的不确定性。目前缺乏关于如何监测人工智能的实用指南,这导致医疗保健提供者和供应商提出了各种方法。建议使用混合方法的集合来跨多个领域和度量监视AI。这将通过适当分配资源和组建强大和多样化的多学科人工智能治理小组来实现。
{"title":"How Do Radiologists Currently Monitor AI in Radiology and What Challenges Do They Face? An Interview Study and Qualitative Analysis.","authors":"Jamie Chow, Ryan Lee, Honghan Wu","doi":"10.1007/s10278-025-01493-8","DOIUrl":"10.1007/s10278-025-01493-8","url":null,"abstract":"<p><p>Artificial intelligence (AI) in radiology is becoming increasingly prevalent; however, there is not a clear picture of how AI is being monitored today and how this should practically be done given the inherent risk of AI model performance degradation over time. This research investigates current practices and what difficulties radiologists face in monitoring AI. Semi-structured virtual interviews were conducted with 6 USA and 10 Europe-based radiologists. The interviews were automatically transcribed and underwent thematic analysis. The findings suggest that AI monitoring in radiology is still relatively nascent as most of the AI projects had not yet progressed into a fully live clinical deployment. The most common method of monitoring involved a manual process of retrospectively comparing the AI results against the radiology report. Automated and statistical methods of monitoring were much less common. The biggest challenges are a lack of resources to support AI monitoring and uncertainty about how to create a robust and scalable process of monitoring the breadth and variety of radiology AI applications available. There is currently a lack of practical guidelines on how to monitor AI which has led to a variety of approaches being proposed from both healthcare providers and vendors. An ensemble of mixed methods is recommended to monitor AI across multiple domains and metrics. This will be enabled by appropriate allocation of resources and the formation of robust and diverse multidisciplinary AI governance groups.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"6-19"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920929/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143813388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Preoperative Prediction of Non-functional Pituitary Neuroendocrine Tumors and Posterior Pituitary Tumors Based on MRI Radiomic Features. 基于MRI放射学特征的非功能性垂体神经内分泌肿瘤和垂体后叶肿瘤术前预测。
Pub Date : 2026-02-01 Epub Date: 2025-04-14 DOI: 10.1007/s10278-025-01400-1
Shucheng Jin, Qin Xu, Chen Sun, Yuan Zhang, Yangyang Wang, Xi Wang, Xiudong Guan, Deling Li, Yiming Li, Chuanbao Zhang, Wang Jia

Compared to non-functional pituitary neuroendocrine tumors (NF-PitNETs), posterior pituitary tumors (PPTs) require more intraoperative protection of the pituitary stalk and hypothalamus, and their perioperative management is more complex than NF-PitNETs. However, they are difficult to be distinguished via magnetic resonance images (MRI) before operation. Based on clinical features and radiological signature extracted from MRI, this study aims to establish a model for distinguishing NF-PitNETs and PPTs. Preoperative MRI of 110 patients with NF-PitNETs and 55 patients with PPTs were retrospectively obtained. Patients were randomly assigned to the training (n = 110) and validation (n = 55) cohorts in a 2:1 ratio. The lest absolute shrinkage and selection operator (LASSO) algorithm was applied to develop a radiomic signature. Afterwards, an individualized predictive model (nomogram) incorporating radiomic signatures and predictive clinical features was developed. The nomogram's performance was evaluated by calibration and decision curve analyses. Five features derived from contrast-enhanced images were selected using the LASSO algorithm. Based on the mentioned methods, the calculation formula of radiomic score was obtained. The constructed nomogram incorporating radiomic signature and predictive clinical features showed a good calibration and outperformed the clinical features for predicting NF-PitNETs and PPTs (area under the curve [AUC]: 0.937 vs. 0.595 in training cohort [p < 0.001]; 0.907 vs. 0.782 in validation cohort [p = 0.03]). The decision curve shows that the individualized predictive model adds more benefit than clinical feature when the threshold probability ranges from 10 to 100%. Individualized predictive model provides a novel noninvasive imaging biomarker and could be conveniently used to distinguish NF-PitNETs and PPTs, which provides a significant reference for preoperative preparation and intraoperative decision-making.

与非功能性垂体神经内分泌肿瘤(NF-PitNETs)相比,垂体后叶肿瘤(pts)术中需要对垂体柄和下丘脑进行更多的保护,其围手术期处理也比NF-PitNETs更为复杂。然而,在手术前通过磁共振成像(MRI)很难区分它们。基于临床特征和MRI提取的放射学特征,本研究旨在建立区分NF-PitNETs和PPTs的模型。回顾性分析110例NF-PitNETs患者和55例PPTs患者的术前MRI。患者按2:1的比例随机分配到训练组(n = 110)和验证组(n = 55)。最小绝对收缩和选择算子(LASSO)算法应用于开发放射性特征。随后,结合放射学特征和预测临床特征的个体化预测模型(nomogram)被开发出来。通过标定和决策曲线分析来评价nomogram的性能。利用LASSO算法从对比度增强图像中选择5个特征。根据上述方法,得到了放射学评分的计算公式。结合放射学特征和预测临床特征构建的nomogram(诺图图)在预测NF-PitNETs和PPTs方面具有良好的校准效果,并且优于临床特征(曲线下面积[AUC]: 0.937 vs.训练队列中的0.595)
{"title":"Preoperative Prediction of Non-functional Pituitary Neuroendocrine Tumors and Posterior Pituitary Tumors Based on MRI Radiomic Features.","authors":"Shucheng Jin, Qin Xu, Chen Sun, Yuan Zhang, Yangyang Wang, Xi Wang, Xiudong Guan, Deling Li, Yiming Li, Chuanbao Zhang, Wang Jia","doi":"10.1007/s10278-025-01400-1","DOIUrl":"10.1007/s10278-025-01400-1","url":null,"abstract":"<p><p>Compared to non-functional pituitary neuroendocrine tumors (NF-PitNETs), posterior pituitary tumors (PPTs) require more intraoperative protection of the pituitary stalk and hypothalamus, and their perioperative management is more complex than NF-PitNETs. However, they are difficult to be distinguished via magnetic resonance images (MRI) before operation. Based on clinical features and radiological signature extracted from MRI, this study aims to establish a model for distinguishing NF-PitNETs and PPTs. Preoperative MRI of 110 patients with NF-PitNETs and 55 patients with PPTs were retrospectively obtained. Patients were randomly assigned to the training (n = 110) and validation (n = 55) cohorts in a 2:1 ratio. The lest absolute shrinkage and selection operator (LASSO) algorithm was applied to develop a radiomic signature. Afterwards, an individualized predictive model (nomogram) incorporating radiomic signatures and predictive clinical features was developed. The nomogram's performance was evaluated by calibration and decision curve analyses. Five features derived from contrast-enhanced images were selected using the LASSO algorithm. Based on the mentioned methods, the calculation formula of radiomic score was obtained. The constructed nomogram incorporating radiomic signature and predictive clinical features showed a good calibration and outperformed the clinical features for predicting NF-PitNETs and PPTs (area under the curve [AUC]: 0.937 vs. 0.595 in training cohort [p < 0.001]; 0.907 vs. 0.782 in validation cohort [p = 0.03]). The decision curve shows that the individualized predictive model adds more benefit than clinical feature when the threshold probability ranges from 10 to 100%. Individualized predictive model provides a novel noninvasive imaging biomarker and could be conveniently used to distinguish NF-PitNETs and PPTs, which provides a significant reference for preoperative preparation and intraoperative decision-making.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"115-126"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920986/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144056627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of imaging informatics in medicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1