首页 > 最新文献

Journal of imaging informatics in medicine最新文献

英文 中文
Improving Radiology Report Conciseness and Structure via Local Large Language Models. 利用局部大语言模型改进放射学报告的简洁性和结构。
Pub Date : 2026-02-01 Epub Date: 2025-04-21 DOI: 10.1007/s10278-025-01510-w
Iryna Hartsock, Cyrillo Araujo, Les Folio, Ghulam Rasool

Radiology reports are often lengthy and unstructured, posing challenges for referring physicians to quickly identify critical imaging findings while increasing risk of missed information. This retrospective study aimed to enhance radiology reports by making them concise and well-structured, with findings organized by relevant organs. To achieve this, we utilized private large language models (LLMs) deployed locally within our institution's firewall, ensuring data security and minimizing computational costs. Using a dataset of 814 radiology reports from seven board-certified body radiologists at [-blinded for review-], we tested five prompting strategies within the LangChain framework. After evaluating several models, the Mixtral LLM demonstrated superior adherence to formatting requirements compared to alternatives like Llama. The optimal strategy involved condensing reports first and then applying structured formatting based on specific instructions, reducing verbosity while improving clarity. Across all radiologists and reports, the Mixtral LLM reduced redundant word counts by more than 53%. These findings highlight the potential of locally deployed, open-source LLMs to streamline radiology reporting. By generating concise, well-structured reports, these models enhance information retrieval and better meet the needs of referring physicians, ultimately improving clinical workflows.

放射学报告通常冗长且无结构,这给转诊医生快速识别关键影像发现带来了挑战,同时增加了遗漏信息的风险。本回顾性研究的目的是提高放射学报告的简练和结构良好,结果由相关器官组织。为了实现这一点,我们使用了私有的大型语言模型(llm),部署在我们机构的防火墙中,以确保数据安全性并将计算成本降至最低。使用来自7位委员会认证的放射学家的814份放射学报告的数据集,我们在LangChain框架内测试了5种提示策略。在对几种模型进行评估后,与Llama等替代方案相比,Mixtral LLM显示出对格式要求的优越依从性。最优策略包括首先压缩报告,然后根据具体指示应用结构化格式,在提高清晰度的同时减少冗长。在所有放射科医生和报告中,midtral LLM减少了53%以上的冗余字数。这些发现突出了本地部署的开源法学硕士在简化放射学报告方面的潜力。通过生成简洁、结构良好的报告,这些模型增强了信息检索,更好地满足转诊医生的需求,最终改善了临床工作流程。
{"title":"Improving Radiology Report Conciseness and Structure via Local Large Language Models.","authors":"Iryna Hartsock, Cyrillo Araujo, Les Folio, Ghulam Rasool","doi":"10.1007/s10278-025-01510-w","DOIUrl":"10.1007/s10278-025-01510-w","url":null,"abstract":"<p><p>Radiology reports are often lengthy and unstructured, posing challenges for referring physicians to quickly identify critical imaging findings while increasing risk of missed information. This retrospective study aimed to enhance radiology reports by making them concise and well-structured, with findings organized by relevant organs. To achieve this, we utilized private large language models (LLMs) deployed locally within our institution's firewall, ensuring data security and minimizing computational costs. Using a dataset of 814 radiology reports from seven board-certified body radiologists at [-blinded for review-], we tested five prompting strategies within the LangChain framework. After evaluating several models, the Mixtral LLM demonstrated superior adherence to formatting requirements compared to alternatives like Llama. The optimal strategy involved condensing reports first and then applying structured formatting based on specific instructions, reducing verbosity while improving clarity. Across all radiologists and reports, the Mixtral LLM reduced redundant word counts by more than 53%. These findings highlight the potential of locally deployed, open-source LLMs to streamline radiology reporting. By generating concise, well-structured reports, these models enhance information retrieval and better meet the needs of referring physicians, ultimately improving clinical workflows.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"1005-1016"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921097/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144059558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning for Classification of Solid Renal Parenchymal Tumors Using Contrast-Enhanced Ultrasound. 深度学习在增强超声实性肾实质肿瘤分类中的应用。
Pub Date : 2026-02-01 Epub Date: 2025-05-06 DOI: 10.1007/s10278-025-01525-3
Yun Bai, Zi-Chen An, Lian-Fang Du, Fan Li, Ying-Yu Cai

The purpose of this study is to assess the ability of deep learning models to classify different subtypes of solid renal parenchymal tumors using contrast-enhanced ultrasound (CEUS) images and to compare their classification performance. A retrospective study was conducted using CEUS images of 237 kidney tumors, including 46 angiomyolipomas (AML), 118 clear cell renal cell carcinomas (ccRCC), 48 papillary RCCs (pRCC), and 25 chromophobe RCCs (chRCC), collected from January 2017 to December 2019. Two deep learning models, based on the ResNet-18 and RepVGG architectures, were trained and validated to distinguish between these subtypes. The models' performance was assessed using sensitivity, specificity, positive predictive value, negative predictive value, F1 score, Matthews correlation coefficient, accuracy, area under the receiver operating characteristic curve (AUC), and confusion matrix analysis. Class activation mapping (CAM) was applied to visualize the specific regions that contributed to the models' predictions. The ResNet-18 and RepVGG-A0 models achieved an overall accuracy of 76.7% and 84.5% across all four subtypes. The AUCs for AML, ccRCC, pRCC, and chRCC were 0.832, 0.829, 0.806, and 0.795 for the ResNet-18 model, compared to 0.906, 0.911, 0.840, and 0.827 for the RepVGG-A0 model, respectively. The deep learning models could reliably differentiate between various histological subtypes of renal tumors using CEUS images in an objective and non-invasive manner.

本研究的目的是评估深度学习模型使用对比增强超声(CEUS)图像对实体肾实质肿瘤不同亚型进行分类的能力,并比较它们的分类性能。对2017年1月至2019年12月收集的237例肾脏肿瘤的超声造影图像进行回顾性研究,其中包括46例血管平滑肌脂肪瘤(AML)、118例透明细胞肾细胞癌(ccRCC)、48例乳头状肾细胞癌(pRCC)和25例疏色肾细胞癌(chRCC)。基于ResNet-18和RepVGG架构的两种深度学习模型进行了训练和验证,以区分这些亚型。采用敏感性、特异性、阳性预测值、阴性预测值、F1评分、Matthews相关系数、准确性、受试者工作特征曲线下面积(AUC)和混淆矩阵分析来评估模型的性能。类激活映射(CAM)应用于可视化特定区域,这些区域有助于模型的预测。ResNet-18和RepVGG-A0模型对所有四种亚型的总体准确率分别为76.7%和84.5%。ResNet-18模型的AML、ccRCC、pRCC和chRCC的auc分别为0.832、0.829、0.806和0.795,而RepVGG-A0模型的auc分别为0.906、0.911、0.840和0.827。深度学习模型可以客观、无创地利用超声造影图像可靠地区分肾肿瘤的各种组织学亚型。
{"title":"Deep Learning for Classification of Solid Renal Parenchymal Tumors Using Contrast-Enhanced Ultrasound.","authors":"Yun Bai, Zi-Chen An, Lian-Fang Du, Fan Li, Ying-Yu Cai","doi":"10.1007/s10278-025-01525-3","DOIUrl":"10.1007/s10278-025-01525-3","url":null,"abstract":"<p><p>The purpose of this study is to assess the ability of deep learning models to classify different subtypes of solid renal parenchymal tumors using contrast-enhanced ultrasound (CEUS) images and to compare their classification performance. A retrospective study was conducted using CEUS images of 237 kidney tumors, including 46 angiomyolipomas (AML), 118 clear cell renal cell carcinomas (ccRCC), 48 papillary RCCs (pRCC), and 25 chromophobe RCCs (chRCC), collected from January 2017 to December 2019. Two deep learning models, based on the ResNet-18 and RepVGG architectures, were trained and validated to distinguish between these subtypes. The models' performance was assessed using sensitivity, specificity, positive predictive value, negative predictive value, F1 score, Matthews correlation coefficient, accuracy, area under the receiver operating characteristic curve (AUC), and confusion matrix analysis. Class activation mapping (CAM) was applied to visualize the specific regions that contributed to the models' predictions. The ResNet-18 and RepVGG-A0 models achieved an overall accuracy of 76.7% and 84.5% across all four subtypes. The AUCs for AML, ccRCC, pRCC, and chRCC were 0.832, 0.829, 0.806, and 0.795 for the ResNet-18 model, compared to 0.906, 0.911, 0.840, and 0.827 for the RepVGG-A0 model, respectively. The deep learning models could reliably differentiate between various histological subtypes of renal tumors using CEUS images in an objective and non-invasive manner.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"277-285"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920943/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144060016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Role of Model Size and Prompting Strategies in Extracting Labels from Free-Text Radiology Reports with Open-Source Large Language Models. 模型大小和提示策略在使用开源大型语言模型从自由文本放射学报告中提取标签中的作用。
Pub Date : 2026-02-01 Epub Date: 2025-05-05 DOI: 10.1007/s10278-025-01505-7
Bardia Khosravi, Theo Dapamede, Frank Li, Zvipo Chisango, Anirudh Bikmal, Sara Garg, Babajide Owosela, Amirali Khosravi, Mohammadreza Chavoshi, Hari M Trivedi, Cody C Wyles, Saptarshi Purkayastha, Bradley J Erickson, Judy W Gichoya

Extracting accurate labels from radiology reports is essential for training medical image analysis models. Large language models (LLMs) show promise for automating this process. The purpose of this study is to evaluate how model size and prompting strategies affect label extraction accuracy and downstream performance in open-source LLMs. Three open-source LLMs (Llama-3, Phi-3 mini, and Zephyr-beta) were used to extract labels from 227,827 MIMIC-CXR radiology reports. Performance was evaluated against human annotations on 2000 MIMIC-CXR reports, and through training image classifiers for pneumothorax and rib fracture detection tested on the CANDID-PTX dataset (n = 19,237). LLM-based labeling outperformed the CheXpert labeler, with the best LLM achieving 95% sensitivity for fracture detection versus CheXpert's 51%. Larger models showed better sensitivity, while chain-of-thought prompting had variable effects. Image classifiers showed resilience to labeling noise when tested externally. The choice of test set labeling schema significantly affected reported performance-a classifier trained on Llama-3 with chain-of-thought labels achieved AUCs of 0.96 and 0.84 for pneumothorax and fracture detection respectively when evaluated against human annotations, compared to 0.91 and 0.73 when evaluated on CheXpert labels. Open-source LLMs effectively extract labels from radiology reports at scale. While larger pre-trained models generally perform better, the choice of model size and prompting strategy should be task specific. Careful consideration of evaluation methods is critical for interpreting classifier performance.

从放射学报告中提取准确的标签对于训练医学图像分析模型至关重要。大型语言模型(llm)显示了自动化这一过程的希望。本研究的目的是评估模型大小和提示策略如何影响开源llm中的标签提取准确性和下游性能。三个开源llm (lama-3, Phi-3 mini和Zephyr-beta)用于从227,827份MIMIC-CXR放射学报告中提取标签。通过对2000份mimi - cxr报告的人类注释进行性能评估,并通过在CANDID-PTX数据集(n = 19,237)上测试的气胸和肋骨骨折检测的训练图像分类器进行性能评估。基于LLM的标记优于CheXpert标记器,最佳LLM对裂缝检测的灵敏度达到95%,而CheXpert为51%。较大的模型显示出更好的灵敏度,而思维链提示的效果则不同。在外部测试时,图像分类器显示出对标记噪声的弹性。测试集标记模式的选择显著影响了报告的性能——在Llama-3上训练的分类器,使用思维链标签对人类注释进行评估时,气胸和骨折检测的auc分别为0.96和0.84,而在CheXpert标签上进行评估时,auc分别为0.91和0.73。开源llm有效地从放射学报告中大规模提取标签。虽然较大的预训练模型通常表现更好,但模型大小和提示策略的选择应该是特定于任务的。仔细考虑评估方法对于解释分类器性能至关重要。
{"title":"Role of Model Size and Prompting Strategies in Extracting Labels from Free-Text Radiology Reports with Open-Source Large Language Models.","authors":"Bardia Khosravi, Theo Dapamede, Frank Li, Zvipo Chisango, Anirudh Bikmal, Sara Garg, Babajide Owosela, Amirali Khosravi, Mohammadreza Chavoshi, Hari M Trivedi, Cody C Wyles, Saptarshi Purkayastha, Bradley J Erickson, Judy W Gichoya","doi":"10.1007/s10278-025-01505-7","DOIUrl":"10.1007/s10278-025-01505-7","url":null,"abstract":"<p><p>Extracting accurate labels from radiology reports is essential for training medical image analysis models. Large language models (LLMs) show promise for automating this process. The purpose of this study is to evaluate how model size and prompting strategies affect label extraction accuracy and downstream performance in open-source LLMs. Three open-source LLMs (Llama-3, Phi-3 mini, and Zephyr-beta) were used to extract labels from 227,827 MIMIC-CXR radiology reports. Performance was evaluated against human annotations on 2000 MIMIC-CXR reports, and through training image classifiers for pneumothorax and rib fracture detection tested on the CANDID-PTX dataset (n = 19,237). LLM-based labeling outperformed the CheXpert labeler, with the best LLM achieving 95% sensitivity for fracture detection versus CheXpert's 51%. Larger models showed better sensitivity, while chain-of-thought prompting had variable effects. Image classifiers showed resilience to labeling noise when tested externally. The choice of test set labeling schema significantly affected reported performance-a classifier trained on Llama-3 with chain-of-thought labels achieved AUCs of 0.96 and 0.84 for pneumothorax and fracture detection respectively when evaluated against human annotations, compared to 0.91 and 0.73 when evaluated on CheXpert labels. Open-source LLMs effectively extract labels from radiology reports at scale. While larger pre-trained models generally perform better, the choice of model size and prompting strategy should be task specific. Careful consideration of evaluation methods is critical for interpreting classifier performance.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"995-1004"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920854/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144061006","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Pulmonary Embolism Detection with Integrated Deep Learning Architectures. 利用集成深度学习架构推进肺栓塞检测。
Pub Date : 2026-02-01 Epub Date: 2025-04-25 DOI: 10.1007/s10278-025-01506-6
Can Berk Biret, Sukru Gurbuz, Erhan Akbal, Mehmet Baygin, Evren Ekingen, Serdar Derya, I Okan Yıldırım, Ilknur Sercek, Sengul Dogan, Turker Tuncer

The main aim of this study is to introduce a new hybrid deep learning model for biomedical image classification. We propose a novel convolutional neural network (CNN), named HybridNeXt, for detecting pulmonary embolism (PE) from computed tomography (CT) images. To evaluate the HybridNeXt model, we created a new dataset consisting of two classes: (1) PE and (2) control. The HybridNeXt architecture combines different advanced CNN blocks, including MobileNet, ResNet, ConvNeXt, and Swin Transformer. We specifically designed this model to combine the strengths of these well-known CNNs. The architecture also includes stem, downsampling, and output stages. By adjusting the parameters, we developed a lightweight version of HybridNeXt, suitable for clinical use. To further improve the classification performance and demonstrate transfer learning capability, we proposed a deep feature engineering (DFE) method using a multilevel discrete wavelet transform (MDWT). This DFE model has three main phases: (i) feature extraction from raw images and wavelet bands, (ii) feature selection using iterative neighborhood component analysis (INCA), and (iii) classification using a k-nearest neighbors (kNN) classifier. We first trained HybridNeXt on the training images, creating a pretrained HybridNeXt model. Then, using this pretrained model, we extracted features and applied the proposed DFE method for classification. The HybridNeXt model achieved a test accuracy of 90.14%, while our DFE model improved accuracy to 96.35%. Overall, the results confirm that our HybridNeXt architecture is highly accurate and effective for biomedical image classification. The presented HybridNeXt and HybridNeXt-based DFE methods can potentially be applied to other image classification tasks.

本研究的主要目的是引入一种新的混合深度学习模型用于生物医学图像分类。我们提出了一种新的卷积神经网络(CNN),命名为HybridNeXt,用于从计算机断层扫描(CT)图像中检测肺栓塞(PE)。为了评估HybridNeXt模型,我们创建了一个由两类组成的新数据集:(1)PE和(2)control。HybridNeXt架构结合了不同的高级CNN模块,包括MobileNet、ResNet、ConvNeXt和Swin Transformer。我们特别设计了这个模型来结合这些知名cnn的优势。该体系结构还包括干、下采样和输出级。通过调整参数,我们开发了适合临床使用的轻量级HybridNeXt。为了进一步提高分类性能和展示迁移学习能力,我们提出了一种基于多级离散小波变换(MDWT)的深度特征工程(DFE)方法。该DFE模型有三个主要阶段:(i)从原始图像和小波带中提取特征,(ii)使用迭代邻域成分分析(INCA)进行特征选择,以及(iii)使用k近邻(kNN)分类器进行分类。我们首先在训练图像上训练HybridNeXt,创建一个预训练的HybridNeXt模型。然后,利用该预训练模型提取特征,并应用所提出的DFE方法进行分类。HybridNeXt模型的测试准确率为90.14%,而我们的DFE模型将准确率提高到96.35%。总的来说,结果证实了我们的HybridNeXt架构对于生物医学图像分类是高度准确和有效的。本文提出的基于HybridNeXt和HybridNeXt的DFE方法可以应用于其他图像分类任务。
{"title":"Advancing Pulmonary Embolism Detection with Integrated Deep Learning Architectures.","authors":"Can Berk Biret, Sukru Gurbuz, Erhan Akbal, Mehmet Baygin, Evren Ekingen, Serdar Derya, I Okan Yıldırım, Ilknur Sercek, Sengul Dogan, Turker Tuncer","doi":"10.1007/s10278-025-01506-6","DOIUrl":"10.1007/s10278-025-01506-6","url":null,"abstract":"<p><p>The main aim of this study is to introduce a new hybrid deep learning model for biomedical image classification. We propose a novel convolutional neural network (CNN), named HybridNeXt, for detecting pulmonary embolism (PE) from computed tomography (CT) images. To evaluate the HybridNeXt model, we created a new dataset consisting of two classes: (1) PE and (2) control. The HybridNeXt architecture combines different advanced CNN blocks, including MobileNet, ResNet, ConvNeXt, and Swin Transformer. We specifically designed this model to combine the strengths of these well-known CNNs. The architecture also includes stem, downsampling, and output stages. By adjusting the parameters, we developed a lightweight version of HybridNeXt, suitable for clinical use. To further improve the classification performance and demonstrate transfer learning capability, we proposed a deep feature engineering (DFE) method using a multilevel discrete wavelet transform (MDWT). This DFE model has three main phases: (i) feature extraction from raw images and wavelet bands, (ii) feature selection using iterative neighborhood component analysis (INCA), and (iii) classification using a k-nearest neighbors (kNN) classifier. We first trained HybridNeXt on the training images, creating a pretrained HybridNeXt model. Then, using this pretrained model, we extracted features and applied the proposed DFE method for classification. The HybridNeXt model achieved a test accuracy of 90.14%, while our DFE model improved accuracy to 96.35%. Overall, the results confirm that our HybridNeXt architecture is highly accurate and effective for biomedical image classification. The presented HybridNeXt and HybridNeXt-based DFE methods can potentially be applied to other image classification tasks.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"186-201"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921004/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144065465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain Shift Analysis in Chest Radiographs Classification in a Veterans Healthcare Administration Population. 退伍军人医疗管理人群胸片分类的域移分析。
Pub Date : 2026-02-01 Epub Date: 2025-04-11 DOI: 10.1007/s10278-025-01494-7
Mayanka Chandrashekar, Ian Goethert, Md Inzamam Ul Haque, Benjamin McMahon, Sayera Dhaubhadel, Kathryn Knight, Joseph Erdos, Donna Reagan, Caroline Taylor, Peter Kuzmak, John Michael Gaziano, Eileen McAllister, Lauren Costa, Yuk-Lam Ho, Kelly Cho, Suzanne Tamang, Samah Fodeh-Jarad, Olga S Ovchinnikova, Amy C Justice, Jacob Hinkle, Ioana Danciu

This study aims to assess the impact of domain shift on chest X-ray classification accuracy and to analyze the influence of ground truth label quality and demographic factors such as age group, sex, and study year. We used a DenseNet121 model pre-trained MIMIC-CXR dataset for deep learning-based multi-label classification using ground truth labels from radiology reports extracted using the CheXpert and CheXbert Labeler. We compared the performance of the 14 chest X-ray labels on the MIMIC-CXR and Veterans Healthcare Administration chest X-ray dataset (VA-CXR). The validation of ground truth and the assessment of multi-label classification performance across various NLP extraction tools revealed that the VA-CXR dataset exhibited lower disagreement rates than the MIMIC-CXR datasets. Additionally, there were notable differences in AUC scores between models utilizing CheXpert and CheXbert. When evaluating multi-label classification performance across different datasets, minimal domain shift was observed in the unseen VA dataset, except for the label "Enlarged Cardiomediastinum." The subgroup with the most significant variations in multi-label classification performance was study year. These findings underscore the importance of considering domain shift in chest X-ray classification tasks, paying particular attention to the temporality of the exam. Our study reveals the significant impact of domain shift and demographic factors on chest X-ray classification, emphasizing the need for improved transfer learning and robust model development. Addressing these challenges is crucial for advancing medical imaging research and improving patient care.

本研究旨在评估领域移位对胸部x线分类准确性的影响,并分析地面真相标签质量和人口统计学因素(如年龄、性别和研究年份)的影响。我们使用DenseNet121模型预训练的MIMIC-CXR数据集进行基于深度学习的多标签分类,使用CheXpert和CheXbert Labeler提取的放射学报告中的地面真实标签。我们比较了MIMIC-CXR和退伍军人医疗管理局胸片数据集(VA-CXR)上14个胸片标签的性能。对各种NLP提取工具的基础真实性验证和多标签分类性能评估表明,VA-CXR数据集的歧异率低于MIMIC-CXR数据集。此外,使用CheXpert和CheXbert的模型之间的AUC评分存在显着差异。当评估跨不同数据集的多标签分类性能时,除了标签“心膈增大”外,在未见的VA数据集中观察到最小的域移位。多标签分类表现变化最显著的亚组为研究年度。这些发现强调了在胸部x线分类任务中考虑域转移的重要性,特别注意检查的时间性。我们的研究揭示了领域转移和人口统计学因素对胸部x线分类的重要影响,强调了改进迁移学习和稳健模型开发的必要性。解决这些挑战对于推进医学成像研究和改善患者护理至关重要。
{"title":"Domain Shift Analysis in Chest Radiographs Classification in a Veterans Healthcare Administration Population.","authors":"Mayanka Chandrashekar, Ian Goethert, Md Inzamam Ul Haque, Benjamin McMahon, Sayera Dhaubhadel, Kathryn Knight, Joseph Erdos, Donna Reagan, Caroline Taylor, Peter Kuzmak, John Michael Gaziano, Eileen McAllister, Lauren Costa, Yuk-Lam Ho, Kelly Cho, Suzanne Tamang, Samah Fodeh-Jarad, Olga S Ovchinnikova, Amy C Justice, Jacob Hinkle, Ioana Danciu","doi":"10.1007/s10278-025-01494-7","DOIUrl":"10.1007/s10278-025-01494-7","url":null,"abstract":"<p><p>This study aims to assess the impact of domain shift on chest X-ray classification accuracy and to analyze the influence of ground truth label quality and demographic factors such as age group, sex, and study year. We used a DenseNet121 model pre-trained MIMIC-CXR dataset for deep learning-based multi-label classification using ground truth labels from radiology reports extracted using the CheXpert and CheXbert Labeler. We compared the performance of the 14 chest X-ray labels on the MIMIC-CXR and Veterans Healthcare Administration chest X-ray dataset (VA-CXR). The validation of ground truth and the assessment of multi-label classification performance across various NLP extraction tools revealed that the VA-CXR dataset exhibited lower disagreement rates than the MIMIC-CXR datasets. Additionally, there were notable differences in AUC scores between models utilizing CheXpert and CheXbert. When evaluating multi-label classification performance across different datasets, minimal domain shift was observed in the unseen VA dataset, except for the label \"Enlarged Cardiomediastinum.\" The subgroup with the most significant variations in multi-label classification performance was study year. These findings underscore the importance of considering domain shift in chest X-ray classification tasks, paying particular attention to the temporality of the exam. Our study reveals the significant impact of domain shift and demographic factors on chest X-ray classification, emphasizing the need for improved transfer learning and robust model development. Addressing these challenges is crucial for advancing medical imaging research and improving patient care.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"484-499"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921069/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144016371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge, Utilisation, and Challenges of Medical Doctors Using Picture Archiving and Communication Systems at a Tertiary Academic Hospital in the Eastern Cape, South Africa. 在南非东开普省的一家三级学术医院,医生使用图片存档和通信系统的知识、利用和挑战。
Pub Date : 2026-02-01 Epub Date: 2025-05-16 DOI: 10.1007/s10278-025-01526-2
Mpatisi Lobi, Anne Faith Namugenyi, Oladele Vincent Adeniyi

Despite the investments in the picture archiving and communication system (PACS) in the South African health facilities, it is unclear whether clinicians in the rural tertiary hospitals are using this tool maximally. This study determines the level of knowledge, utilisation, and challenges associated with the use of PACS among doctors in a rural tertiary hospital in the Eastern Cape Province. In this cross-sectional descriptive study, a total of 66 medical doctors drawn from different departments completed a structured questionnaire at the Nelson Mandela Academic Hospital, Mthatha. Relevant items on knowledge and use of PACS, including challenges experienced, were obtained. The mean age of the respondents was 36 (± standard deviation 9.58) years. The majority of the doctors (n = 42; 63.7%) demonstrated moderate to good knowledge of PACS. Similarly, a substantial majority (n = 55; 83.3%) have used PACS for years, for both images and reports (49.2%). The highest proportion of the respondents had at least 1 year of PACS experience (63.5%). Though there was no association between the sociodemographics and level of knowledge, the duration of use (p = 0.025) and frequency of use (p = 0.025) were significantly associated with moderate to good knowledge of PACS. Internet connectivity and mobile PACS were the major challenges identified. The study found moderate to good knowledge of PACS among the final sample of 66 clinicians. A substantial majority of the clinicians had used PACS for years; however, there is considerable room for strengthening and expanding the use of PACS in the study setting.

尽管南非卫生设施对图片存档和通信系统(PACS)进行了投资,但尚不清楚农村三级医院的临床医生是否最大限度地利用了这一工具。本研究确定了东开普省一家农村三级医院医生使用PACS的知识水平、利用率和挑战。在这项横断面描述性研究中,共有66名来自不同部门的医生在姆塔塔纳尔逊·曼德拉学术医院完成了一份结构化问卷。获得了关于PACS的知识和使用的相关项目,包括所经历的挑战。受访者平均年龄36岁(±标准差9.58)岁。大多数医生(n = 42;63.7%)对PACS有中等到良好的了解。同样,绝大多数(n = 55;83.3%)使用PACS多年,用于图像和报告(49.2%)。受访者中至少有1年PACS经验的比例最高(63.5%)。虽然社会人口统计学与知识水平之间没有关联,但使用时间(p = 0.025)和使用频率(p = 0.025)与中度至良好的PACS知识显著相关。互联网连接和移动PACS是确定的主要挑战。研究发现,在66名临床医生的最终样本中,对PACS有中等到良好的了解。绝大多数临床医生已使用PACS多年;然而,在研究环境中加强和扩大PACS的使用仍有相当大的空间。
{"title":"Knowledge, Utilisation, and Challenges of Medical Doctors Using Picture Archiving and Communication Systems at a Tertiary Academic Hospital in the Eastern Cape, South Africa.","authors":"Mpatisi Lobi, Anne Faith Namugenyi, Oladele Vincent Adeniyi","doi":"10.1007/s10278-025-01526-2","DOIUrl":"10.1007/s10278-025-01526-2","url":null,"abstract":"<p><p>Despite the investments in the picture archiving and communication system (PACS) in the South African health facilities, it is unclear whether clinicians in the rural tertiary hospitals are using this tool maximally. This study determines the level of knowledge, utilisation, and challenges associated with the use of PACS among doctors in a rural tertiary hospital in the Eastern Cape Province. In this cross-sectional descriptive study, a total of 66 medical doctors drawn from different departments completed a structured questionnaire at the Nelson Mandela Academic Hospital, Mthatha. Relevant items on knowledge and use of PACS, including challenges experienced, were obtained. The mean age of the respondents was 36 (± standard deviation 9.58) years. The majority of the doctors (n = 42; 63.7%) demonstrated moderate to good knowledge of PACS. Similarly, a substantial majority (n = 55; 83.3%) have used PACS for years, for both images and reports (49.2%). The highest proportion of the respondents had at least 1 year of PACS experience (63.5%). Though there was no association between the sociodemographics and level of knowledge, the duration of use (p = 0.025) and frequency of use (p = 0.025) were significantly associated with moderate to good knowledge of PACS. Internet connectivity and mobile PACS were the major challenges identified. The study found moderate to good knowledge of PACS among the final sample of 66 clinicians. A substantial majority of the clinicians had used PACS for years; however, there is considerable room for strengthening and expanding the use of PACS in the study setting.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"34-45"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920820/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144087438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial Intelligence and Data Science Methods for Automatic Detection of White Blood Cells in Images. 图像中白细胞自动检测的人工智能和数据科学方法。
Pub Date : 2026-02-01 Epub Date: 2025-05-16 DOI: 10.1007/s10278-025-01538-y
Yawo M Kobara, Ikpe Justice Akpan, Alima Damipe Nam, Firas H AlMukthar, Mbuotidem Peter

Data scieQuerynce (DS) methods and artificial intelligence (AI) are critical in today's healthcare services operations. This study focuses on evaluating the effectiveness of AI and DS in biomedical diagnostics, including automatic detection and counting of white blood cells (WBCs) and types, which provide valuable information for diagnosing and treating blood diseases such as leukemia. Automating these tasks using AI and DS saves time and avoids or minimizes errors compared to manual processes, which can be complex and error prone. The study utilizes bibliographic data from SCOPUS to evaluate research on applying AI algorithms and DS methods for mapping and classifying WBC images for treatment of blood diseases, such as leukemia using literature survey and science mapping methodology. The results show the potency of different DS methods and AI algorithms, such as machine learning, deep learning, and classification algorithms that enable the automatic detection of WBC images. AI and DS algorithms offer critical benefits in effectively and efficiently analyzing microscopic images of blood cells. The automatic identification, localization, and classification of WBCs speed up the patient diagnosis process, allowing hematologists to focus on interpreting results. Automatic processes identify specific abnormalities and patterns, enhancing accuracy and timely diagnoses. Future work will examine the application of generative AI in blood cells diagnostics.

数据科学(DS)方法和人工智能(AI)在当今的医疗保健服务运营中至关重要。本研究的重点是评估AI和DS在生物医学诊断中的有效性,包括自动检测和计数白细胞(wbc)和类型,为诊断和治疗白血病等血液疾病提供有价值的信息。与复杂且容易出错的手动流程相比,使用AI和DS自动化这些任务可以节省时间,避免或最大限度地减少错误。本研究利用SCOPUS的书目数据,利用文献调查和科学制图方法,对应用AI算法和DS方法对白细胞图像进行制图和分类以治疗血液病(如白血病)的研究进行了评价。结果显示了不同的DS方法和人工智能算法的效力,例如机器学习、深度学习和分类算法,这些算法可以自动检测WBC图像。人工智能和DS算法在有效和高效地分析血细胞显微图像方面提供了关键的好处。白细胞的自动识别、定位和分类加快了患者的诊断过程,使血液学家能够专注于解释结果。自动过程识别特定的异常和模式,提高准确性和及时诊断。未来的工作将研究生成人工智能在血细胞诊断中的应用。
{"title":"Artificial Intelligence and Data Science Methods for Automatic Detection of White Blood Cells in Images.","authors":"Yawo M Kobara, Ikpe Justice Akpan, Alima Damipe Nam, Firas H AlMukthar, Mbuotidem Peter","doi":"10.1007/s10278-025-01538-y","DOIUrl":"10.1007/s10278-025-01538-y","url":null,"abstract":"<p><p>Data scieQuerynce (DS) methods and artificial intelligence (AI) are critical in today's healthcare services operations. This study focuses on evaluating the effectiveness of AI and DS in biomedical diagnostics, including automatic detection and counting of white blood cells (WBCs) and types, which provide valuable information for diagnosing and treating blood diseases such as leukemia. Automating these tasks using AI and DS saves time and avoids or minimizes errors compared to manual processes, which can be complex and error prone. The study utilizes bibliographic data from SCOPUS to evaluate research on applying AI algorithms and DS methods for mapping and classifying WBC images for treatment of blood diseases, such as leukemia using literature survey and science mapping methodology. The results show the potency of different DS methods and AI algorithms, such as machine learning, deep learning, and classification algorithms that enable the automatic detection of WBC images. AI and DS algorithms offer critical benefits in effectively and efficiently analyzing microscopic images of blood cells. The automatic identification, localization, and classification of WBCs speed up the patient diagnosis process, allowing hematologists to focus on interpreting results. Automatic processes identify specific abnormalities and patterns, enhancing accuracy and timely diagnoses. Future work will examine the application of generative AI in blood cells diagnostics.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"583-603"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920980/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144087434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Histopathology-Based Prostate Cancer Classification Using ResNet: A Comprehensive Deep Learning Analysis. 基于组织病理学的前列腺癌分类使用ResNet:一个全面的深度学习分析。
Pub Date : 2026-02-01 Epub Date: 2025-05-20 DOI: 10.1007/s10278-025-01543-1
Declan Ikechukwu Emegano, Mubarak Taiwo Mustapha, Dilber Uzun Ozsahin, Ilker Ozsahin, Berna Uzun

Prostate cancer is the most prevalent solid tumor in males and one of the most common causes of male mortality. It is the most common type of cancer in men, a major global public health issue, and accounts for up to 7.3% of all male cancer diagnoses worldwide. To optimize patient outcomes and ensure therapeutic success, an accurate diagnosis must be made promptly. To achieve this, we focused on using ResNet50, a convolutional neural network (CNN) architecture, to analyze prostate histological images to classify prostate cancer. ResNet50, due to its efficiency in medical image classification, was used to classify the histological images as benign or malignant. In this study, a total of 1276 prostate biopsy images were used on the ResNet50 model. We employed evaluation metrics such as accuracy, precision, recall, and F1 score. The results showed that the ResNet50 model performed excellently with an overall accuracy of 0.98, 1.00 as precision, 0.98 as recall, and 0.97 as F1 score for benign. The malignant histological image has 0.99, 0.98, and 0.97 as precision, recall, and F1 scores. It also recorded a 95% confidence interval (CI) for accuracy as (0.91, 1.00) and a performance gain of 4.26% compared to MobileNet and CNN-RNN. The result of our model was also compared with the state-of-the-art (SOTA) DL models to ensure robustness. This study has demonstrated the potential of the ResNet50 model in the classification of prostate cancer. Again, the clinical integration of the results of this study will aid decision-makers in enhancing patient outcomes.

前列腺癌是男性中最常见的实体肿瘤,也是男性死亡的最常见原因之一。它是男性中最常见的癌症类型,是一个主要的全球公共卫生问题,占全球男性癌症诊断的7.3%。为了优化患者的治疗效果并确保治疗成功,必须及时做出准确的诊断。为了实现这一目标,我们专注于使用卷积神经网络(CNN)架构ResNet50来分析前列腺组织学图像以对前列腺癌进行分类。由于ResNet50在医学图像分类方面的有效性,我们使用ResNet50对组织学图像进行良性和恶性的分类。在本研究中,共有1276张前列腺活检图像用于ResNet50模型。我们采用了评估指标,如准确性、精密度、召回率和F1分数。结果表明,ResNet50模型表现优异,总体准确率为0.98,精密度为1.00,召回率为0.98,良性评分为0.97。恶性组织学图像的准确率、召回率和F1得分分别为0.99、0.98和0.97。它还记录了准确率的95%置信区间(CI)为(0.91,1.00),与MobileNet和CNN-RNN相比,性能提高了4.26%。我们的模型的结果也与最先进的(SOTA)深度学习模型进行了比较,以确保鲁棒性。本研究证明了ResNet50模型在前列腺癌分类中的潜力。再一次,本研究结果的临床整合将有助于决策者提高患者的治疗效果。
{"title":"Histopathology-Based Prostate Cancer Classification Using ResNet: A Comprehensive Deep Learning Analysis.","authors":"Declan Ikechukwu Emegano, Mubarak Taiwo Mustapha, Dilber Uzun Ozsahin, Ilker Ozsahin, Berna Uzun","doi":"10.1007/s10278-025-01543-1","DOIUrl":"10.1007/s10278-025-01543-1","url":null,"abstract":"<p><p>Prostate cancer is the most prevalent solid tumor in males and one of the most common causes of male mortality. It is the most common type of cancer in men, a major global public health issue, and accounts for up to 7.3% of all male cancer diagnoses worldwide. To optimize patient outcomes and ensure therapeutic success, an accurate diagnosis must be made promptly. To achieve this, we focused on using ResNet50, a convolutional neural network (CNN) architecture, to analyze prostate histological images to classify prostate cancer. ResNet50, due to its efficiency in medical image classification, was used to classify the histological images as benign or malignant. In this study, a total of 1276 prostate biopsy images were used on the ResNet50 model. We employed evaluation metrics such as accuracy, precision, recall, and F1 score. The results showed that the ResNet50 model performed excellently with an overall accuracy of 0.98, 1.00 as precision, 0.98 as recall, and 0.97 as F1 score for benign. The malignant histological image has 0.99, 0.98, and 0.97 as precision, recall, and F1 scores. It also recorded a 95% confidence interval (CI) for accuracy as (0.91, 1.00) and a performance gain of 4.26% compared to MobileNet and CNN-RNN. The result of our model was also compared with the state-of-the-art (SOTA) DL models to ensure robustness. This study has demonstrated the potential of the ResNet50 model in the classification of prostate cancer. Again, the clinical integration of the results of this study will aid decision-makers in enhancing patient outcomes.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"604-619"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921011/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144113203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Standardizing Heterogeneous MRI Series Description Metadata Using Large Language Models. 使用大型语言模型标准化异构MRI系列描述元数据。
Pub Date : 2026-02-01 Epub Date: 2025-05-29 DOI: 10.1007/s10278-025-01541-3
Peter I Kamel, Florence X Doo, Dharmam Savani, Adway Kanhere, Paul H Yi, Vishwa S Parekh

MRI metadata, particularly free-text series descriptions (SDs) used to identify sequences, are highly heterogeneous due to variable inputs by manufacturers and technologists. This variability poses challenges in correctly identifying series for hanging protocols and dataset curation. The purpose of this study was to evaluate the ability of large language models (LLMs) to automatically classify MRI SDs. We analyzed non-contrast brain MRIs performed between 2016 and 2022 at our institution, identifying all unique SDs in the metadata. A practicing neuroradiologist manually classified the SD text into: "T1," "T2," "T2/FLAIR," "SWI," "DWI," ADC," or "Other." Then, various LLMs, including GPT 3.5 Turbo, GPT-4, GPT-4o, Llama 3 8b, and Llama 3 70b, were asked to classify each SD into one of the sequence categories. Model performances were compared to ground truth classification using area under the curve (AUC) as the primary metric. Additionally, GPT-4o was tasked with generating regular expression templates to match each category. In 2510 MRI brain examinations, there were 1395 unique SDs, with 727/1395 (52.1%) appearing only once, indicating high variability. GPT-4o demonstrated the highest performance, achieving an average AUC of 0.983 ± 0.020 for all series with detailed prompting. GPT models significantly outperformed Llama models, with smaller differences within the GPT family. Regular expression generation was inconsistent, demonstrating an average AUC of 0.774 ± 0.161 for all sequences. Our findings suggest that LLMs are effective for interpreting and standardizing heterogeneous MRI SDs.

MRI元数据,特别是用于识别序列的自由文本序列描述(SDs),由于制造商和技术人员的不同输入,是高度异构的。这种可变性对正确识别悬挂协议和数据集管理的序列提出了挑战。本研究的目的是评估大型语言模型(LLMs)自动分类MRI SDs的能力。我们分析了2016年至2022年在我们机构进行的非对比脑mri,识别了元数据中所有独特的SDs。一位执业神经放射学家手动将SD文本分类为:“T1”、“T2”、“T2/FLAIR”、“SWI”、“DWI”、“ADC”或“其他”。然后,不同的LLMs,包括GPT 3.5 Turbo、GPT-4、GPT- 40、Llama 38 8b和Llama 3 70b,被要求将每个SD分类到一个序列类别中。将模型性能与以曲线下面积(AUC)为主要度量的地面真值分类进行比较。此外,gpt - 40的任务是生成正则表达式模板来匹配每个类别。2510例MRI脑部检查中,有1395个独特的SDs,其中727/1395(52.1%)只出现一次,显示出较高的变异性。gpt - 40表现出最高的性能,在详细提示的情况下,所有系列的平均AUC为0.983±0.020。GPT模型明显优于Llama模型,GPT家族内部的差异较小。正则表达式生成不一致,所有序列的平均AUC为0.774±0.161。我们的研究结果表明,llm对于解释和标准化异质MRI SDs是有效的。
{"title":"Standardizing Heterogeneous MRI Series Description Metadata Using Large Language Models.","authors":"Peter I Kamel, Florence X Doo, Dharmam Savani, Adway Kanhere, Paul H Yi, Vishwa S Parekh","doi":"10.1007/s10278-025-01541-3","DOIUrl":"10.1007/s10278-025-01541-3","url":null,"abstract":"<p><p>MRI metadata, particularly free-text series descriptions (SDs) used to identify sequences, are highly heterogeneous due to variable inputs by manufacturers and technologists. This variability poses challenges in correctly identifying series for hanging protocols and dataset curation. The purpose of this study was to evaluate the ability of large language models (LLMs) to automatically classify MRI SDs. We analyzed non-contrast brain MRIs performed between 2016 and 2022 at our institution, identifying all unique SDs in the metadata. A practicing neuroradiologist manually classified the SD text into: \"T1,\" \"T2,\" \"T2/FLAIR,\" \"SWI,\" \"DWI,\" ADC,\" or \"Other.\" Then, various LLMs, including GPT 3.5 Turbo, GPT-4, GPT-4o, Llama 3 8b, and Llama 3 70b, were asked to classify each SD into one of the sequence categories. Model performances were compared to ground truth classification using area under the curve (AUC) as the primary metric. Additionally, GPT-4o was tasked with generating regular expression templates to match each category. In 2510 MRI brain examinations, there were 1395 unique SDs, with 727/1395 (52.1%) appearing only once, indicating high variability. GPT-4o demonstrated the highest performance, achieving an average AUC of 0.983 ± 0.020 for all series with detailed prompting. GPT models significantly outperformed Llama models, with smaller differences within the GPT family. Regular expression generation was inconsistent, demonstrating an average AUC of 0.774 ± 0.161 for all sequences. Our findings suggest that LLMs are effective for interpreting and standardizing heterogeneous MRI SDs.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"962-972"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12921075/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144183157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Personalized Breast Cancer Prognosis Using a Model Based on MRI and Clinicopathological Variables. 基于MRI和临床病理变量的个性化乳腺癌预后模型。
Pub Date : 2026-02-01 Epub Date: 2025-04-15 DOI: 10.1007/s10278-025-01500-y
Alisa Mohebbi, Saeed Mohammadzadeh, Afshin Mohammadi, Seyed Mohammad Tavangar

This study aimed to develop and internally validate a prognostic prediction model based on MRI, pathological, and clinical findings to predict breast cancer recurrence and death. A retrospective study prediction model was developed using data from 922 breast cancer patients recruited in Duke University Hospital from January 2000 to March 2014. Cox and binary logistic regressions were implemented for hazard score and 2-, 3-, 5-, and 8-year survivals and recurrences. After assessing the collinearity of predictors, both univariable and multivariable analyses were performed. Qualitative and quantitative MRI variables were selected based on clinical expert opinion and literature review. Bootstrap and leave-one-out methods were used for internal validation. Calibration, shrinkage, time-dependent receiver operating characteristic (ROC) curve, and decision-curve analyses were also performed. Finally, a user-friendly calculator was built. Of included participants, 62 (6.72%) died with a mean patient-year follow-up of 8.89 years (CI = 8.74 to 9.04), while 90 (9.76%) experienced recurrence with mean patient-year follow-up of 8.20 years (CI = 7.92 to 8.48). The Akaike information criterion (AIC) value of survival and recurrence models were 752.9 and 1020.7, indicating a good balance between model complexity and fit. Validation model adjusted area under curve (AUC) in 8-, 5-, 3-, and 2-year survivals were 0.740 (CI = 0.711 to 0.768), 0.741 (CI = 0.712 to 0.770), 0.788 (CI = 0.761 to 0.816), and 0.783 (CI = 0.755 to 0.809), while in 8-, 5-, and 3-year recurrences were 0.678 (CI = 0.647 to 0.708), 0.696 (CI = 0.664 to 0.727), and 0.769 (CI = 0.740 to 0.798), respectively. Good calibration and shrinkage parameters were achieved. The internal validation and decision curve analyses highlighted the usefulness of the model across all probability levels. The combined MRI-pathological-clinical model has excellent performance in predicting overall survival and recurrence of breast cancer and may have a role to play in daily personalized breast cancer therapy.

本研究旨在建立并内部验证基于MRI,病理和临床结果的预后预测模型,以预测乳腺癌复发和死亡。利用2000年1月至2014年3月在杜克大学医院招募的922名乳腺癌患者的数据,建立了回顾性研究预测模型。对危险评分、2年、3年、5年和8年生存率和复发率进行Cox和二元logistic回归。在评估了预测因子的共线性后,进行了单变量和多变量分析。根据临床专家意见和文献回顾选择定性和定量MRI变量。Bootstrap和leave- out方法用于内部验证。还进行了校准、收缩、随时间变化的受试者工作特征(ROC)曲线和决策曲线分析。最后,构建了一个用户友好的计算器。在纳入的参与者中,62例(6.72%)死亡,平均患者年随访8.89年(CI = 8.74至9.04),而90例(9.76%)复发,平均患者年随访8.20年(CI = 7.92至8.48)。生存模型和复发模型的赤池信息准则(Akaike information criterion, AIC)值分别为752.9和1020.7,表明模型复杂度与拟合之间取得了较好的平衡。验证模型校正曲线下面积(AUC)在8年、5年、3年和2年生存率中分别为0.740 (CI = 0.711 ~ 0.768)、0.741 (CI = 0.712 ~ 0.770)、0.788 (CI = 0.761 ~ 0.816)和0.783 (CI = 0.755 ~ 0.809),而在8年、5年和3年生存率中分别为0.678 (CI = 0.647 ~ 0.708)、0.696 (CI = 0.664 ~ 0.727)和0.769 (CI = 0.740 ~ 0.798)。获得了良好的校准和收缩参数。内部验证和决策曲线分析强调了模型在所有概率水平上的有用性。mri -病理-临床联合模型在预测乳腺癌的总生存期和复发方面具有优异的性能,可能在乳腺癌的日常个性化治疗中发挥作用。
{"title":"Personalized Breast Cancer Prognosis Using a Model Based on MRI and Clinicopathological Variables.","authors":"Alisa Mohebbi, Saeed Mohammadzadeh, Afshin Mohammadi, Seyed Mohammad Tavangar","doi":"10.1007/s10278-025-01500-y","DOIUrl":"10.1007/s10278-025-01500-y","url":null,"abstract":"<p><p>This study aimed to develop and internally validate a prognostic prediction model based on MRI, pathological, and clinical findings to predict breast cancer recurrence and death. A retrospective study prediction model was developed using data from 922 breast cancer patients recruited in Duke University Hospital from January 2000 to March 2014. Cox and binary logistic regressions were implemented for hazard score and 2-, 3-, 5-, and 8-year survivals and recurrences. After assessing the collinearity of predictors, both univariable and multivariable analyses were performed. Qualitative and quantitative MRI variables were selected based on clinical expert opinion and literature review. Bootstrap and leave-one-out methods were used for internal validation. Calibration, shrinkage, time-dependent receiver operating characteristic (ROC) curve, and decision-curve analyses were also performed. Finally, a user-friendly calculator was built. Of included participants, 62 (6.72%) died with a mean patient-year follow-up of 8.89 years (CI = 8.74 to 9.04), while 90 (9.76%) experienced recurrence with mean patient-year follow-up of 8.20 years (CI = 7.92 to 8.48). The Akaike information criterion (AIC) value of survival and recurrence models were 752.9 and 1020.7, indicating a good balance between model complexity and fit. Validation model adjusted area under curve (AUC) in 8-, 5-, 3-, and 2-year survivals were 0.740 (CI = 0.711 to 0.768), 0.741 (CI = 0.712 to 0.770), 0.788 (CI = 0.761 to 0.816), and 0.783 (CI = 0.755 to 0.809), while in 8-, 5-, and 3-year recurrences were 0.678 (CI = 0.647 to 0.708), 0.696 (CI = 0.664 to 0.727), and 0.769 (CI = 0.740 to 0.798), respectively. Good calibration and shrinkage parameters were achieved. The internal validation and decision curve analyses highlighted the usefulness of the model across all probability levels. The combined MRI-pathological-clinical model has excellent performance in predicting overall survival and recurrence of breast cancer and may have a role to play in daily personalized breast cancer therapy.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":"500-517"},"PeriodicalIF":0.0,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12920835/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144003568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of imaging informatics in medicine
全部 Geobiology Appl. Clay Sci. Geochim. Cosmochim. Acta J. Hydrol. Org. Geochem. Carbon Balance Manage. Contrib. Mineral. Petrol. Int. J. Biometeorol. IZV-PHYS SOLID EART+ J. Atmos. Chem. Acta Oceanolog. Sin. Acta Geophys. ACTA GEOL POL ACTA PETROL SIN ACTA GEOL SIN-ENGL AAPG Bull. Acta Geochimica Adv. Atmos. Sci. Adv. Meteorol. Am. J. Phys. Anthropol. Am. J. Sci. Am. Mineral. Annu. Rev. Earth Planet. Sci. Appl. Geochem. Aquat. Geochem. Ann. Glaciol. Archaeol. Anthropol. Sci. ARCHAEOMETRY ARCT ANTARCT ALP RES Asia-Pac. J. Atmos. Sci. ATMOSPHERE-BASEL Atmos. Res. Aust. J. Earth Sci. Atmos. Chem. Phys. Atmos. Meas. Tech. Basin Res. Big Earth Data BIOGEOSCIENCES Geostand. Geoanal. Res. GEOLOGY Geosci. J. Geochem. J. Geochem. Trans. Geosci. Front. Geol. Ore Deposits Global Biogeochem. Cycles Gondwana Res. Geochem. Int. Geol. J. Geophys. Prospect. Geosci. Model Dev. GEOL BELG GROUNDWATER Hydrogeol. J. Hydrol. Earth Syst. Sci. Hydrol. Processes Int. J. Climatol. Int. J. Earth Sci. Int. Geol. Rev. Int. J. Disaster Risk Reduct. Int. J. Geomech. Int. J. Geog. Inf. Sci. Isl. Arc J. Afr. Earth. Sci. J. Adv. Model. Earth Syst. J APPL METEOROL CLIM J. Atmos. Oceanic Technol. J. Atmos. Sol. Terr. Phys. J. Clim. J. Earth Sci. J. Earth Syst. Sci. J. Environ. Eng. Geophys. J. Geog. Sci. Mineral. Mag. Miner. Deposita Mon. Weather Rev. Nat. Hazards Earth Syst. Sci. Nat. Clim. Change Nat. Geosci. Ocean Dyn. Ocean and Coastal Research npj Clim. Atmos. Sci. Ocean Modell. Ocean Sci. Ore Geol. Rev. OCEAN SCI J Paleontol. J. PALAEOGEOGR PALAEOCL PERIOD MINERAL PETROLOGY+ Phys. Chem. Miner. Polar Sci. Prog. Oceanogr. Quat. Sci. Rev. Q. J. Eng. Geol. Hydrogeol. RADIOCARBON Pure Appl. Geophys. Resour. Geol. Rev. Geophys. Sediment. Geol.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1