首页 > 最新文献

Radiology-Artificial Intelligence最新文献

英文 中文
One System to Rule Them All? Task- and Data-specific Considerations for Automated Data Extraction. 用一个系统来统治所有人?自动数据提取的任务和数据特定考虑因素。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.250175
Ali S Tejani, Andreas M Rauschecker
{"title":"One System to Rule Them All? Task- and Data-specific Considerations for Automated Data Extraction.","authors":"Ali S Tejani, Andreas M Rauschecker","doi":"10.1148/ryai.250175","DOIUrl":"10.1148/ryai.250175","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 3","pages":"e250175"},"PeriodicalIF":13.2,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144018982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Seeing the Unseen: How Unsupervised Learning Can Predict Genetic Mutations from Radiologic Images. 看到看不见的:无监督学习如何从放射图像预测基因突变。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.250243
Eduardo Moreno Júdice de Mattos Farina, Paulo Eduardo de Aguiar Kuriki
{"title":"Seeing the Unseen: How Unsupervised Learning Can Predict Genetic Mutations from Radiologic Images.","authors":"Eduardo Moreno Júdice de Mattos Farina, Paulo Eduardo de Aguiar Kuriki","doi":"10.1148/ryai.250243","DOIUrl":"10.1148/ryai.250243","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 3","pages":"e250243"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144001837","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Dual-Task Deep Learning for Automated Thyroid Cancer Triaging at Screening US. 自适应双任务深度学习用于筛查美国甲状腺癌的自动分类。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.240271
Shao-Hong Wu, Ming-De Li, Wen-Juan Tong, Yi-Hao Liu, Rui Cui, Jin-Bo Hu, Mei-Qing Cheng, Wei-Ping Ke, Xin-Xin Lin, Jia-Yi Lv, Long-Zhong Liu, Jie Ren, Guang-Jian Liu, Hong Yang, Wei Wang

Purpose To develop an adaptive dual-task deep learning model (ThyNet-S) for detection and classification of thyroid lesions at US screening. Materials and Methods This retrospective study used a multicenter dataset comprising 35 008 thyroid US images of 23 294 individual examinations (mean age, 40.4 years ± 13.1 [SD]; 17 587 female) from seven medical centers from January 2009 to December 2021. Of these, 29 004 images were used for model development and 6004 images for validation. The model determined cancer risk for each image and automatically triaged images with normal thyroid and benign nodules by dynamically integrating lesion detection through pixel-level feature analysis and lesion classification through deep semantic features analysis. Diagnostic performance of screening assisted by the model (ThyNet-S triaged screening) and traditional screening (radiologists alone) was assessed by comparing sensitivity, specificity, accuracy, and area under the receiver operating characteristic curve using the McNemar test and DeLong test. The influence of ThyNet-S on radiologist workload and clinical decision-making was also assessed. Results ThyNet-S-assisted triaged screening achieved a higher area under the receiver operating characteristic curve than original screening with six senior and six junior radiologists (0.93 vs 0.91 and 0.92 vs 0.88, respectively; all P < .001). The model improved sensitivity for junior radiologists (88.2% vs 86.8%; P < .001). Notably, the model reduced radiologists' workload by triaging 60.4% of cases as not potentially malignant, which did not require further interpretation. The model simultaneously decreased the unnecessary fine needle aspiration rate from 38.7% to 14.9% and 11.5% when used independently or in combination with the Thyroid Imaging Reporting and Data System, respectively. Conclusion ThyNet-S improved the efficiency of thyroid cancer screening and optimized clinical decision-making. Keywords: Artificial Intelligence, Adaptive, Dual Task, Thyroid Cancer, Screening, Ultrasound Supplemental material is available for this article. © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的建立一种自适应双任务深度学习模型(ThyNet-S),用于超声筛查中甲状腺病变的检测和分类。材料与方法回顾性研究采用多中心数据集,包括2009年1月至2021年12月来自7个医疗中心的23294例个体检查(平均年龄40.4岁±13.1[SD], 17587例女性)的35008张甲状腺超声图像。其中,29004张图像用于模型开发,6004张图像用于验证。该模型通过像素级特征分析动态整合病变检测和深度语义特征分析的病变分类,确定每张图像的癌症风险,并自动对正常甲状腺结节和良性结节进行分类。通过比较McNemar试验和Delong试验的敏感性、特异性、准确性和AUC,评估模型辅助筛查(ThyNet-S分级筛查)和传统筛查(放射科医师单独筛查)的诊断性能。还评估了ThyNet-S对放射科医生工作量和临床决策的影响。结果6名高级放射科医师和6名初级放射科医师的thynet - s辅助分类筛查的AUC高于原始筛查(分别为0.93比0.91,0.92比0.88,P均< 0.001)。该模型提高了初级放射科医生的敏感性(88.2%对86.8%,P
{"title":"Adaptive Dual-Task Deep Learning for Automated Thyroid Cancer Triaging at Screening US.","authors":"Shao-Hong Wu, Ming-De Li, Wen-Juan Tong, Yi-Hao Liu, Rui Cui, Jin-Bo Hu, Mei-Qing Cheng, Wei-Ping Ke, Xin-Xin Lin, Jia-Yi Lv, Long-Zhong Liu, Jie Ren, Guang-Jian Liu, Hong Yang, Wei Wang","doi":"10.1148/ryai.240271","DOIUrl":"10.1148/ryai.240271","url":null,"abstract":"<p><p>Purpose To develop an adaptive dual-task deep learning model (ThyNet-S) for detection and classification of thyroid lesions at US screening. Materials and Methods This retrospective study used a multicenter dataset comprising 35 008 thyroid US images of 23 294 individual examinations (mean age, 40.4 years ± 13.1 [SD]; 17 587 female) from seven medical centers from January 2009 to December 2021. Of these, 29 004 images were used for model development and 6004 images for validation. The model determined cancer risk for each image and automatically triaged images with normal thyroid and benign nodules by dynamically integrating lesion detection through pixel-level feature analysis and lesion classification through deep semantic features analysis. Diagnostic performance of screening assisted by the model (ThyNet-S triaged screening) and traditional screening (radiologists alone) was assessed by comparing sensitivity, specificity, accuracy, and area under the receiver operating characteristic curve using the McNemar test and DeLong test. The influence of ThyNet-S on radiologist workload and clinical decision-making was also assessed. Results ThyNet-S-assisted triaged screening achieved a higher area under the receiver operating characteristic curve than original screening with six senior and six junior radiologists (0.93 vs 0.91 and 0.92 vs 0.88, respectively; all <i>P</i> < .001). The model improved sensitivity for junior radiologists (88.2% vs 86.8%; <i>P</i> < .001). Notably, the model reduced radiologists' workload by triaging 60.4% of cases as not potentially malignant, which did not require further interpretation. The model simultaneously decreased the unnecessary fine needle aspiration rate from 38.7% to 14.9% and 11.5% when used independently or in combination with the Thyroid Imaging Reporting and Data System, respectively. Conclusion ThyNet-S improved the efficiency of thyroid cancer screening and optimized clinical decision-making. <b>Keywords:</b> Artificial Intelligence, Adaptive, Dual Task, Thyroid Cancer, Screening, Ultrasound <i>Supplemental material is available for this article.</i> © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240271"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143812639","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pixels to Prognosis: Using Deep Learning to Rethink Cardiac Risk Prediction from CT Angiography. 像素到预后:利用深度学习重新思考CT血管造影的心脏风险预测。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.250260
Rohit Reddy
{"title":"Pixels to Prognosis: Using Deep Learning to Rethink Cardiac Risk Prediction from CT Angiography.","authors":"Rohit Reddy","doi":"10.1148/ryai.250260","DOIUrl":"10.1148/ryai.250260","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 3","pages":"e250260"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144162461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Quantification of Serial PET/CT Images for Pediatric Hodgkin Lymphoma Using a Longitudinally Aware Segmentation Network. 基于纵向感知分割网络的儿童霍奇金淋巴瘤系列PET/CT图像自动量化。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.240229
Xin Tie, Muheon Shin, Changhee Lee, Scott B Perlman, Zachary Huemann, Amy J Weisman, Sharon M Castellino, Kara M Kelly, Kathleen M McCarten, Adina L Alazraki, Junjie Hu, Steve Y Cho, Tyler J Bradshaw

Purpose To develop a longitudinally aware segmentation network (LAS-Net) that can quantify serial PET/CT images for pediatric patients with Hodgkin lymphoma. Materials and Methods This retrospective study included baseline (PET1) and interim (PET2) PET/CT images from 297 pediatric patients enrolled in two Children's Oncology Group clinical trials (AHOD1331 and AHOD0831). The internal dataset included 200 patients (enrolled between March 2015 and August 2019; median age, 15.4 years [range, 5.6-22.0 years]; 107 male), and the external testing dataset included 97 patients (enrolled between December 2009 and January 2012; median age, 15.8 years [range, 5.2-21.4 years]; 59 male). LAS-Net incorporates longitudinal cross-attention, allowing relevant features from PET1 to inform the analysis of PET2. The model's lesion segmentation performance on PET1 images was evaluated using Dice coefficients, and lesion detection performance on PET2 images was evaluated using F1 scores. In addition, quantitative PET metrics, including metabolic tumor volume (MTV) and total lesion glycolysis (TLG) in PET1, as well as qPET and percentage difference between baseline and interim maximum standardized uptake value (∆SUVmax) in PET2, were extracted and compared against physician-derived measurements. Agreement between model and physician-derived measurements was quantified using Spearman correlation, and bootstrap resampling was used for statistical analysis. Results LAS-Net detected residual lymphoma on PET2 scans with an F1 score of 0.61 (precision/recall: 0.62/0.60), outperforming all comparator methods (P < .01). For baseline segmentation, LAS-Net achieved a mean Dice score of 0.77. In PET quantification, LAS-Net's measurements of qPET, ∆SUVmax, MTV, and TLG were strongly correlated with physician measurements, with Spearman ρ values of 0.78, 0.80, 0.93, and 0.96, respectively. The quantification performance remained high, with a slight decrease, in an external testing cohort. Conclusion LAS-Net demonstrated significant improvements in quantifying PET metrics across serial scans in pediatric patients with Hodgkin lymphoma, highlighting the value of longitudinal awareness in evaluating multi-time-point imaging datasets. Keywords: Pediatrics, PET/CT, Lymphoma, Segmentation, Quantification, Supervised Learning, Convolutional Neural Network (CNN), Quantitative PET, Longitudinal Analysis, Deep Learning, Image Segmentation Supplemental material is available for this article. Clinical trial registration no. NCT02166463 and NCT01026220 © RSNA, 2025 See also commentary by Khosravi and Gichoya in this issue.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的建立纵向感知的分割网络(LAS-Net),用于量化儿童霍奇金淋巴瘤患者的连续PET/CT图像。材料和方法本回顾性研究纳入了297名儿童肿瘤组临床试验(AHOD1331和AHOD0831)的儿童患者的基线(PET1)和中期(PET2) PET/CT图像。内部数据集包括200名患者(2015年3月至2019年8月;中位年龄15.4岁[IQR: 5.6, 22.0]岁;外部测试数据集包括97例患者(2009年12月至2012年1月入组;中位年龄15.8岁[IQR: 5.2, 21.4]岁;59岁男性)。LAS-Net结合了纵向交叉注意,允许PET1的相关特征为PET2的分析提供信息。使用Dice系数评价模型对PET1图像的病灶分割性能,使用F1分数评价模型对PET2图像的病灶检测性能。此外,提取定量PET指标,包括PET1中的代谢肿瘤体积(MTV)和总病变糖酵解(TLG),以及PET2中的qPET和∆SUVmax,并与医生提供的测量结果进行比较。采用Spearman相关对模型和医生测量结果之间的一致性进行量化,并采用自举重采样进行统计分析。结果LAS-Net在PET2扫描中检测到残留淋巴瘤,F1评分为0.61(精密度/召回率:0.62/0.60),优于所有比较方法(P < 0.01)。对于基线分割,LAS-Net的平均Dice得分为0.77。在PET定量中,LAS-Net测量的qPET、∆SUVmax、MTV和TLG与医生测量值密切相关,Spearman的ρ值分别为0.78、0.80、0.93和0.96。在外部测试队列中,量化表现仍然很高,略有下降。结论:LAS-Net在量化霍奇金淋巴瘤儿童患者串行扫描的PET指标方面有显著改善,突出了纵向感知在评估多时间点成像数据集中的价值。©RSNA, 2025年。
{"title":"Automatic Quantification of Serial PET/CT Images for Pediatric Hodgkin Lymphoma Using a Longitudinally Aware Segmentation Network.","authors":"Xin Tie, Muheon Shin, Changhee Lee, Scott B Perlman, Zachary Huemann, Amy J Weisman, Sharon M Castellino, Kara M Kelly, Kathleen M McCarten, Adina L Alazraki, Junjie Hu, Steve Y Cho, Tyler J Bradshaw","doi":"10.1148/ryai.240229","DOIUrl":"10.1148/ryai.240229","url":null,"abstract":"<p><p>Purpose To develop a longitudinally aware segmentation network (LAS-Net) that can quantify serial PET/CT images for pediatric patients with Hodgkin lymphoma. Materials and Methods This retrospective study included baseline (PET1) and interim (PET2) PET/CT images from 297 pediatric patients enrolled in two Children's Oncology Group clinical trials (AHOD1331 and AHOD0831). The internal dataset included 200 patients (enrolled between March 2015 and August 2019; median age, 15.4 years [range, 5.6-22.0 years]; 107 male), and the external testing dataset included 97 patients (enrolled between December 2009 and January 2012; median age, 15.8 years [range, 5.2-21.4 years]; 59 male). LAS-Net incorporates longitudinal cross-attention, allowing relevant features from PET1 to inform the analysis of PET2. The model's lesion segmentation performance on PET1 images was evaluated using Dice coefficients, and lesion detection performance on PET2 images was evaluated using F1 scores. In addition, quantitative PET metrics, including metabolic tumor volume (MTV) and total lesion glycolysis (TLG) in PET1, as well as qPET and percentage difference between baseline and interim maximum standardized uptake value (∆SUV<sub>max</sub>) in PET2, were extracted and compared against physician-derived measurements. Agreement between model and physician-derived measurements was quantified using Spearman correlation, and bootstrap resampling was used for statistical analysis. Results LAS-Net detected residual lymphoma on PET2 scans with an F1 score of 0.61 (precision/recall: 0.62/0.60), outperforming all comparator methods (<i>P</i> < .01). For baseline segmentation, LAS-Net achieved a mean Dice score of 0.77. In PET quantification, LAS-Net's measurements of qPET, ∆SUV<sub>max</sub>, MTV, and TLG were strongly correlated with physician measurements, with Spearman ρ values of 0.78, 0.80, 0.93, and 0.96, respectively. The quantification performance remained high, with a slight decrease, in an external testing cohort. Conclusion LAS-Net demonstrated significant improvements in quantifying PET metrics across serial scans in pediatric patients with Hodgkin lymphoma, highlighting the value of longitudinal awareness in evaluating multi-time-point imaging datasets. <b>Keywords:</b> Pediatrics, PET/CT, Lymphoma, Segmentation, Quantification, Supervised Learning, Convolutional Neural Network (CNN), Quantitative PET, Longitudinal Analysis, Deep Learning, Image Segmentation <i>Supplemental material is available for this article.</i> Clinical trial registration no. NCT02166463 and NCT01026220 © RSNA, 2025 See also commentary by Khosravi and Gichoya in this issue.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240229"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12127956/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143450322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Pipeline for Automated Quality Control of Chest Radiographs. 胸片自动质量控制的流水线。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.240003
Ian A Selby, Eduardo González Solares, Anna Breger, Michael Roberts, Lorena Escudero Sánchez, Judith Babar, James H F Rudd, Nicholas A Walton, Evis Sala, Carola-Bibiane Schönlieb, Jonathan R Weir-McCall
{"title":"A Pipeline for Automated Quality Control of Chest Radiographs.","authors":"Ian A Selby, Eduardo González Solares, Anna Breger, Michael Roberts, Lorena Escudero Sánchez, Judith Babar, James H F Rudd, Nicholas A Walton, Evis Sala, Carola-Bibiane Schönlieb, Jonathan R Weir-McCall","doi":"10.1148/ryai.240003","DOIUrl":"10.1148/ryai.240003","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240003"},"PeriodicalIF":13.2,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12127945/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143558237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing Large Language Models with Retrieval-Augmented Generation: A Radiology-Specific Approach. 用检索增强生成增强大型语言模型:一种放射学专用方法。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.240313
Dane A Weinert, Andreas M Rauschecker

Retrieval-augmented generation (RAG) is a strategy to improve the performance of large language models (LLMs) by providing an LLM with an updated corpus of knowledge that can be used for answer generation in real time. RAG may improve LLM performance and clinical applicability in radiology by providing citable, up-to-date information without requiring model fine-tuning. In this retrospective study, a radiology-specific RAG system was developed using a vector database of 3689 RadioGraphics articles published from January 1999 to December 2023. Performance of five LLMs with (RAG-Systems) and without RAG on a 192-question radiology examination was compared. RAG significantly improved examination scores for GPT-4 (OpenAI; 81.2% vs 75.5%, P = .04) and Command R+ (Cohere; 70.3% vs 62.0%, P = .02), but not for Claude Opus (Anthropic), Mixtral (Mistral AI), or Gemini 1.5 Pro (Google DeepMind). RAG-Systems performed significantly better than pure LLMs on a 24-question subset directly sourced from RadioGraphics (85% vs 76%, P = .03). RAG-Systems retrieved 21 of 24 (87.5%, P < .001) relevant RadioGraphics references cited in the examination's answer explanations and successfully cited them in 18 of 21 (85.7%, P < .001) outputs. The results suggest that RAG is a promising approach to enhance LLM capabilities for radiology knowledge tasks, providing transparent, domain-specific information retrieval. Keywords: Computer Applications-General (Informatics), Technology Assessment Supplemental material is available for this article. © RSNA, 2025 See also commentary by Mansuri and Gichoya in this issue.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。检索增强生成(RAG)是一种提高大型语言模型(LLM)性能的策略,它为LLM提供可用于实时生成答案的更新知识语料库。RAG可以通过提供可引用的、最新的信息而不需要模型微调来提高LLM在放射学中的性能和临床适用性。在这项回顾性研究中,利用1999年1月至2023年12月发表的3,689篇放射学文章的矢量数据库开发了放射学特异性RAG。我们比较了5例有和没有RAG的LLMs在192个问题的放射学检查中的表现。RAG显著提高了GPT-4(81.2%对75.5%,P = .04)和Command R+(70.3%对62.0%,P = .02)的考试成绩,但对Claude Opus、Mixtral或Gemini 1.5 Pro没有显著提高。在直接来自RadioGraphics的24个问题子集上,RAG-System的表现明显优于纯LLMs(85%对76%,P = .03)。ragg系统检索了考试答案解释中引用的21/24 (87.5%,P < .001)篇相关放射学文献,并在18/21 (85.7%,P < .001)篇输出中成功引用了这些文献。结果表明,RAG是一种很有前途的方法,可以提高LLM在放射学知识任务中的能力,提供透明的、特定领域的信息检索。©RSNA, 2025年。
{"title":"Enhancing Large Language Models with Retrieval-Augmented Generation: A Radiology-Specific Approach.","authors":"Dane A Weinert, Andreas M Rauschecker","doi":"10.1148/ryai.240313","DOIUrl":"10.1148/ryai.240313","url":null,"abstract":"<p><p>Retrieval-augmented generation (RAG) is a strategy to improve the performance of large language models (LLMs) by providing an LLM with an updated corpus of knowledge that can be used for answer generation in real time. RAG may improve LLM performance and clinical applicability in radiology by providing citable, up-to-date information without requiring model fine-tuning. In this retrospective study, a radiology-specific RAG system was developed using a vector database of 3689 <i>RadioGraphics</i> articles published from January 1999 to December 2023. Performance of five LLMs with (RAG-Systems) and without RAG on a 192-question radiology examination was compared. RAG significantly improved examination scores for GPT-4 (OpenAI; 81.2% vs 75.5%, <i>P</i> = .04) and Command R+ (Cohere; 70.3% vs 62.0%, <i>P</i> = .02), but not for Claude Opus (Anthropic), Mixtral (Mistral AI), or Gemini 1.5 Pro (Google DeepMind). RAG-Systems performed significantly better than pure LLMs on a 24-question subset directly sourced from <i>RadioGraphics</i> (85% vs 76%, <i>P</i> = .03). RAG-Systems retrieved 21 of 24 (87.5%, <i>P</i> < .001) relevant <i>RadioGraphics</i> references cited in the examination's answer explanations and successfully cited them in 18 of 21 (85.7%, <i>P</i> < .001) outputs. The results suggest that RAG is a promising approach to enhance LLM capabilities for radiology knowledge tasks, providing transparent, domain-specific information retrieval. <b>Keywords:</b> Computer Applications-General (Informatics), Technology Assessment <i>Supplemental material is available for this article.</i> © RSNA, 2025 See also commentary by Mansuri and Gichoya in this issue.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240313"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143606543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Context Is Everything: Understanding Variable LLM Performance for Radiology Retrieval-Augmented Generation. 背景决定一切:理解放射学检索增强生成的可变LLM性能。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.250187
Aawez Mansuri, Judy W Gichoya
{"title":"Context Is Everything: Understanding Variable LLM Performance for Radiology Retrieval-Augmented Generation.","authors":"Aawez Mansuri, Judy W Gichoya","doi":"10.1148/ryai.250187","DOIUrl":"10.1148/ryai.250187","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 3","pages":"e250187"},"PeriodicalIF":13.2,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12127949/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144036799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating Performance of a Deep Learning Multilabel Segmentation Model to Quantify Acute and Chronic Brain Lesions at MRI after Stroke and Predict Prognosis. 评估深度学习多标签分割模型在脑卒中后MRI上量化急慢性脑损伤和预测预后的性能。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.240072
Tianyu Tang, Ying Cui, Chunqiang Lu, Huiming Li, Jiaying Zhou, Xiaoyu Zhang, Yujie Zhou, Ying Zhang, Yi Zhang, Yuhao Xu, Yuefeng Li, Shenghong Ju

Purpose To develop and evaluate a multilabel deep learning network to identify and quantify acute and chronic brain lesions at multisequence MRI after acute ischemic stroke (AIS) and assess relationships between clinical and model-extracted radiologic features of the lesions and patient prognosis. Materials and Methods This retrospective study included patients with AIS from multiple centers, who experienced stroke onset between September 2008 and October 2022 and underwent MRI as well as thrombolytic therapy and/or treatment with antiplatelets or anticoagulants. A SegResNet-based deep learning model was developed to segment core infarcts and white matter hyperintensity (WMH) burdens on diffusion-weighted and fluid-attenuated inversion recovery images. The model was trained, validated, and tested with manual labels (260, 60, and 40 patients in each dataset, respectively). Radiologic features extracted from the model, including regional infarct size and periventricular and deep WMH volumes and cluster numbers, combined with clinical variables, were used to predict favorable versus unfavorable patient outcomes at 7 days (modified Rankin Scale [mRS] score). Mediation analyses explored associations between radiologic features and AIS outcomes within different treatment groups. Results A total of 1008 patients (mean age, 67.0 years ± 11.8 [SD]; 686 male, 322 female) were included. The training and validation dataset comprised 702 patients with AIS, and the two external testing datasets included 206 and 100 patients, respectively. The prognostic model combining clinical and radiologic features achieved areas under the receiver operating characteristic curve of 0.81 (95% CI: 0.74, 0.88) and 0.77 (95% CI: 0.68, 0.86) for predicting 7-day outcomes in the two external testing datasets, respectively. Mediation analyses revealed that deep WMH in patients treated with thrombolysis had a significant direct effect (17.7%, P = .01) and indirect effect (10.7%, P = .01) on unfavorable outcomes, as indicated by higher mRS scores, which was not observed in patients treated with antiplatelets and/or anticoagulants. Conclusion The proposed deep learning model quantitatively analyzed radiologic features of acute and chronic brain lesions, and the extracted radiologic features combined with clinical variables predicted short-term AIS outcomes. WMH burden, particularly deep WMH, emerged as a risk factor for poor outcomes in patients treated with thrombolysis. Keywords: MR-Diffusion Weighted Imaging, Thrombolysis, Head/Neck, Brain/Brain Stem, Stroke, Outcomes Analysis, Segmentation, Prognosis, Supervised Learning, Convolutional Neural Network (CNN), Support Vector Machines Supplemental material is available for this article. © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的开发和评估一个多标签深度学习网络,用于在急性缺血性卒中(AIS)后的多序列MRI上识别和量化急性和慢性脑病变,并评估病变的临床和模型提取的放射学特征与患者预后之间的关系。材料和方法本回顾性研究纳入了来自多个中心的AIS患者(2008年9月至2022年10月),这些患者接受了MRI和溶栓或抗血小板和/或抗凝治疗。开发了基于segresnet的深度学习模型,用于在弥散加权成像和流体衰减反演恢复图像上分割核心梗死和白质高强度(WMH)负担。该模型使用手动标签进行训练、验证和测试(每个数据集中分别有n = 260、60和40名患者)。从模型中提取的放射学特征,包括局部梗死面积、心室周围和深部WMH体积和簇数,结合临床变量,用于预测患者7天的有利和不利结果(改良Rankin量表[mRS]评分)。中介分析探讨了不同治疗组放射学特征与AIS结果之间的关系。结果共1008例患者(平均年龄67.0±11.8岁;男性686例,女性322例)。训练和验证数据集包括702例AIS患者,两个外部测试数据集分别包括206例和100例患者。结合临床和放射学特征的预后模型在两个外部测试数据集中预测7天预后的auc分别为0.81 (95% CI: 0.74-0.88)和0.77 (95% CI: 0.68-0.86)。中介分析显示,接受溶栓治疗的患者的深度WMH对不良结局有显著的直接影响(17.7%,P = 0.01)和间接影响(10.7%,P = 0.01),这表明mRS评分较高,而在接受抗血小板和/或抗凝药物治疗的患者中没有观察到这一点。结论提出的深度学习模型定量分析急慢性脑病变的影像学特征,并结合临床变量提取影像学特征预测AIS的短期预后。WMH负担,特别是深度WMH,已成为溶栓治疗患者预后不良的一个危险因素。©RSNA, 2025年。
{"title":"Evaluating Performance of a Deep Learning Multilabel Segmentation Model to Quantify Acute and Chronic Brain Lesions at MRI after Stroke and Predict Prognosis.","authors":"Tianyu Tang, Ying Cui, Chunqiang Lu, Huiming Li, Jiaying Zhou, Xiaoyu Zhang, Yujie Zhou, Ying Zhang, Yi Zhang, Yuhao Xu, Yuefeng Li, Shenghong Ju","doi":"10.1148/ryai.240072","DOIUrl":"10.1148/ryai.240072","url":null,"abstract":"<p><p>Purpose To develop and evaluate a multilabel deep learning network to identify and quantify acute and chronic brain lesions at multisequence MRI after acute ischemic stroke (AIS) and assess relationships between clinical and model-extracted radiologic features of the lesions and patient prognosis. Materials and Methods This retrospective study included patients with AIS from multiple centers, who experienced stroke onset between September 2008 and October 2022 and underwent MRI as well as thrombolytic therapy and/or treatment with antiplatelets or anticoagulants. A SegResNet-based deep learning model was developed to segment core infarcts and white matter hyperintensity (WMH) burdens on diffusion-weighted and fluid-attenuated inversion recovery images. The model was trained, validated, and tested with manual labels (260, 60, and 40 patients in each dataset, respectively). Radiologic features extracted from the model, including regional infarct size and periventricular and deep WMH volumes and cluster numbers, combined with clinical variables, were used to predict favorable versus unfavorable patient outcomes at 7 days (modified Rankin Scale [mRS] score). Mediation analyses explored associations between radiologic features and AIS outcomes within different treatment groups. Results A total of 1008 patients (mean age, 67.0 years ± 11.8 [SD]; 686 male, 322 female) were included. The training and validation dataset comprised 702 patients with AIS, and the two external testing datasets included 206 and 100 patients, respectively. The prognostic model combining clinical and radiologic features achieved areas under the receiver operating characteristic curve of 0.81 (95% CI: 0.74, 0.88) and 0.77 (95% CI: 0.68, 0.86) for predicting 7-day outcomes in the two external testing datasets, respectively. Mediation analyses revealed that deep WMH in patients treated with thrombolysis had a significant direct effect (17.7%, <i>P</i> = .01) and indirect effect (10.7%, <i>P</i> = .01) on unfavorable outcomes, as indicated by higher mRS scores, which was not observed in patients treated with antiplatelets and/or anticoagulants. Conclusion The proposed deep learning model quantitatively analyzed radiologic features of acute and chronic brain lesions, and the extracted radiologic features combined with clinical variables predicted short-term AIS outcomes. WMH burden, particularly deep WMH, emerged as a risk factor for poor outcomes in patients treated with thrombolysis. <b>Keywords:</b> MR-Diffusion Weighted Imaging, Thrombolysis, Head/Neck, Brain/Brain Stem, Stroke, Outcomes Analysis, Segmentation, Prognosis, Supervised Learning, Convolutional Neural Network (CNN), Support Vector Machines <i>Supplemental material is available for this article.</i> © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240072"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143711426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Establishing a Chain of Evidence for AI in Radiology: Sham AI and Randomized Controlled Trials. 建立人工智能在放射学中的证据链:假人工智能和随机对照试验。
IF 8.1 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-05-01 DOI: 10.1148/ryai.250334
John D Mayfield, Javier Romero
{"title":"Establishing a Chain of Evidence for AI in Radiology: Sham AI and Randomized Controlled Trials.","authors":"John D Mayfield, Javier Romero","doi":"10.1148/ryai.250334","DOIUrl":"10.1148/ryai.250334","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 3","pages":"e250334"},"PeriodicalIF":8.1,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144162292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Radiology-Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1