首页 > 最新文献

Radiology-Artificial Intelligence最新文献

英文 中文
Unlocking Robust Segmentation: Decoding Domain Randomization for Radiologists. 解锁鲁棒分割:解码领域随机化放射科医生。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.250384
John D Mayfield
{"title":"Unlocking Robust Segmentation: Decoding Domain Randomization for Radiologists.","authors":"John D Mayfield","doi":"10.1148/ryai.250384","DOIUrl":"10.1148/ryai.250384","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 4","pages":"e250384"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144592463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cybersecurity Threats and Mitigation Strategies for Large Language Models in Health Care. 医疗保健中大型语言模型的网络安全威胁和缓解策略。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240739
Tugba Akinci D'Antonoli, Ali S Tejani, Bardia Khosravi, Christian Bluethgen, Felix Busch, Keno K Bressem, Lisa C Adams, Mana Moassefi, Shahriar Faghani, Judy Wawira Gichoya

The integration of large language models (LLMs) into health care offers tremendous opportunities to improve medical practice and patient care. Besides being susceptible to biases and threats common to all artificial intelligence (AI) systems, LLMs pose unique cybersecurity risks that must be carefully evaluated before these AI models are deployed in health care. LLMs can be exploited in several ways, such as malicious attacks, privacy breaches, and unauthorized manipulation of patient data. Moreover, malicious actors could use LLMs to infer sensitive patient information from training data. Furthermore, manipulated or poisoned data fed into these models could change their results in a way that is beneficial for the malicious actors. This report presents the cybersecurity challenges posed by LLMs in health care and provides strategies for mitigation. By implementing robust security measures and adhering to best practices during the model development, training, and deployment stages, stakeholders can help minimize these risks and protect patient privacy. Keywords: Computer Applications-General (Informatics), Application Domain, Large Language Models, Artificial Intelligence, Cybersecurity © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。将大型语言模型(llm)集成到医疗保健中为改善医疗实践和患者护理提供了巨大的机会。除了容易受到所有人工智能系统共同存在的偏见和威胁之外,法学硕士还带来了独特的网络安全风险,在将这些人工智能模型部署到医疗保健领域之前,必须仔细评估这些风险。llm可以以多种方式被利用,例如恶意攻击、隐私泄露和未经授权的患者数据操纵。此外,恶意行为者可以使用llm从训练数据中推断出敏感的患者信息。此外,输入到这些模型中的被操纵或有毒的数据可能会以一种有利于恶意行为者的方式改变它们的结果。本报告介绍了医疗保健法学硕士带来的网络安全挑战,并提供了缓解策略。通过在模型开发、培训和部署阶段实施健壮的安全措施并遵循最佳实践,涉众可以帮助最小化这些风险并保护患者隐私。©RSNA, 2025年。
{"title":"Cybersecurity Threats and Mitigation Strategies for Large Language Models in Health Care.","authors":"Tugba Akinci D'Antonoli, Ali S Tejani, Bardia Khosravi, Christian Bluethgen, Felix Busch, Keno K Bressem, Lisa C Adams, Mana Moassefi, Shahriar Faghani, Judy Wawira Gichoya","doi":"10.1148/ryai.240739","DOIUrl":"10.1148/ryai.240739","url":null,"abstract":"<p><p>The integration of large language models (LLMs) into health care offers tremendous opportunities to improve medical practice and patient care. Besides being susceptible to biases and threats common to all artificial intelligence (AI) systems, LLMs pose unique cybersecurity risks that must be carefully evaluated before these AI models are deployed in health care. LLMs can be exploited in several ways, such as malicious attacks, privacy breaches, and unauthorized manipulation of patient data. Moreover, malicious actors could use LLMs to infer sensitive patient information from training data. Furthermore, manipulated or poisoned data fed into these models could change their results in a way that is beneficial for the malicious actors. This report presents the cybersecurity challenges posed by LLMs in health care and provides strategies for mitigation. By implementing robust security measures and adhering to best practices during the model development, training, and deployment stages, stakeholders can help minimize these risks and protect patient privacy. <b>Keywords:</b> Computer Applications-General (Informatics), Application Domain, Large Language Models, Artificial Intelligence, Cybersecurity © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240739"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143999117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning with Domain Randomization in Image and Feature Spaces for Abdominal Multiorgan Segmentation on CT and MRI Scans. 基于图像和特征空间随机化的深度学习用于腹部CT和MRI多器官分割。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240586
Yu Shi, Lixia Wang, Touseef Ahmad Qureshi, Zengtian Deng, Yibin Xie, Debiao Li

Purpose To develop a deep learning segmentation model that can segment abdominal organs on CT and MRI scans with high accuracy and generalization ability. Materials and Methods In this study, an extended nnU-Net model was trained for abdominal organ segmentation. A domain randomization method in both the image and feature space was developed to improve the generalization ability under cross-site and cross-modality settings on public prostate MRI and abdominal CT and MRI datasets. The prostate MRI dataset contains data from multiple health care institutions, with domain shifts. The abdominal CT and MRI dataset is structured for cross-modality evaluation: training on one modality (eg, MRI) and testing on the other (eg, CT). This domain randomization method was then used to train a segmentation model with enhanced generalization ability on the abdominal multiorgan segmentation challenge dataset to improve abdominal CT and MR multiorgan segmentation, and the model was compared with two commonly used segmentation algorithms (TotalSegmentator and MRSegmentator). Model performance was evaluated using the Dice similarity coefficient (DSC). Results The proposed domain randomization method showed improved generalization ability on the cross-site and cross-modality datasets compared with the state-of-the-art methods. The segmentation model using this method outperformed two other publicly available segmentation models on data from unseen test domains (mean DSC, 0.88 vs 0.79 [P < .001] and 0.88 vs 0.76 [P < .001]). Conclusion The combination of image and feature domain randomizations improved the accuracy and generalization ability of deep learning-based abdominal segmentation on CT and MR images. Keywords: Segmentation, CT, MR Imaging, Neural Networks, MRI, Domain Randomization Supplemental material is available for this article. © RSNA, 2025 See also commentary by Mayfield in this issue.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的建立一种具有较高准确率和泛化能力的腹部器官CT和MR图像深度学习分割模型。在本研究中,我们训练了一个扩展的nnU-Net模型用于腹部器官分割。提出了一种图像和特征空间的域随机化方法,以提高前列腺MRI、腹部CT和MRI数据集在跨位点和跨模态设置下的泛化能力。前列腺MRI数据集包含来自多个医疗保健机构的数据,具有域移位。腹部CT和MRI数据集的结构用于跨模态评估,在一种模态(例如MRI)上进行训练,在另一种模态(例如CT)上进行测试。然后利用该领域随机化方法在腹部多器官分割挑战(AMOS)数据集上训练具有增强泛化能力的分割模型来改进腹部CT和MR多器官分割,并将该模型与两种常用的分割算法(TotalSegmentator和MRSegmentator)进行比较。采用Dice相似系数(DSC)对模型性能进行评价。结果与现有方法相比,所提出的领域随机化方法在跨站点和跨模态数据集上的泛化能力有所提高。使用该方法的分割模型在未见测试域的数据上优于其他两种公开可用的分割模型(平均DSC: 0.88 vs 0.79;P < 0.001和0.88比0.76;P < 0.001)。结论图像随机化与特征域随机化相结合,提高了基于深度学习的CT和MR图像腹部分割的准确率和泛化能力。©rsna, 2025。
{"title":"Deep Learning with Domain Randomization in Image and Feature Spaces for Abdominal Multiorgan Segmentation on CT and MRI Scans.","authors":"Yu Shi, Lixia Wang, Touseef Ahmad Qureshi, Zengtian Deng, Yibin Xie, Debiao Li","doi":"10.1148/ryai.240586","DOIUrl":"10.1148/ryai.240586","url":null,"abstract":"<p><p>Purpose To develop a deep learning segmentation model that can segment abdominal organs on CT and MRI scans with high accuracy and generalization ability. Materials and Methods In this study, an extended nnU-Net model was trained for abdominal organ segmentation. A domain randomization method in both the image and feature space was developed to improve the generalization ability under cross-site and cross-modality settings on public prostate MRI and abdominal CT and MRI datasets. The prostate MRI dataset contains data from multiple health care institutions, with domain shifts. The abdominal CT and MRI dataset is structured for cross-modality evaluation: training on one modality (eg, MRI) and testing on the other (eg, CT). This domain randomization method was then used to train a segmentation model with enhanced generalization ability on the abdominal multiorgan segmentation challenge dataset to improve abdominal CT and MR multiorgan segmentation, and the model was compared with two commonly used segmentation algorithms (TotalSegmentator and MRSegmentator). Model performance was evaluated using the Dice similarity coefficient (DSC). Results The proposed domain randomization method showed improved generalization ability on the cross-site and cross-modality datasets compared with the state-of-the-art methods. The segmentation model using this method outperformed two other publicly available segmentation models on data from unseen test domains (mean DSC, 0.88 vs 0.79 [<i>P</i> < .001] and 0.88 vs 0.76 [<i>P</i> < .001]). Conclusion The combination of image and feature domain randomizations improved the accuracy and generalization ability of deep learning-based abdominal segmentation on CT and MR images. <b>Keywords:</b> Segmentation, CT, MR Imaging, Neural Networks, MRI, Domain Randomization <i>Supplemental material is available for this article.</i> © RSNA, 2025 See also commentary by Mayfield in this issue.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240586"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12476582/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144112076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Optimizing the Trade-off between Privacy and Utility in Medical Imaging Federated Learning. 医学影像联合学习中隐私与效用的权衡优化。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.250434
Zekai Yu
{"title":"Optimizing the Trade-off between Privacy and Utility in Medical Imaging Federated Learning.","authors":"Zekai Yu","doi":"10.1148/ryai.250434","DOIUrl":"10.1148/ryai.250434","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 4","pages":"e250434"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12319695/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144745348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI to Measure Nuchal Translucency: Improved Speed and Accuracy, but Is It Still Relevant? 人工智能测量颈部透明度:提高速度和准确性,但它仍然相关吗?
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.250231
Steven C Horii
{"title":"AI to Measure Nuchal Translucency: Improved Speed and Accuracy, but Is It Still Relevant?","authors":"Steven C Horii","doi":"10.1148/ryai.250231","DOIUrl":"10.1148/ryai.250231","url":null,"abstract":"","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":"7 4","pages":"e250231"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144486217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning Model for Real-Time Nuchal Translucency Assessment at Prenatal US. 实时产前颈部透明度评估的人工智能模型:高性能和工作流集成。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240498
Yuanji Zhang, Xin Yang, Chunya Ji, Xindi Hu, Yan Cao, Chaoyu Chen, He Sui, Binghan Li, Chaojiong Zhen, Weijun Huang, Xuedong Deng, Linliang Yin, Dong Ni

Purpose To develop and evaluate an artificial intelligence-based model for real-time nuchal translucency (NT) plane identification and measurement in prenatal US assessments. Materials and Methods In this retrospective multicenter study conducted from January 2022 to October 2023, the Automated Identification and Measurement of NT (AIM-NT) model was developed and evaluated using internal and external datasets. NT plane assessment, including identification of the NT plane and measurement of NT thickness, was independently conducted by AIM-NT and experienced radiologists, with the results subsequently audited by radiology specialists and accuracy compared between groups. To assess alignment of artificial intelligence with radiologist workflow, discrepancies between the AIM-NT model and radiologists in NT plane identification time and thickness measurements were evaluated. Results The internal dataset included a total of 3959 NT images from 3153 fetuses, and the external dataset included 267 US videos from 267 fetuses. On the internal testing dataset, AIM-NT achieved an area under the receiver operating characteristic curve of 0.92 for NT plane identification. On the external testing dataset, there was no evidence of differences between AIM-NT and radiologists in NT plane identification accuracy (88.8% vs 87.6%, P = .69) or NT thickness measurements on standard and nonstandard NT planes (P = .29 and .59). AIM-NT demonstrated high consistency with radiologists in NT plane identification time, with 1-minute discrepancies observed in 77.9% of cases, and NT thickness measurements, with a mean difference of 0.03 mm and mean absolute error of 0.22 mm (95% CI: 0.19, 0.25). Conclusion AIM-NT demonstrated high accuracy in identifying the NT plane and measuring NT thickness on prenatal US images, showing minimal discrepancies with radiologist workflow. Keywords: Ultrasound, Fetus, Segmentation, Feature Detection, Diagnosis, Convolutional Neural Network (CNN) Supplemental material is available for this article. © RSNA, 2025 See also commentary by Horii in this issue.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的建立一种基于人工智能的实时颈部半透明(NT)平面识别和测量模型,用于产前超声评估。材料和方法在这项于2022年1月至2023年10月进行的回顾性多中心研究中,利用内部和外部数据集开发并评估了NT的自动识别和测量(AIM-NT)模型。NT平面评估,包括NT平面的识别和NT厚度的测量,由AIM-NT和经验丰富的放射科医生独立进行,随后由放射科专家审核结果并比较两组之间的准确性。为了评估人工智能与放射科医生工作流程的一致性,我们评估了AIM-NT模型与放射科医生在NT平面识别时间和厚度测量方面的差异。结果内部数据集包括来自3153个胎儿的3959张NT图像,外部数据集包括来自267个胎儿的267个US视频。在内部测试数据集上,AIM-NT实现了接收机工作特征曲线下面积为0.92的NT平面识别。在外部测试数据集上,AIM-NT和放射科医生在NT平面识别准确率(88.8%对87.6%,P = 0.69)或标准和非标准NT平面上NT厚度测量(P = 0.29, 0.59)方面没有差异。AIM-NT在NT平面识别时间和NT厚度测量方面与放射科医生高度一致,77.9%的病例存在1分钟的差异,平均差异为0.03 mm,平均绝对误差为0.22 mm [95% CI: 0.19 mm, 0.25 mm]。结论AIM-NT在产前超声中识别NT平面和测量NT厚度具有较高的准确性,与放射科医生的工作流程差异很小。©RSNA, 2025年。
{"title":"Deep Learning Model for Real-Time Nuchal Translucency Assessment at Prenatal US.","authors":"Yuanji Zhang, Xin Yang, Chunya Ji, Xindi Hu, Yan Cao, Chaoyu Chen, He Sui, Binghan Li, Chaojiong Zhen, Weijun Huang, Xuedong Deng, Linliang Yin, Dong Ni","doi":"10.1148/ryai.240498","DOIUrl":"10.1148/ryai.240498","url":null,"abstract":"<p><p>Purpose To develop and evaluate an artificial intelligence-based model for real-time nuchal translucency (NT) plane identification and measurement in prenatal US assessments. Materials and Methods In this retrospective multicenter study conducted from January 2022 to October 2023, the Automated Identification and Measurement of NT (AIM-NT) model was developed and evaluated using internal and external datasets. NT plane assessment, including identification of the NT plane and measurement of NT thickness, was independently conducted by AIM-NT and experienced radiologists, with the results subsequently audited by radiology specialists and accuracy compared between groups. To assess alignment of artificial intelligence with radiologist workflow, discrepancies between the AIM-NT model and radiologists in NT plane identification time and thickness measurements were evaluated. Results The internal dataset included a total of 3959 NT images from 3153 fetuses, and the external dataset included 267 US videos from 267 fetuses. On the internal testing dataset, AIM-NT achieved an area under the receiver operating characteristic curve of 0.92 for NT plane identification. On the external testing dataset, there was no evidence of differences between AIM-NT and radiologists in NT plane identification accuracy (88.8% vs 87.6%, <i>P</i> = .69) or NT thickness measurements on standard and nonstandard NT planes (<i>P</i> = .29 and .59). AIM-NT demonstrated high consistency with radiologists in NT plane identification time, with 1-minute discrepancies observed in 77.9% of cases, and NT thickness measurements, with a mean difference of 0.03 mm and mean absolute error of 0.22 mm (95% CI: 0.19, 0.25). Conclusion AIM-NT demonstrated high accuracy in identifying the NT plane and measuring NT thickness on prenatal US images, showing minimal discrepancies with radiologist workflow. <b>Keywords:</b> Ultrasound, Fetus, Segmentation, Feature Detection, Diagnosis, Convolutional Neural Network (CNN) <i>Supplemental material is available for this article.</i> © RSNA, 2025 See also commentary by Horii in this issue.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240498"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143999114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RadioRAG: Online Retrieval-Augmented Generation for Radiology Question Answering. RadioRAG:放射学问答的在线检索增强生成。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240476
Soroosh Tayebi Arasteh, Mahshad Lotfinia, Keno Bressem, Robert Siepmann, Lisa Adams, Dyke Ferber, Christiane Kuhl, Jakob Nikolas Kather, Sven Nebelung, Daniel Truhn

Purpose To evaluate diagnostic accuracy of various large language models (LLMs) when answering radiology-specific questions with and without access to additional online, up-to-date information via retrieval-augmented generation (RAG). Materials and Methods The authors developed radiology RAG (RadioRAG), an end-to-end framework that retrieves data from authoritative radiologic online sources in real-time. RAG incorporates information retrieval from external sources to supplement the initial prompt, grounding the model's response in relevant information. Using 80 questions from the RSNA Case Collection across radiologic subspecialties and 24 additional expert-curated questions with reference standard answers, LLMs (GPT-3.5-turbo [OpenAI], GPT-4, Mistral 7B, Mixtral 8×7B [Mistral], and Llama3-8B and -70B [Meta]) were prompted with and without RadioRAG in a zero-shot inference scenario (temperature ≤ 0.1, top-p = 1). RadioRAG retrieved context-specific information from www.radiopaedia.org. Accuracy of LLMs with and without RadioRAG in answering questions from each dataset was assessed. Statistical analyses were performed using bootstrapping while preserving pairing. Additional assessments included comparison of model with human performance and comparison of time required for conventional versus RadioRAG-powered question answering. Results RadioRAG improved accuracy for some LLMs, including GPT-3.5-turbo (74% [59 of 80] vs 66% [53 of 80], false discovery rate [FDR] = 0.03) and Mixtral 8×7B (76% [61 of 80] vs 65% [52 of 80], FDR = 0.02) on the RSNA radiology question answering (RSNA-RadioQA) dataset, with similar trends in the ExtendedQA dataset. Accuracy exceeded that of a human expert (63% [50 of 80], FDR ≤ 0.007) for these LLMs, although not for Mistral 7B-instruct-v0.2, Llama3-8B, and Llama3-70B (FDR ≥ 0.21). RadioRAG reduced hallucinations for all LLMs (rate, 6%-25%). RadioRAG increased estimated response time fourfold. Conclusion RadioRAG shows potential to improve LLM accuracy and factuality in radiology QA by integrating real-time, domain-specific data. Keywords: Retrieval-augmented Generation, Informatics, Computer-aided Diagnosis, Large Language Models Supplemental material is available for this article. © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的评估各种大型语言模型(llm)在回答放射学特定问题时的诊断准确性,并通过检索增强生成(RAG)获取额外的在线最新信息。材料和方法作者开发了Radiology RAG (RadioRAG),这是一个端到端的框架,可以实时从权威的放射学在线资源中检索数据。RAG结合了来自外部来源的信息检索,以补充初始提示,将模型的响应建立在相关信息中。利用来自RSNA病例集的80个问题和24个额外的专家设计的问题,以及参考标准答案,LLMs (GPT-3.5-turbo、GPT-4、Mistral-7B、Mixtral-8 × 7B和Llama3 [8B和70B])在零shot推断场景(温度≤0.1,top- P = 1)中提示是否使用RadioRAG。RadioRAG从www.radiopaedia.org检索特定于上下文的信息。评估了使用和不使用RadioRAG的llm在回答每个数据集的问题时的准确性。统计分析采用自举法,同时保持配对。其他评估包括将模型与人工表现进行比较,以及将传统问答与radiorag问答所需的时间进行比较。结果RadioRAG提高了一些LLMs的准确性,包括GPT-3.5-turbo[74%(59/80)对66% (53/80),FDR = 0.03]和Mixtral-8 × 7B[76%(61/80)对65% (52/80),FDR = 0.02],在RSNA-RadioQA数据集中也有类似的趋势。对于这些llm,准确率超过人类专家(FDR≤0.007)(63%,(50/80)),而对于mistral - 7b - directive -v0.2, Llama3-8B和Llama3-70B则没有(FDR≥0.21)。RadioRAG减少了所有llm的幻觉(比率从6-25%)。RadioRAG将估计响应时间提高了四倍。结论RadioRAG通过整合实时领域特定数据,有可能提高LLM在放射学问题回答中的准确性和真实性。©RSNA, 2025年。
{"title":"RadioRAG: Online Retrieval-Augmented Generation for Radiology Question Answering.","authors":"Soroosh Tayebi Arasteh, Mahshad Lotfinia, Keno Bressem, Robert Siepmann, Lisa Adams, Dyke Ferber, Christiane Kuhl, Jakob Nikolas Kather, Sven Nebelung, Daniel Truhn","doi":"10.1148/ryai.240476","DOIUrl":"10.1148/ryai.240476","url":null,"abstract":"<p><p>Purpose To evaluate diagnostic accuracy of various large language models (LLMs) when answering radiology-specific questions with and without access to additional online, up-to-date information via retrieval-augmented generation (RAG). Materials and Methods The authors developed radiology RAG (RadioRAG), an end-to-end framework that retrieves data from authoritative radiologic online sources in real-time. RAG incorporates information retrieval from external sources to supplement the initial prompt, grounding the model's response in relevant information. Using 80 questions from the RSNA Case Collection across radiologic subspecialties and 24 additional expert-curated questions with reference standard answers, LLMs (GPT-3.5-turbo [OpenAI], GPT-4, Mistral 7B, Mixtral 8×7B [Mistral], and Llama3-8B and -70B [Meta]) were prompted with and without RadioRAG in a zero-shot inference scenario (temperature ≤ 0.1, top-p = 1). RadioRAG retrieved context-specific information from <i>www.radiopaedia.org</i>. Accuracy of LLMs with and without RadioRAG in answering questions from each dataset was assessed. Statistical analyses were performed using bootstrapping while preserving pairing. Additional assessments included comparison of model with human performance and comparison of time required for conventional versus RadioRAG-powered question answering. Results RadioRAG improved accuracy for some LLMs, including GPT-3.5-turbo (74% [59 of 80] vs 66% [53 of 80], false discovery rate [FDR] = 0.03) and Mixtral 8×7B (76% [61 of 80] vs 65% [52 of 80], FDR = 0.02) on the RSNA radiology question answering (RSNA-RadioQA) dataset, with similar trends in the ExtendedQA dataset. Accuracy exceeded that of a human expert (63% [50 of 80], FDR ≤ 0.007) for these LLMs, although not for Mistral 7B-instruct-v0.2, Llama3-8B, and Llama3-70B (FDR ≥ 0.21). RadioRAG reduced hallucinations for all LLMs (rate, 6%-25%). RadioRAG increased estimated response time fourfold. Conclusion RadioRAG shows potential to improve LLM accuracy and factuality in radiology QA by integrating real-time, domain-specific data. <b>Keywords:</b> Retrieval-augmented Generation, Informatics, Computer-aided Diagnosis, Large Language Models <i>Supplemental material is available for this article.</i> © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240476"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12326075/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144327031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy-preserving Federated Learning and Uncertainty Quantification in Medical Imaging. 医学影像中隐私保护的联邦学习和不确定性量化。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240637
Nikolas Koutsoubis, Asim Waqas, Yasin Yilmaz, Ravi P Ramachandran, Matthew B Schabath, Ghulam Rasool

Artificial intelligence (AI) has demonstrated strong potential in automating medical imaging tasks, with potential applications across disease diagnosis, prognosis, treatment planning, and posttreatment surveillance. However, privacy concerns surrounding patient data remain a major barrier to the widespread adoption of AI in clinical practice, because large and diverse training datasets are essential for developing accurate, robust, and generalizable AI models. Federated learning offers a privacy-preserving solution by enabling collaborative model training across institutions without sharing sensitive data. Instead, model parameters, such as model weights, are exchanged between participating sites. Despite its potential, federated learning is still in its early stages of development and faces several challenges. Notably, sensitive information can still be inferred from the shared model parameters. Additionally, postdeployment data distribution shifts can degrade model performance, making uncertainty quantification essential. In federated learning, this task is particularly challenging due to data heterogeneity across participating sites. This review provides a comprehensive overview of federated learning, privacy-preserving federated learning, and uncertainty quantification in federated learning. Key limitations in current methodologies are identified, and future research directions are proposed to enhance data privacy and trustworthiness in medical imaging applications. Keywords: Supervised Learning, Perception, Neural Networks, Radiology-Pathology Integration Supplemental material is available for this article. © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。人工智能(AI)在自动化医学成像任务方面显示出强大的潜力,在疾病诊断、预后、治疗计划和治疗后监测方面具有潜在的应用前景。然而,围绕患者数据的隐私问题仍然是人工智能在临床实践中广泛采用的主要障碍,因为大型和多样化的训练数据集对于开发准确、稳健和可推广的人工智能模型至关重要。联邦学习通过在不共享敏感数据的情况下实现跨机构的协作模型培训,提供了一种保护隐私的解决方案。相反,模型参数(如模型权重)在参与站点之间交换。尽管具有潜力,但联邦学习仍处于发展的早期阶段,并面临着一些挑战。值得注意的是,敏感信息仍然可以从共享模型参数中推断出来。此外,部署后数据分布的变化会降低模型的性能,使得不确定性量化变得至关重要。在联邦学习中,由于参与站点之间的数据异构性,这项任务尤其具有挑战性。本文对联邦学习、隐私保护联邦学习和联邦学习中的不确定性量化进行了全面的综述。指出了当前方法的主要局限性,并提出了未来的研究方向,以增强医学成像应用中的数据隐私和可信度。©RSNA, 2025年。
{"title":"Privacy-preserving Federated Learning and Uncertainty Quantification in Medical Imaging.","authors":"Nikolas Koutsoubis, Asim Waqas, Yasin Yilmaz, Ravi P Ramachandran, Matthew B Schabath, Ghulam Rasool","doi":"10.1148/ryai.240637","DOIUrl":"10.1148/ryai.240637","url":null,"abstract":"<p><p>Artificial intelligence (AI) has demonstrated strong potential in automating medical imaging tasks, with potential applications across disease diagnosis, prognosis, treatment planning, and posttreatment surveillance. However, privacy concerns surrounding patient data remain a major barrier to the widespread adoption of AI in clinical practice, because large and diverse training datasets are essential for developing accurate, robust, and generalizable AI models. Federated learning offers a privacy-preserving solution by enabling collaborative model training across institutions without sharing sensitive data. Instead, model parameters, such as model weights, are exchanged between participating sites. Despite its potential, federated learning is still in its early stages of development and faces several challenges. Notably, sensitive information can still be inferred from the shared model parameters. Additionally, postdeployment data distribution shifts can degrade model performance, making uncertainty quantification essential. In federated learning, this task is particularly challenging due to data heterogeneity across participating sites. This review provides a comprehensive overview of federated learning, privacy-preserving federated learning, and uncertainty quantification in federated learning. Key limitations in current methodologies are identified, and future research directions are proposed to enhance data privacy and trustworthiness in medical imaging applications. <b>Keywords:</b> Supervised Learning, Perception, Neural Networks, Radiology-Pathology Integration <i>Supplemental material is available for this article.</i> © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240637"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12319697/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144053086","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimating Total Lung Volume from Pixel-Level Thickness Maps of Chest Radiographs Using Deep Learning. 利用深度学习从胸片像素级厚度图估计总肺容量。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240484
Tina Dorosti, Manuel Schultheiß, Philipp Schmette, Jule Heuchert, Johannes Thalhammer, Florian T Gassert, Thorsten Sellerer, Rafael Schick, Kirsten Taphorn, Korbinian Mechlem, Lorenz Birnbacher, Florian Schaff, Franz Pfeiffer, Daniela Pfeiffer

Purpose To estimate the total lung volume (TLV) from real and synthetic frontal chest radiographs on a pixel level using lung thickness maps generated by a U-Net deep learning model. Materials and Methods This retrospective study included 5959 chest CT scans from two public datasets, the Lung Nodule Analysis 2016 (Luna16) (n = 656) and the Radiological Society of North America Pulmonary Embolism Detection Challenge 2020 (n = 5303). Additionally, 72 participants were selected from the Klinikum Rechts der Isar dataset (October 2018 through December 2019), each with a corresponding chest radiograph obtained within 7 days. Synthetic radiographs and lung thickness maps were generated using forward projection of CT scans and their lung segmentations. A U-Net model was trained on synthetic radiographs to predict lung thickness maps and estimate TLV. Model performance was assessed using mean squared error (MSE), Pearson correlation coefficient, and two-sided Student t distribution. Results The study included 72 participants (45 male and 27 female participants; 33 healthy participants: mean age, 62 years [range, 34-80 years]; 39 with chronic obstructive pulmonary disease: mean age, 69 years [range, 47-91 years]). TLV predictions showed low error rates (MSEPublic-Synthetic, 0.16 L2; MSEKRI-Synthetic, 0.20 L2; MSEKRI-Real, 0.35 L2) and strong correlations with CT-derived reference standard TLV (nPublic-Synthetic, 1191; r = 0.99; P < .001) (nKRI-Synthetic, 72; r = 0.97; P < .001) (nKRI-Real, 72; r = 0.91; P < .001). When evaluated on different datasets, the U-Net model achieved the highest performance for TLV estimation on the Luna16 test dataset, with the lowest MSE (0.09 L2) and strongest correlation (r = 0.99; P < .001) compared with CT-derived TLV. Conclusion The U-Net-generated pixel-level lung thickness maps successfully estimated TLV for both synthetic and real radiographs. Keywords: Frontal Chest Radiographs, Lung Thickness Map, Pixel-Level, Total Lung Volume, U-Net Supplemental material is available for this article. © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的利用U-Net深度学习模型生成的肺厚度图,在像素水平上估计真实和合成胸部x线片(CXR)的肺总容积(TLV)。本回顾性研究包括来自两个公共数据集的5959个胸部CT扫描:2016年肺结节分析(n = 656)和北美放射学会(RSNA) 2020年肺栓塞检测挑战(n = 5303)。此外,从Klinikum Rechts der Isar数据集(2018年10月至2019年12月)中选择72名参与者,每个参与者在7天内拍摄相应的胸片。合成x线片和肺厚度图是利用CT扫描的正投影及其肺分割生成的。在合成x线片上训练U-Net模型来预测肺厚度图和估计TLV。采用均方误差(MSE)、Pearson相关系数(r)和双侧Student's t分布评估模型性能。结果共纳入72例受试者,其中男45例,女27例,健康33例,平均年龄62岁[范围34 ~ 80岁];39例慢性阻塞性肺疾病:平均年龄69岁[范围47-91])。TLV预测的误差率低(MSEPublic-Synthetic = 0.16 L2, MSEKRI-Synthetic = 0.20 L2, MSEKRI-Real = 0.35 L2),与ct衍生的参考标准TLV有很强的相关性(nPublic-Synthetic = 1191, r = 0.99, P < .001;nKRI-Synthetic = 72, r = 0.97, P < 0.001;nKRI-Real = 72, r = 0.91, P < 0.001)。当在不同的数据集上进行评估时,U-Net模型在Luna16测试数据集上获得了最高的TLV估计性能,均方误差最小(MSE = 0.09 L2),相关性最强(r = 0.99, P
{"title":"Estimating Total Lung Volume from Pixel-Level Thickness Maps of Chest Radiographs Using Deep Learning.","authors":"Tina Dorosti, Manuel Schultheiß, Philipp Schmette, Jule Heuchert, Johannes Thalhammer, Florian T Gassert, Thorsten Sellerer, Rafael Schick, Kirsten Taphorn, Korbinian Mechlem, Lorenz Birnbacher, Florian Schaff, Franz Pfeiffer, Daniela Pfeiffer","doi":"10.1148/ryai.240484","DOIUrl":"10.1148/ryai.240484","url":null,"abstract":"<p><p>Purpose To estimate the total lung volume (TLV) from real and synthetic frontal chest radiographs on a pixel level using lung thickness maps generated by a U-Net deep learning model. Materials and Methods This retrospective study included 5959 chest CT scans from two public datasets, the Lung Nodule Analysis 2016 (Luna16) (<i>n</i> = 656) and the Radiological Society of North America Pulmonary Embolism Detection Challenge 2020 (<i>n</i> = 5303). Additionally, 72 participants were selected from the Klinikum Rechts der Isar dataset (October 2018 through December 2019), each with a corresponding chest radiograph obtained within 7 days. Synthetic radiographs and lung thickness maps were generated using forward projection of CT scans and their lung segmentations. A U-Net model was trained on synthetic radiographs to predict lung thickness maps and estimate TLV. Model performance was assessed using mean squared error (MSE), Pearson correlation coefficient, and two-sided Student <i>t</i> distribution. Results The study included 72 participants (45 male and 27 female participants; 33 healthy participants: mean age, 62 years [range, 34-80 years]; 39 with chronic obstructive pulmonary disease: mean age, 69 years [range, 47-91 years]). TLV predictions showed low error rates (MSE<sub>Public-Synthetic</sub>, 0.16 L<sup>2</sup>; MSE<sub>KRI-Synthetic</sub>, 0.20 L<sup>2</sup>; MSE<sub>KRI-Real</sub>, 0.35 L<sup>2</sup>) and strong correlations with CT-derived reference standard TLV (<i>n</i><sub>Public-Synthetic</sub>, 1191; <i>r</i> = 0.99; <i>P</i> < .001) (<i>n</i><sub>KRI-Synthetic</sub>, 72; <i>r</i> = 0.97; <i>P</i> < .001) (<i>n</i><sub>KRI-Real</sub>, 72; <i>r</i> = 0.91; <i>P</i> < .001). When evaluated on different datasets, the U-Net model achieved the highest performance for TLV estimation on the Luna16 test dataset, with the lowest MSE (0.09 L<sup>2</sup>) and strongest correlation (<i>r</i> = 0.99; <i>P</i> < .001) compared with CT-derived TLV. Conclusion The U-Net-generated pixel-level lung thickness maps successfully estimated TLV for both synthetic and real radiographs. <b>Keywords:</b> Frontal Chest Radiographs, Lung Thickness Map, Pixel-Level, Total Lung Volume, U-Net <i>Supplemental material is available for this article.</i> © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240484"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144162222","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial Intelligence in Breast US Diagnosis and Report Generation. 人工智能在乳腺诊断和报告生成中的应用。
IF 13.2 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-07-01 DOI: 10.1148/ryai.240625
Jian Wang, HongTian Tian, Xin Yang, HuaiYu Wu, XiLiang Zhu, RuSi Chen, Ao Chang, YanLin Chen, HaoRan Dou, RuoBing Huang, Jun Cheng, YongSong Zhou, Rui Gao, KeEn Yang, GuoQiu Li, Jing Chen, Dong Ni, JinFeng Xu, Ning Gu, FaJin Dong

Purpose To develop and evaluate an artificial intelligence (AI) system for generating breast US reports. Materials and Methods This retrospective study included 104 364 cases from three hospitals (January 2020-December 2022). The AI system was trained on 82 896 cases, validated on 10 385 cases, and tested on an internal set (10 383 cases) and two external sets (300 and 400 cases). Under blind review, three senior radiologists (each with >10 years of experience) evaluated AI-generated reports and those written by one midlevel radiologist (with 7 years of experience), as well as reports from three junior radiologists (each with 2-3 years of experience) with and without AI assistance. The primary outcomes included the acceptance rates of Breast Imaging Reporting and Data System (BI-RADS) categories and lesion characteristics. Statistical analysis included one-sided and two-sided McNemar tests for noninferiority and significance testing. Results In external test set 1 (300 cases), the midlevel radiologist and AI system achieved BI-RADS acceptance rates of 95.00% (285 of 300) versus 92.33% (277 of 300) (P < .001, noninferiority test with a prespecified margin of 10%). In external test set 2 (400 cases), three junior radiologists had BI-RADS acceptance rates of 87.00% (348 of 400) versus 90.75% (363 of 400) (P = .06), 86.50% (346 of 400) versus 92.00% (368 of 400) (P = .007), and 84.75% (339 of 400) versus 90.25% (361 of 400) (P = .02) without and with AI assistance, respectively. Conclusion The AI system performed comparably to a midlevel radiologist and aided junior radiologists in BI-RADS classification. Keywords: Neural Networks, Computer-aided Diagnosis, CAD, Ultrasound Supplemental material is available for this article. © RSNA, 2025.

“刚刚接受”的论文经过了全面的同行评审,并已被接受发表在《放射学:人工智能》杂志上。这篇文章将经过编辑,布局和校样审查,然后在其最终版本出版。请注意,在最终编辑文章的制作过程中,可能会发现可能影响内容的错误。目的开发和评估用于生成乳腺超声(BUS)报告的人工智能(AI)系统。材料与方法回顾性研究纳入2020年1月- 2022年12月三家医院104364例病例。人工智能系统在82896个案例上进行了训练,在10385个案例上进行了验证,并在一个内部集(10383个案例)和两个外部集(300和400个案例)上进行了测试。在盲评的情况下,三名资深放射科医生(50 - 10年经验)评估了人工智能生成的报告、一名中级放射科医生(7年经验)撰写的报告,以及三名初级放射科医生(2-3年经验)在有和没有人工智能帮助的情况下的报告。主要结果包括乳腺成像报告和数据系统(BI-RADS)分类的接受率和病变特征。统计分析包括单侧和双侧McNemar非劣效性检验和显著性检验。结果在外部测试集1(300例)中,中级放射科医师和人工智能系统的BI-RADS满意率分别为95.00%[285/300]和92.33% [277/300](P < .001;非劣效性检验,预设裕度为10%)。在外部测试集2(400例)中,3名初级放射科医师在人工智能辅助和非人工智能辅助下的BI-RADS满意率分别为87.00%[348/400]对90.75% [363/400](P = .06)、86.50%[346/400]对92.00% [368/400](P = .007)和84.75%[339/400]对90.25% [361/400](P = .02)。结论人工智能系统对中级放射科医生和辅助初级放射科医生进行BI-RADS分类的效果相当。©RSNA, 2025年。
{"title":"Artificial Intelligence in Breast US Diagnosis and Report Generation.","authors":"Jian Wang, HongTian Tian, Xin Yang, HuaiYu Wu, XiLiang Zhu, RuSi Chen, Ao Chang, YanLin Chen, HaoRan Dou, RuoBing Huang, Jun Cheng, YongSong Zhou, Rui Gao, KeEn Yang, GuoQiu Li, Jing Chen, Dong Ni, JinFeng Xu, Ning Gu, FaJin Dong","doi":"10.1148/ryai.240625","DOIUrl":"10.1148/ryai.240625","url":null,"abstract":"<p><p>Purpose To develop and evaluate an artificial intelligence (AI) system for generating breast US reports. Materials and Methods This retrospective study included 104 364 cases from three hospitals (January 2020-December 2022). The AI system was trained on 82 896 cases, validated on 10 385 cases, and tested on an internal set (10 383 cases) and two external sets (300 and 400 cases). Under blind review, three senior radiologists (each with >10 years of experience) evaluated AI-generated reports and those written by one midlevel radiologist (with 7 years of experience), as well as reports from three junior radiologists (each with 2-3 years of experience) with and without AI assistance. The primary outcomes included the acceptance rates of Breast Imaging Reporting and Data System (BI-RADS) categories and lesion characteristics. Statistical analysis included one-sided and two-sided McNemar tests for noninferiority and significance testing. Results In external test set 1 (300 cases), the midlevel radiologist and AI system achieved BI-RADS acceptance rates of 95.00% (285 of 300) versus 92.33% (277 of 300) (<i>P</i> < .001, noninferiority test with a prespecified margin of 10%). In external test set 2 (400 cases), three junior radiologists had BI-RADS acceptance rates of 87.00% (348 of 400) versus 90.75% (363 of 400) (<i>P</i> = .06), 86.50% (346 of 400) versus 92.00% (368 of 400) (<i>P</i> = .007), and 84.75% (339 of 400) versus 90.25% (361 of 400) (<i>P</i> = .02) without and with AI assistance, respectively. Conclusion The AI system performed comparably to a midlevel radiologist and aided junior radiologists in BI-RADS classification. <b>Keywords:</b> Neural Networks, Computer-aided Diagnosis, CAD, Ultrasound <i>Supplemental material is available for this article.</i> © RSNA, 2025.</p>","PeriodicalId":29787,"journal":{"name":"Radiology-Artificial Intelligence","volume":" ","pages":"e240625"},"PeriodicalIF":13.2,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144327030","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Radiology-Artificial Intelligence
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1