{"title":"Perivascular Epithelioid Cell Tumor of the Liver: A Rare and Difficult Case Diagnosis.","authors":"Amara A Cuello, Miguel E Nazar","doi":"10.1148/radiol.241611","DOIUrl":"https://doi.org/10.1148/radiol.241611","url":null,"abstract":"","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e241611"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143053271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Brian H Mu, Faris Galambo, Hadeer W Al-Ali, Sumeet G Dua, Chanae D Dixon, Xinhai R Zhang, Mustafa A Mafraji
History: A 38-year-old previously healthy male patient presented with left-sided facial pain over the prior 5 weeks. He first noticed the pain while washing and applying pressure to his face. The pain was described as shock-like, sharp and shooting, and radiating along the left cheek and temple. It began as 1-2-second episodes occurring two to three times per day, sometimes spontaneously, progressing in severity and frequency over time. Mild progressive left facial weakness also developed a few weeks after initial symptoms. Physical examination demonstrated reproducible pain in the distribution of the maxillary division of the trigeminal nerve (V2), with normal motor and sensory function. A recent routine dental examination demonstrated healthy teeth and gums, and there was no history of dental procedures or trauma. The rest of the physical and neurologic examinations revealed no abnormalities. The patient was afebrile with normal vital signs. Findings of routine laboratory testing, including complete blood count, metabolic panel with electrolytes, kidney and liver function, and inflammatory markers such as C-reactive protein, were all within normal limits. Following the neurologic and otolaryngologic evaluations, imaging was recommended. The patient was also started on treatment with carbamazepine for trigeminal neuralgia, with modest improvement of symptoms. He initially underwent MRI of the temporal bones at an outside hospital. After subsequent referral to our hospital, follow-up concomitant MRI and CT (Figs 1-4) were performed approximately 3 months after the initial imaging.
{"title":"Case 337.","authors":"Brian H Mu, Faris Galambo, Hadeer W Al-Ali, Sumeet G Dua, Chanae D Dixon, Xinhai R Zhang, Mustafa A Mafraji","doi":"10.1148/radiol.241909","DOIUrl":"https://doi.org/10.1148/radiol.241909","url":null,"abstract":"<p><strong>History: </strong>A 38-year-old previously healthy male patient presented with left-sided facial pain over the prior 5 weeks. He first noticed the pain while washing and applying pressure to his face. The pain was described as shock-like, sharp and shooting, and radiating along the left cheek and temple. It began as 1-2-second episodes occurring two to three times per day, sometimes spontaneously, progressing in severity and frequency over time. Mild progressive left facial weakness also developed a few weeks after initial symptoms. Physical examination demonstrated reproducible pain in the distribution of the maxillary division of the trigeminal nerve (V2), with normal motor and sensory function. A recent routine dental examination demonstrated healthy teeth and gums, and there was no history of dental procedures or trauma. The rest of the physical and neurologic examinations revealed no abnormalities. The patient was afebrile with normal vital signs. Findings of routine laboratory testing, including complete blood count, metabolic panel with electrolytes, kidney and liver function, and inflammatory markers such as C-reactive protein, were all within normal limits. Following the neurologic and otolaryngologic evaluations, imaging was recommended. The patient was also started on treatment with carbamazepine for trigeminal neuralgia, with modest improvement of symptoms. He initially underwent MRI of the temporal bones at an outside hospital. After subsequent referral to our hospital, follow-up concomitant MRI and CT (Figs 1-4) were performed approximately 3 months after the initial imaging.</p>","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e241909"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143053305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Deep Learning MRI Reconstruction Delivers Superior Resolution and Improved Diagnostics.","authors":"Mika T Nevalainen","doi":"10.1148/radiol.242952","DOIUrl":"https://doi.org/10.1148/radiol.242952","url":null,"abstract":"","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e242952"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143053308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Cody H Savage, Adway Kanhere, Vishwa Parekh, Curtis P Langlotz, Anupam Joshi, Heng Huang, Florence X Doo
Integrating large language models (LLMs) into health care holds substantial potential to enhance clinical workflows and care delivery. However, LLMs also pose serious risks if integration is not thoughtfully executed, with complex challenges spanning accuracy, accessibility, privacy, and regulation. Proprietary commercial LLMs (eg, GPT-4 [OpenAI], Claude 3 Sonnet and Claude 3 Opus [Anthropic], Gemini [Google]) have received much attention from researchers in the medical domain, including radiology. Interestingly, open-source LLMs (eg, Llama 3 and LLaVA-Med) have received comparatively little attention. Yet, open-source LLMs hold several key advantages over proprietary LLMs for medical institutions, hospitals, and individual researchers. The wider adoption of open-source LLMs has been slower, perhaps in part due to the lack of familiarity, accessible computational infrastructure, and community-built tools to streamline their local implementation and customize them for specific use cases. Thus, this article provides a tutorial for the implementation of open-source LLMs in radiology, including examples of commonly used tools for text generation and techniques for troubleshooting issues with prompt engineering, retrieval-augmented generation, and fine-tuning. Implementation-ready code for each tool is provided at https://github.com/UM2ii/Open-Source-LLM-Tools-for-Radiology. In addition, this article compares the benefits and drawbacks of open-source and proprietary LLMs, discusses the differentiating characteristics of popular open-source LLMs, and highlights recent advancements that may affect their adoption.
{"title":"Open-Source Large Language Models in Radiology: A Review and Tutorial for Practical Research and Clinical Deployment.","authors":"Cody H Savage, Adway Kanhere, Vishwa Parekh, Curtis P Langlotz, Anupam Joshi, Heng Huang, Florence X Doo","doi":"10.1148/radiol.241073","DOIUrl":"10.1148/radiol.241073","url":null,"abstract":"<p><p>Integrating large language models (LLMs) into health care holds substantial potential to enhance clinical workflows and care delivery. However, LLMs also pose serious risks if integration is not thoughtfully executed, with complex challenges spanning accuracy, accessibility, privacy, and regulation. Proprietary commercial LLMs (eg, GPT-4 [OpenAI], Claude 3 Sonnet and Claude 3 Opus [Anthropic], Gemini [Google]) have received much attention from researchers in the medical domain, including radiology. Interestingly, open-source LLMs (eg, Llama 3 and LLaVA-Med) have received comparatively little attention. Yet, open-source LLMs hold several key advantages over proprietary LLMs for medical institutions, hospitals, and individual researchers. The wider adoption of open-source LLMs has been slower, perhaps in part due to the lack of familiarity, accessible computational infrastructure, and community-built tools to streamline their local implementation and customize them for specific use cases. Thus, this article provides a tutorial for the implementation of open-source LLMs in radiology, including examples of commonly used tools for text generation and techniques for troubleshooting issues with prompt engineering, retrieval-augmented generation, and fine-tuning. Implementation-ready code for each tool is provided at <i>https://github.com/UM2ii/Open-Source-LLM-Tools-for-Radiology</i>. In addition, this article compares the benefits and drawbacks of open-source and proprietary LLMs, discusses the differentiating characteristics of popular open-source LLMs, and highlights recent advancements that may affect their adoption.</p>","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e241073"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11783163/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143053245","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Unraveling the Obesity Paradox in Non-Small Cell Lung Cancer.","authors":"Michael W Vannier","doi":"10.1148/radiol.243509","DOIUrl":"https://doi.org/10.1148/radiol.243509","url":null,"abstract":"","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e243509"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143053301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Consideration of Thermal Ablation for Secondary Hyperparathyroidism in Patients with Chronic Kidney Disease.","authors":"Joseph J Gemmete","doi":"10.1148/radiol.243288","DOIUrl":"https://doi.org/10.1148/radiol.243288","url":null,"abstract":"","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e243288"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142954066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Sebastian Nowak, Benjamin Wulff, Yannik C Layer, Maike Theis, Alexander Isaak, Babak Salam, Wolfgang Block, Daniel Kuetting, Claus C Pieper, Julian A Luetkens, Ulrike Attenberger, Alois M Sprinkart
Background Large-scale secondary use of clinical databases requires automated tools for retrospective extraction of structured content from free-text radiology reports. Purpose To share data and insights on the application of privacy-preserving open-weights large language models (LLMs) for reporting content extraction with comparison to standard rule-based systems and the closed-weights LLMs from OpenAI. Materials and Methods In this retrospective exploratory study conducted between May 2024 and September 2024, zero-shot prompting of 17 open-weights LLMs was preformed. These LLMs with model weights released under open licenses were compared with rule-based annotation and with OpenAI's GPT-4o, GPT-4o-mini, GPT-4-turbo, and GPT-3.5-turbo on a manually annotated public English chest radiography dataset (Indiana University, 3927 patients and reports). An annotated nonpublic German chest radiography dataset (18 500 reports, 16 844 patients [10 340 male; mean age, 62.6 years ± 21.5 {SD}]) was used to compare local fine-tuning of all open-weights LLMs via low-rank adaptation and 4-bit quantization to bidirectional encoder representations from transformers (BERT) with different subsets of reports (from 10 to 14 580). Nonoverlapping 95% CIs of macro-averaged F1 scores were defined as relevant differences. Results For the English reports, the highest zero-shot macro-averaged F1 score was observed for GPT-4o (92.4% [95% CI: 87.9, 95.9]); GPT-4o outperformed the rule-based CheXpert [Stanford University] (73.1% [95% CI: 65.1, 79.7]) but was comparable in performance to several open-weights LLMs (top three: Mistral-Large [Mistral AI], 92.6% [95% CI: 88.2, 96.0]; Llama-3.1-70b [Meta AI], 92.2% [95% CI: 87.1, 95.8]; and Llama-3.1-405b [Meta AI]: 90.3% [95% CI: 84.6, 94.5]). For the German reports, Mistral-Large (91.6% [95% CI: 90.5, 92.7]) had the highest zero-shot macro-averaged F1 score among the six other open-weights LLMs and outperformed the rule-based annotation (74.8% [95% CI: 73.3, 76.1]). Using 1000 reports for fine-tuning, all LLMs (top three: Mistral-Large, 94.3% [95% CI: 93.5, 95.2]; OpenBioLLM-70b [Saama]: 93.9% [95% CI: 92.9, 94.8]; and Mixtral-8×22b [Mistral AI]: 93.8% [95% CI: 92.8, 94.7]) achieved significantly higher macro-averaged F1 score than did BERT (86.7% [95% CI: 85.0, 88.3]); however, the differences were not relevant when 2000 or more reports were used for fine-tuning. Conclusion LLMs have the potential to outperform rule-based systems for zero-shot "out-of-the-box" structuring of report databases, with privacy-ensuring open-weights LLMs being competitive with closed-weights GPT-4o. Additionally, the open-weights LLM outperformed BERT when moderate numbers of reports were used for fine-tuning. Published under a CC BY 4.0 license. Supplemental material is available for this article. See also the editorial by Gee and Yao in this issue.
{"title":"Privacy-ensuring Open-weights Large Language Models Are Competitive with Closed-weights GPT-4o in Extracting Chest Radiography Findings from Free-Text Reports.","authors":"Sebastian Nowak, Benjamin Wulff, Yannik C Layer, Maike Theis, Alexander Isaak, Babak Salam, Wolfgang Block, Daniel Kuetting, Claus C Pieper, Julian A Luetkens, Ulrike Attenberger, Alois M Sprinkart","doi":"10.1148/radiol.240895","DOIUrl":"https://doi.org/10.1148/radiol.240895","url":null,"abstract":"<p><p>Background Large-scale secondary use of clinical databases requires automated tools for retrospective extraction of structured content from free-text radiology reports. Purpose To share data and insights on the application of privacy-preserving open-weights large language models (LLMs) for reporting content extraction with comparison to standard rule-based systems and the closed-weights LLMs from OpenAI. Materials and Methods In this retrospective exploratory study conducted between May 2024 and September 2024, zero-shot prompting of 17 open-weights LLMs was preformed. These LLMs with model weights released under open licenses were compared with rule-based annotation and with OpenAI's GPT-4o, GPT-4o-mini, GPT-4-turbo, and GPT-3.5-turbo on a manually annotated public English chest radiography dataset (Indiana University, 3927 patients and reports). An annotated nonpublic German chest radiography dataset (18 500 reports, 16 844 patients [10 340 male; mean age, 62.6 years ± 21.5 {SD}]) was used to compare local fine-tuning of all open-weights LLMs via low-rank adaptation and 4-bit quantization to bidirectional encoder representations from transformers (BERT) with different subsets of reports (from 10 to 14 580). Nonoverlapping 95% CIs of macro-averaged F1 scores were defined as relevant differences. Results For the English reports, the highest zero-shot macro-averaged F1 score was observed for GPT-4o (92.4% [95% CI: 87.9, 95.9]); GPT-4o outperformed the rule-based CheXpert [Stanford University] (73.1% [95% CI: 65.1, 79.7]) but was comparable in performance to several open-weights LLMs (top three: Mistral-Large [Mistral AI], 92.6% [95% CI: 88.2, 96.0]; Llama-3.1-70b [Meta AI], 92.2% [95% CI: 87.1, 95.8]; and Llama-3.1-405b [Meta AI]: 90.3% [95% CI: 84.6, 94.5]). For the German reports, Mistral-Large (91.6% [95% CI: 90.5, 92.7]) had the highest zero-shot macro-averaged F1 score among the six other open-weights LLMs and outperformed the rule-based annotation (74.8% [95% CI: 73.3, 76.1]). Using 1000 reports for fine-tuning, all LLMs (top three: Mistral-Large, 94.3% [95% CI: 93.5, 95.2]; OpenBioLLM-70b [Saama]: 93.9% [95% CI: 92.9, 94.8]; and Mixtral-8×22b [Mistral AI]: 93.8% [95% CI: 92.8, 94.7]) achieved significantly higher macro-averaged F1 score than did BERT (86.7% [95% CI: 85.0, 88.3]); however, the differences were not relevant when 2000 or more reports were used for fine-tuning. Conclusion LLMs have the potential to outperform rule-based systems for zero-shot \"out-of-the-box\" structuring of report databases, with privacy-ensuring open-weights LLMs being competitive with closed-weights GPT-4o. Additionally, the open-weights LLM outperformed BERT when moderate numbers of reports were used for fine-tuning. Published under a CC BY 4.0 license. <i>Supplemental material is available for this article.</i> See also the editorial by Gee and Yao in this issue.</p>","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e240895"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142979764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Building Rome: TNM Lung Cancer Staging and an Illustration of the Scientific Method.","authors":"Carolyn Horst","doi":"10.1148/radiol.243715","DOIUrl":"https://doi.org/10.1148/radiol.243715","url":null,"abstract":"","PeriodicalId":20896,"journal":{"name":"Radiology","volume":"314 1","pages":"e243715"},"PeriodicalIF":12.1,"publicationDate":"2025-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143010652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}