Fine-Tuning Large Language Models for Specialized Use Cases

D.M. Anisuzzaman PhD, Jeffrey G. Malins PhD, Paul A. Friedman MD, Zachi I. Attia PhD
{"title":"Fine-Tuning Large Language Models for Specialized Use Cases","authors":"D.M. Anisuzzaman PhD,&nbsp;Jeffrey G. Malins PhD,&nbsp;Paul A. Friedman MD,&nbsp;Zachi I. Attia PhD","doi":"10.1016/j.mcpdig.2024.11.005","DOIUrl":null,"url":null,"abstract":"<div><div>Large language models (LLMs) are a type of artificial intelligence, which operate by predicting and assembling sequences of words that are statistically likely to follow from a given text input. With this basic ability, LLMs are able to answer complex questions and follow extremely complex instructions. Products created using LLMs such as ChatGPT by OpenAI and Claude by Anthropic have created a huge amount of traction and user engagements and revolutionized the way we interact with technology, bringing a new dimension to human-computer interaction. Fine-tuning is a process in which a pretrained model, such as an LLM, is further trained on a custom data set to adapt it for specialized tasks or domains. In this review, we outline some of the major methodologic approaches and techniques that can be used to fine-tune LLMs for specialized use cases and enumerate the general steps required for carrying out LLM fine-tuning. We then illustrate a few of these methodologic approaches by describing several specific use cases of fine-tuning LLMs across medical subspecialties. Finally, we close with a consideration of some of the benefits and limitations associated with fine-tuning LLMs for specialized use cases, with an emphasis on specific concerns in the field of medicine.</div></div>","PeriodicalId":74127,"journal":{"name":"Mayo Clinic Proceedings. Digital health","volume":"3 1","pages":"Article 100184"},"PeriodicalIF":0.0000,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Mayo Clinic Proceedings. Digital health","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949761224001147","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/11/29 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Large language models (LLMs) are a type of artificial intelligence, which operate by predicting and assembling sequences of words that are statistically likely to follow from a given text input. With this basic ability, LLMs are able to answer complex questions and follow extremely complex instructions. Products created using LLMs such as ChatGPT by OpenAI and Claude by Anthropic have created a huge amount of traction and user engagements and revolutionized the way we interact with technology, bringing a new dimension to human-computer interaction. Fine-tuning is a process in which a pretrained model, such as an LLM, is further trained on a custom data set to adapt it for specialized tasks or domains. In this review, we outline some of the major methodologic approaches and techniques that can be used to fine-tune LLMs for specialized use cases and enumerate the general steps required for carrying out LLM fine-tuning. We then illustrate a few of these methodologic approaches by describing several specific use cases of fine-tuning LLMs across medical subspecialties. Finally, we close with a consideration of some of the benefits and limitations associated with fine-tuning LLMs for specialized use cases, with an emphasis on specific concerns in the field of medicine.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
为专门用例微调大型语言模型
大型语言模型(llm)是一种人工智能,它通过预测和组装统计上可能从给定文本输入中跟随的单词序列来操作。有了这个基本能力,法学硕士能够回答复杂的问题,并遵循极其复杂的指令。使用llm创建的产品,如OpenAI的ChatGPT和Anthropic的Claude,创造了巨大的吸引力和用户参与,并彻底改变了我们与技术交互的方式,为人机交互带来了一个新的维度。微调是一个过程,在这个过程中,预先训练的模型(如LLM)在自定义数据集上进一步训练,以使其适应专门的任务或领域。在这篇综述中,我们概述了一些主要的方法方法和技术,这些方法和技术可用于针对特定用例对LLM进行微调,并列举了执行LLM微调所需的一般步骤。然后,我们通过描述跨医学亚专科微调法学硕士的几个具体用例来说明其中一些方法方法。最后,我们以考虑与针对专门用例进行微调的llm相关的一些好处和限制作为结束,重点是医学领域的具体问题。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Mayo Clinic Proceedings. Digital health
Mayo Clinic Proceedings. Digital health Medicine and Dentistry (General), Health Informatics, Public Health and Health Policy
自引率
0.00%
发文量
0
审稿时长
47 days
期刊最新文献
Early Implications for Solid Organ Transplantation With the Use of Artificial Intelligence From a Bibliometric Perspective Wearable Sleep Measures May Improve Machine Learning Prediction of Home-Based Pulmonary Rehabilitation Engagement Among Patients With Chronic Obstructive Pulmonary Disease: A Proof-of-Concept Study Patients’ Perspectives on Telemedicine for Type 2 Diabetes in Primary Care: Systematic Review and Thematic Synthesis Prediction of Heart Failure With Reduced Ejection Fraction With Artificial Intelligence Electrocardiography in Patients With Atrial Fibrillation Criterion Validation for Step Count and Distance Covered During the 6-Minute Walk Test Using a Low-Cost, Commercially Available Activity Monitor
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1