Brihat Sharma, Yanjun Gao, Timothy Miller, Matthew M Churpek, Majid Afshar, Dmitriy Dligach
{"title":"Multi-Task Training with In-Domain Language Models for Diagnostic Reasoning.","authors":"Brihat Sharma, Yanjun Gao, Timothy Miller, Matthew M Churpek, Majid Afshar, Dmitriy Dligach","doi":"","DOIUrl":null,"url":null,"abstract":"<p><p>Generative artificial intelligence (AI) is a promising direction for augmenting clinical diagnostic decision support and reducing diagnostic errors, a leading contributor to medical errors. To further the development of clinical AI systems, the Diagnostic Reasoning Benchmark (DR.BENCH) was introduced as a comprehensive generative AI framework, comprised of six tasks representing key components in clinical reasoning. We present a comparative analysis of in-domain versus out-of-domain language models as well as multi-task versus single task training with a focus on the problem summarization task in DR.BENCH (Gao et al., 2023). We demonstrate that a multi-task, clinically-trained language model outperforms its general domain counterpart by a large margin, establishing a new state-of-the-art performance, with a ROUGE-L score of 28.55. This research underscores the value of domain-specific training for optimizing clinical diagnostic reasoning tasks.</p>","PeriodicalId":74541,"journal":{"name":"Proceedings of the conference. Association for Computational Linguistics. Meeting","volume":"2023 ClinicalNLP","pages":"78-85"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10368094/pdf/nihms-1917256.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the conference. Association for Computational Linguistics. Meeting","FirstCategoryId":"1085","ListUrlMain":"","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Generative artificial intelligence (AI) is a promising direction for augmenting clinical diagnostic decision support and reducing diagnostic errors, a leading contributor to medical errors. To further the development of clinical AI systems, the Diagnostic Reasoning Benchmark (DR.BENCH) was introduced as a comprehensive generative AI framework, comprised of six tasks representing key components in clinical reasoning. We present a comparative analysis of in-domain versus out-of-domain language models as well as multi-task versus single task training with a focus on the problem summarization task in DR.BENCH (Gao et al., 2023). We demonstrate that a multi-task, clinically-trained language model outperforms its general domain counterpart by a large margin, establishing a new state-of-the-art performance, with a ROUGE-L score of 28.55. This research underscores the value of domain-specific training for optimizing clinical diagnostic reasoning tasks.
生成式人工智能(AI)是增强临床诊断决策支持和减少诊断错误的一个有前途的方向,诊断错误是导致医疗错误的主要原因。为了进一步发展临床人工智能系统,引入了诊断推理基准(DR.BENCH)作为一个全面的生成式人工智能框架,由代表临床推理关键组件的六个任务组成。我们对领域内和领域外的语言模型以及多任务和单任务训练进行了比较分析,重点是DR.BENCH中的问题总结任务(Gao et al., 2023)。我们证明了一个多任务、临床训练的语言模型在很大程度上优于其一般领域的对应模型,建立了一个新的最先进的性能,ROUGE-L得分为28.55。这项研究强调了领域特定训练对优化临床诊断推理任务的价值。