Daniil Filienko, Yinzhou Wang, Caroline El Jazmi, Serena Xie, Trevor Cohen, Martine De Cock, Weichao Yuwen
{"title":"Toward Large Language Models as a Therapeutic Tool: Comparing Prompting Techniques to Improve GPT-Delivered Problem-Solving Therapy","authors":"Daniil Filienko, Yinzhou Wang, Caroline El Jazmi, Serena Xie, Trevor Cohen, Martine De Cock, Weichao Yuwen","doi":"arxiv-2409.00112","DOIUrl":null,"url":null,"abstract":"While Large Language Models (LLMs) are being quickly adapted to many domains,\nincluding healthcare, their strengths and pitfalls remain under-explored. In\nour study, we examine the effects of prompt engineering to guide Large Language\nModels (LLMs) in delivering parts of a Problem-Solving Therapy (PST) session\nvia text, particularly during the symptom identification and assessment phase\nfor personalized goal setting. We present evaluation results of the models'\nperformances by automatic metrics and experienced medical professionals. We\ndemonstrate that the models' capability to deliver protocolized therapy can be\nimproved with the proper use of prompt engineering methods, albeit with\nlimitations. To our knowledge, this study is among the first to assess the\neffects of various prompting techniques in enhancing a generalist model's\nability to deliver psychotherapy, focusing on overall quality, consistency, and\nempathy. Exploring LLMs' potential in delivering psychotherapy holds promise\nwith the current shortage of mental health professionals amid significant\nneeds, enhancing the potential utility of AI-based and AI-enhanced care\nservices.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"410 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Emerging Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.00112","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
While Large Language Models (LLMs) are being quickly adapted to many domains,
including healthcare, their strengths and pitfalls remain under-explored. In
our study, we examine the effects of prompt engineering to guide Large Language
Models (LLMs) in delivering parts of a Problem-Solving Therapy (PST) session
via text, particularly during the symptom identification and assessment phase
for personalized goal setting. We present evaluation results of the models'
performances by automatic metrics and experienced medical professionals. We
demonstrate that the models' capability to deliver protocolized therapy can be
improved with the proper use of prompt engineering methods, albeit with
limitations. To our knowledge, this study is among the first to assess the
effects of various prompting techniques in enhancing a generalist model's
ability to deliver psychotherapy, focusing on overall quality, consistency, and
empathy. Exploring LLMs' potential in delivering psychotherapy holds promise
with the current shortage of mental health professionals amid significant
needs, enhancing the potential utility of AI-based and AI-enhanced care
services.