{"title":"Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review.","authors":"Dipak Gautam, Philipp Kellmeyer","doi":"10.2196/62865","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>The rapid evolution of large language models (LLMs), such as Bidirectional Encoder Representations from Transformers (BERT; Google) and GPT (OpenAI), has introduced significant advancements in natural language processing. These models are increasingly integrated into various applications, including mental health support. However, the credibility of LLMs in providing reliable and explainable mental health information and support remains underexplored.</p><p><strong>Objective: </strong>This scoping review systematically maps the factors influencing the credibility of LLMs in mental health support, including reliability, explainability, and ethical considerations. The review is expected to offer critical insights for practitioners, researchers, and policy makers, guiding future research and policy development. These findings will contribute to the responsible integration of LLMs into mental health care, with a focus on maintaining ethical standards and user trust.</p><p><strong>Methods: </strong>This review follows PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines and the Joanna Briggs Institute (JBI) methodology. Eligibility criteria include studies that apply transformer-based generative language models in mental health support, such as BERT and GPT. Sources include PsycINFO, MEDLINE via PubMed, Web of Science, IEEE Xplore, and ACM Digital Library. A systematic search of studies from 2019 onward will be conducted and updated until October 2024. Data will be synthesized qualitatively. The Population, Concept, and Context framework will guide the inclusion criteria. Two independent reviewers will screen and extract data, resolving discrepancies through discussion. Data will be synthesized and presented descriptively.</p><p><strong>Results: </strong>As of September 2024, this study is currently in progress, with the systematic search completed and the screening phase ongoing. We expect to complete data extraction by early November 2024 and synthesis by late November 2024.</p><p><strong>Conclusions: </strong>This scoping review will map the current evidence on the credibility of LLMs in mental health support. It will identify factors influencing the reliability, explainability, and ethical considerations of these models, providing insights for practitioners, researchers, policy makers, and users. These findings will fill a critical gap in the literature and inform future research, practice, and policy development, ensuring the responsible integration of LLMs in mental health services.</p><p><strong>International registered report identifier (irrid): </strong>DERR1-10.2196/62865.</p>","PeriodicalId":14755,"journal":{"name":"JMIR Research Protocols","volume":"14 ","pages":"e62865"},"PeriodicalIF":1.5000,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11822324/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"JMIR Research Protocols","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2196/62865","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
引用次数: 0
Abstract
Background: The rapid evolution of large language models (LLMs), such as Bidirectional Encoder Representations from Transformers (BERT; Google) and GPT (OpenAI), has introduced significant advancements in natural language processing. These models are increasingly integrated into various applications, including mental health support. However, the credibility of LLMs in providing reliable and explainable mental health information and support remains underexplored.
Objective: This scoping review systematically maps the factors influencing the credibility of LLMs in mental health support, including reliability, explainability, and ethical considerations. The review is expected to offer critical insights for practitioners, researchers, and policy makers, guiding future research and policy development. These findings will contribute to the responsible integration of LLMs into mental health care, with a focus on maintaining ethical standards and user trust.
Methods: This review follows PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines and the Joanna Briggs Institute (JBI) methodology. Eligibility criteria include studies that apply transformer-based generative language models in mental health support, such as BERT and GPT. Sources include PsycINFO, MEDLINE via PubMed, Web of Science, IEEE Xplore, and ACM Digital Library. A systematic search of studies from 2019 onward will be conducted and updated until October 2024. Data will be synthesized qualitatively. The Population, Concept, and Context framework will guide the inclusion criteria. Two independent reviewers will screen and extract data, resolving discrepancies through discussion. Data will be synthesized and presented descriptively.
Results: As of September 2024, this study is currently in progress, with the systematic search completed and the screening phase ongoing. We expect to complete data extraction by early November 2024 and synthesis by late November 2024.
Conclusions: This scoping review will map the current evidence on the credibility of LLMs in mental health support. It will identify factors influencing the reliability, explainability, and ethical considerations of these models, providing insights for practitioners, researchers, policy makers, and users. These findings will fill a critical gap in the literature and inform future research, practice, and policy development, ensuring the responsible integration of LLMs in mental health services.
International registered report identifier (irrid): DERR1-10.2196/62865.
背景:大型语言模型(llm)的快速发展,如变形金刚的双向编码器表示(BERT;b谷歌)和GPT (OpenAI),在自然语言处理方面取得了重大进展。这些模型越来越多地集成到各种应用中,包括心理健康支持。然而,法学硕士在提供可靠和可解释的心理健康信息和支持方面的可信度仍未得到充分探讨。目的:本范围综述系统地绘制了影响法学硕士在心理健康支持方面可信度的因素,包括可靠性、可解释性和伦理考虑。该综述有望为从业者、研究人员和政策制定者提供重要见解,指导未来的研究和政策制定。这些发现将有助于负责任地将法学硕士纳入精神卫生保健,重点是保持道德标准和用户信任。方法:本综述遵循PRISMA-ScR(系统评价和荟萃分析扩展范围评价的首选报告项目)指南和乔安娜布里格斯研究所(JBI)的方法。资格标准包括在心理健康支持中应用基于转换的生成语言模型的研究,如BERT和GPT。来源包括PsycINFO, MEDLINE通过PubMed, Web of Science, IEEE explore和ACM数字图书馆。将对2019年以后的研究进行系统检索,并更新到2024年10月。数据将进行定性合成。人口、概念和上下文框架将指导纳入标准。两名独立的审稿人将筛选和提取数据,通过讨论解决差异。数据将被综合并描述性地呈现。结果:截至2024年9月,本研究正在进行中,系统检索已完成,筛选阶段正在进行中。我们预计在2024年11月初完成数据提取,在2024年11月底完成数据合成。结论:这一范围综述将绘制出llm在心理健康支持方面可信度的现有证据。它将确定影响这些模型的可靠性、可解释性和伦理考虑的因素,为从业者、研究人员、政策制定者和用户提供见解。这些发现将填补文献中的关键空白,并为未来的研究、实践和政策制定提供信息,确保法学硕士在精神卫生服务中的负责任整合。国际注册报告标识符(irrid): DERR1-10.2196/62865。