Zac E Imel, Brian Pace, Brad Pendergraft, Jordan Pruett, Michael Tanana, Christina S Soma, Kate A Comtois, David C Atkins
{"title":"Machine Learning-Based Evaluation of Suicide Risk Assessment in Crisis Counseling Calls.","authors":"Zac E Imel, Brian Pace, Brad Pendergraft, Jordan Pruett, Michael Tanana, Christina S Soma, Kate A Comtois, David C Atkins","doi":"10.1176/appi.ps.20230648","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>Counselor assessment of suicide risk is one key component of crisis counseling, and standards require risk assessment in every crisis counseling conversation. Efforts to increase risk assessment frequency are limited by quality improvement tools that rely on human evaluation of conversations, which is labor intensive, slow, and impossible to scale. Advances in machine learning (ML) have made possible the development of tools that can automatically and immediately detect the presence of risk assessment in crisis counseling conversations.</p><p><strong>Methods: </strong>To train models, a coding team labeled every statement in 476 crisis counseling calls (193,257 statements) for a core element of risk assessment. The authors then fine-tuned a transformer-based ML model with the labeled data, utilizing separate training, validation, and test data sets.</p><p><strong>Results: </strong>Generally, the evaluated ML model was highly consistent with human raters. For detecting any risk assessment, ML model agreement with human ratings was 98% of human interrater agreement. Across specific labels, average F1 (the harmonic mean of precision and recall) was 0.86 at the call level and 0.66 at the statement level and often varied as a result of a low base rate for some risk labels.</p><p><strong>Conclusions: </strong>ML models can reliably detect the presence of suicide risk assessment in crisis counseling conversations, presenting an opportunity to scale quality improvement efforts.</p>","PeriodicalId":20878,"journal":{"name":"Psychiatric services","volume":" ","pages":"1068-1074"},"PeriodicalIF":3.3000,"publicationDate":"2024-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11530329/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Psychiatric services","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1176/appi.ps.20230648","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/7/19 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"HEALTH POLICY & SERVICES","Score":null,"Total":0}
引用次数: 0
Abstract
Objective: Counselor assessment of suicide risk is one key component of crisis counseling, and standards require risk assessment in every crisis counseling conversation. Efforts to increase risk assessment frequency are limited by quality improvement tools that rely on human evaluation of conversations, which is labor intensive, slow, and impossible to scale. Advances in machine learning (ML) have made possible the development of tools that can automatically and immediately detect the presence of risk assessment in crisis counseling conversations.
Methods: To train models, a coding team labeled every statement in 476 crisis counseling calls (193,257 statements) for a core element of risk assessment. The authors then fine-tuned a transformer-based ML model with the labeled data, utilizing separate training, validation, and test data sets.
Results: Generally, the evaluated ML model was highly consistent with human raters. For detecting any risk assessment, ML model agreement with human ratings was 98% of human interrater agreement. Across specific labels, average F1 (the harmonic mean of precision and recall) was 0.86 at the call level and 0.66 at the statement level and often varied as a result of a low base rate for some risk labels.
Conclusions: ML models can reliably detect the presence of suicide risk assessment in crisis counseling conversations, presenting an opportunity to scale quality improvement efforts.
期刊介绍:
Psychiatric Services, established in 1950, is published monthly by the American Psychiatric Association. The peer-reviewed journal features research reports on issues related to the delivery of mental health services, especially for people with serious mental illness in community-based treatment programs. Long known as an interdisciplinary journal, Psychiatric Services recognizes that provision of high-quality care involves collaboration among a variety of professionals, frequently working as a team. Authors of research reports published in the journal include psychiatrists, psychologists, pharmacists, nurses, social workers, drug and alcohol treatment counselors, economists, policy analysts, and professionals in related systems such as criminal justice and welfare systems. In the mental health field, the current focus on patient-centered, recovery-oriented care and on dissemination of evidence-based practices is transforming service delivery systems at all levels. Research published in Psychiatric Services contributes to this transformation.