{"title":"Automated question type coding of forensic interviews and trial testimony in child sexual abuse cases.","authors":"Zsofia A Szojka, Suvimal Yashraj, Thomas D Lyon","doi":"10.1037/lhb0000590","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>Question-type classification is widely used as a measure of interview quality. However, question-type coding is a time-consuming process when performed by manual coders. Reliable automated question-type coding approaches would facilitate the assessment of the quality of forensic interviews and court testimony involving victims of child abuse.</p><p><strong>Hypotheses: </strong>We expected that the reliability achieved by the automated model would be comparable to manual coders.</p><p><strong>Method: </strong>We examined whether a large language model (Robustly Optimized Bidirectional Encoder Representations from Transformers Approach) trained on questions (<i>N</i> = 351,920) asked in forensic interviews (<i>n</i> = 1,435) and trial testimony (<i>n</i> = 416) involving 3- to 17-year-old alleged victims of child sexual abuse could distinguish among (a) invitations, (b) wh-questions, (c) option-posing questions, and (d) nonquestions.</p><p><strong>Results: </strong>The model achieved high reliability (95% agreement; κ = .93). To determine whether disagreements were due to machine or manual errors, we recoded inconsistencies between the machine and manual codes. Manual coders erred more often than the machine, particularly by overlooking invitations and nonquestions. Correcting errors in the manual codes further increased the model's reliability (98% agreement; κ = .97).</p><p><strong>Conclusions: </strong>Automated question-type coding can provide a time-efficient and highly accurate alternative to manual coding. We have made the trained model publicly available for use by researchers and practitioners. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":48230,"journal":{"name":"Law and Human Behavior","volume":" ","pages":""},"PeriodicalIF":2.4000,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Law and Human Behavior","FirstCategoryId":"90","ListUrlMain":"https://doi.org/10.1037/lhb0000590","RegionNum":2,"RegionCategory":"社会学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"LAW","Score":null,"Total":0}
引用次数: 0
Abstract
Objective: Question-type classification is widely used as a measure of interview quality. However, question-type coding is a time-consuming process when performed by manual coders. Reliable automated question-type coding approaches would facilitate the assessment of the quality of forensic interviews and court testimony involving victims of child abuse.
Hypotheses: We expected that the reliability achieved by the automated model would be comparable to manual coders.
Method: We examined whether a large language model (Robustly Optimized Bidirectional Encoder Representations from Transformers Approach) trained on questions (N = 351,920) asked in forensic interviews (n = 1,435) and trial testimony (n = 416) involving 3- to 17-year-old alleged victims of child sexual abuse could distinguish among (a) invitations, (b) wh-questions, (c) option-posing questions, and (d) nonquestions.
Results: The model achieved high reliability (95% agreement; κ = .93). To determine whether disagreements were due to machine or manual errors, we recoded inconsistencies between the machine and manual codes. Manual coders erred more often than the machine, particularly by overlooking invitations and nonquestions. Correcting errors in the manual codes further increased the model's reliability (98% agreement; κ = .97).
Conclusions: Automated question-type coding can provide a time-efficient and highly accurate alternative to manual coding. We have made the trained model publicly available for use by researchers and practitioners. (PsycInfo Database Record (c) 2025 APA, all rights reserved).
期刊介绍:
Law and Human Behavior, the official journal of the American Psychology-Law Society/Division 41 of the American Psychological Association, is a multidisciplinary forum for the publication of articles and discussions of issues arising out of the relationships between human behavior and the law, our legal system, and the legal process. This journal publishes original research, reviews of past research, and theoretical studies from professionals in criminal justice, law, psychology, sociology, psychiatry, political science, education, communication, and other areas germane to the field.