Automated question type coding of forensic interviews and trial testimony in child sexual abuse cases.

IF 2.4 2区 社会学 Q1 LAW Law and Human Behavior Pub Date : 2025-03-20 DOI:10.1037/lhb0000590
Zsofia A Szojka, Suvimal Yashraj, Thomas D Lyon
{"title":"Automated question type coding of forensic interviews and trial testimony in child sexual abuse cases.","authors":"Zsofia A Szojka, Suvimal Yashraj, Thomas D Lyon","doi":"10.1037/lhb0000590","DOIUrl":null,"url":null,"abstract":"<p><strong>Objective: </strong>Question-type classification is widely used as a measure of interview quality. However, question-type coding is a time-consuming process when performed by manual coders. Reliable automated question-type coding approaches would facilitate the assessment of the quality of forensic interviews and court testimony involving victims of child abuse.</p><p><strong>Hypotheses: </strong>We expected that the reliability achieved by the automated model would be comparable to manual coders.</p><p><strong>Method: </strong>We examined whether a large language model (Robustly Optimized Bidirectional Encoder Representations from Transformers Approach) trained on questions (<i>N</i> = 351,920) asked in forensic interviews (<i>n</i> = 1,435) and trial testimony (<i>n</i> = 416) involving 3- to 17-year-old alleged victims of child sexual abuse could distinguish among (a) invitations, (b) wh-questions, (c) option-posing questions, and (d) nonquestions.</p><p><strong>Results: </strong>The model achieved high reliability (95% agreement; κ = .93). To determine whether disagreements were due to machine or manual errors, we recoded inconsistencies between the machine and manual codes. Manual coders erred more often than the machine, particularly by overlooking invitations and nonquestions. Correcting errors in the manual codes further increased the model's reliability (98% agreement; κ = .97).</p><p><strong>Conclusions: </strong>Automated question-type coding can provide a time-efficient and highly accurate alternative to manual coding. We have made the trained model publicly available for use by researchers and practitioners. (PsycInfo Database Record (c) 2025 APA, all rights reserved).</p>","PeriodicalId":48230,"journal":{"name":"Law and Human Behavior","volume":" ","pages":""},"PeriodicalIF":2.4000,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Law and Human Behavior","FirstCategoryId":"90","ListUrlMain":"https://doi.org/10.1037/lhb0000590","RegionNum":2,"RegionCategory":"社会学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"LAW","Score":null,"Total":0}
引用次数: 0

Abstract

Objective: Question-type classification is widely used as a measure of interview quality. However, question-type coding is a time-consuming process when performed by manual coders. Reliable automated question-type coding approaches would facilitate the assessment of the quality of forensic interviews and court testimony involving victims of child abuse.

Hypotheses: We expected that the reliability achieved by the automated model would be comparable to manual coders.

Method: We examined whether a large language model (Robustly Optimized Bidirectional Encoder Representations from Transformers Approach) trained on questions (N = 351,920) asked in forensic interviews (n = 1,435) and trial testimony (n = 416) involving 3- to 17-year-old alleged victims of child sexual abuse could distinguish among (a) invitations, (b) wh-questions, (c) option-posing questions, and (d) nonquestions.

Results: The model achieved high reliability (95% agreement; κ = .93). To determine whether disagreements were due to machine or manual errors, we recoded inconsistencies between the machine and manual codes. Manual coders erred more often than the machine, particularly by overlooking invitations and nonquestions. Correcting errors in the manual codes further increased the model's reliability (98% agreement; κ = .97).

Conclusions: Automated question-type coding can provide a time-efficient and highly accurate alternative to manual coding. We have made the trained model publicly available for use by researchers and practitioners. (PsycInfo Database Record (c) 2025 APA, all rights reserved).

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
4.50
自引率
8.00%
发文量
42
期刊介绍: Law and Human Behavior, the official journal of the American Psychology-Law Society/Division 41 of the American Psychological Association, is a multidisciplinary forum for the publication of articles and discussions of issues arising out of the relationships between human behavior and the law, our legal system, and the legal process. This journal publishes original research, reviews of past research, and theoretical studies from professionals in criminal justice, law, psychology, sociology, psychiatry, political science, education, communication, and other areas germane to the field.
期刊最新文献
Automated question type coding of forensic interviews and trial testimony in child sexual abuse cases. Who questions the legitimacy of law? A latent profile analysis using national data in China. Police-induced confessions, 2.0: Risk factors and recommendations. What risk assessment tools can be used with men convicted of child sexual exploitation material offenses? Recommendations from a review of current research. Law and Human Behavior: Status update and new initiatives.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1