实现随机对照试验的偏倚风险自动评估:机器人审查员与人类的比较。

IF 5 2区 生物学 Q1 MATHEMATICAL & COMPUTATIONAL BIOLOGY Research Synthesis Methods Pub Date : 2024-09-26 DOI:10.1002/jrsm.1761
Yuan Tian, Xi Yang, Suhail A. Doi, Luis Furuya-Kanamori, Lifeng Lin, Joey S. W. Kwong, Chang Xu
{"title":"实现随机对照试验的偏倚风险自动评估:机器人审查员与人类的比较。","authors":"Yuan Tian,&nbsp;Xi Yang,&nbsp;Suhail A. Doi,&nbsp;Luis Furuya-Kanamori,&nbsp;Lifeng Lin,&nbsp;Joey S. W. Kwong,&nbsp;Chang Xu","doi":"10.1002/jrsm.1761","DOIUrl":null,"url":null,"abstract":"<p>RobotReviewer is a tool for automatically assessing the risk of bias in randomized controlled trials, but there is limited evidence of its reliability. We evaluated the agreement between RobotReviewer and humans regarding the risk of bias assessment based on 1955 randomized controlled trials. The risk of bias in these trials was assessed via two different approaches: (1) manually by human reviewers, and (2) automatically by the RobotReviewer. The manual assessment was based on two groups independently, with two additional rounds of verification. The agreement between RobotReviewer and humans was measured via the concordance rate and Cohen's kappa statistics, based on the comparison of binary classification of the risk of bias (low vs. high/unclear) as restricted by RobotReviewer. The concordance rates varied by domain, ranging from 63.07% to 83.32%. Cohen's kappa statistics showed a poor agreement between humans and RobotReviewer for allocation concealment (<i>κ</i> = 0.25, 95% CI: 0.21–0.30), blinding of outcome assessors (<i>κ</i> = 0.27, 95% CI: 0.23–0.31); While moderate for random sequence generation (<i>κ</i> = 0.46, 95% CI: 0.41–0.50) and blinding of participants and personnel (<i>κ</i> = 0.59, 95% CI: 0.55–0.64). The findings demonstrate that there were domain-specific differences in the level of agreement between RobotReviewer and humans. We suggest that it might be a useful auxiliary tool, but the specific manner of its integration as a complementary tool requires further discussion.</p>","PeriodicalId":226,"journal":{"name":"Research Synthesis Methods","volume":"15 6","pages":"1111-1119"},"PeriodicalIF":5.0000,"publicationDate":"2024-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards the automatic risk of bias assessment on randomized controlled trials: A comparison of RobotReviewer and humans\",\"authors\":\"Yuan Tian,&nbsp;Xi Yang,&nbsp;Suhail A. Doi,&nbsp;Luis Furuya-Kanamori,&nbsp;Lifeng Lin,&nbsp;Joey S. W. Kwong,&nbsp;Chang Xu\",\"doi\":\"10.1002/jrsm.1761\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>RobotReviewer is a tool for automatically assessing the risk of bias in randomized controlled trials, but there is limited evidence of its reliability. We evaluated the agreement between RobotReviewer and humans regarding the risk of bias assessment based on 1955 randomized controlled trials. The risk of bias in these trials was assessed via two different approaches: (1) manually by human reviewers, and (2) automatically by the RobotReviewer. The manual assessment was based on two groups independently, with two additional rounds of verification. The agreement between RobotReviewer and humans was measured via the concordance rate and Cohen's kappa statistics, based on the comparison of binary classification of the risk of bias (low vs. high/unclear) as restricted by RobotReviewer. The concordance rates varied by domain, ranging from 63.07% to 83.32%. Cohen's kappa statistics showed a poor agreement between humans and RobotReviewer for allocation concealment (<i>κ</i> = 0.25, 95% CI: 0.21–0.30), blinding of outcome assessors (<i>κ</i> = 0.27, 95% CI: 0.23–0.31); While moderate for random sequence generation (<i>κ</i> = 0.46, 95% CI: 0.41–0.50) and blinding of participants and personnel (<i>κ</i> = 0.59, 95% CI: 0.55–0.64). The findings demonstrate that there were domain-specific differences in the level of agreement between RobotReviewer and humans. We suggest that it might be a useful auxiliary tool, but the specific manner of its integration as a complementary tool requires further discussion.</p>\",\"PeriodicalId\":226,\"journal\":{\"name\":\"Research Synthesis Methods\",\"volume\":\"15 6\",\"pages\":\"1111-1119\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2024-09-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Research Synthesis Methods\",\"FirstCategoryId\":\"99\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1002/jrsm.1761\",\"RegionNum\":2,\"RegionCategory\":\"生物学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICAL & COMPUTATIONAL BIOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Research Synthesis Methods","FirstCategoryId":"99","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/jrsm.1761","RegionNum":2,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICAL & COMPUTATIONAL BIOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

RobotReviewer 是一种自动评估随机对照试验偏倚风险的工具,但其可靠性的证据有限。我们以 1955 项随机对照试验为基础,评估了 RobotReviewer 与人类在偏倚风险评估方面的一致性。这些试验的偏倚风险通过两种不同的方法进行评估:(1) 由人类审稿人手动评估;(2) 由机器人审稿器自动评估。人工评估由两组人员独立进行,并额外进行两轮验证。机器人审稿器和人类之间的一致性是通过一致率和科恩卡帕统计来衡量的,基于机器人审稿器限制的偏倚风险二元分类(低与高/不明确)的比较。不同领域的一致率各不相同,从 63.07% 到 83.32% 不等。Cohen's kappa 统计显示,人类与 RobotReviewer 在分配隐藏(κ = 0.25,95% CI:0.21-0.30)、结果评估者盲法(κ = 0.27,95% CI:0.23-0.31)方面的一致性较差;而在随机序列生成(κ = 0.46,95% CI:0.41-0.50)以及参与者和人员盲法(κ = 0.59,95% CI:0.55-0.64)方面的一致性适中。研究结果表明,RobotReviewer 与人类在特定领域的一致性水平存在差异。我们认为,它可能是一个有用的辅助工具,但其作为补充工具的具体整合方式还需要进一步讨论。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Towards the automatic risk of bias assessment on randomized controlled trials: A comparison of RobotReviewer and humans

RobotReviewer is a tool for automatically assessing the risk of bias in randomized controlled trials, but there is limited evidence of its reliability. We evaluated the agreement between RobotReviewer and humans regarding the risk of bias assessment based on 1955 randomized controlled trials. The risk of bias in these trials was assessed via two different approaches: (1) manually by human reviewers, and (2) automatically by the RobotReviewer. The manual assessment was based on two groups independently, with two additional rounds of verification. The agreement between RobotReviewer and humans was measured via the concordance rate and Cohen's kappa statistics, based on the comparison of binary classification of the risk of bias (low vs. high/unclear) as restricted by RobotReviewer. The concordance rates varied by domain, ranging from 63.07% to 83.32%. Cohen's kappa statistics showed a poor agreement between humans and RobotReviewer for allocation concealment (κ = 0.25, 95% CI: 0.21–0.30), blinding of outcome assessors (κ = 0.27, 95% CI: 0.23–0.31); While moderate for random sequence generation (κ = 0.46, 95% CI: 0.41–0.50) and blinding of participants and personnel (κ = 0.59, 95% CI: 0.55–0.64). The findings demonstrate that there were domain-specific differences in the level of agreement between RobotReviewer and humans. We suggest that it might be a useful auxiliary tool, but the specific manner of its integration as a complementary tool requires further discussion.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Research Synthesis Methods
Research Synthesis Methods MATHEMATICAL & COMPUTATIONAL BIOLOGYMULTID-MULTIDISCIPLINARY SCIENCES
CiteScore
16.90
自引率
3.10%
发文量
75
期刊介绍: Research Synthesis Methods is a reputable, peer-reviewed journal that focuses on the development and dissemination of methods for conducting systematic research synthesis. Our aim is to advance the knowledge and application of research synthesis methods across various disciplines. Our journal provides a platform for the exchange of ideas and knowledge related to designing, conducting, analyzing, interpreting, reporting, and applying research synthesis. While research synthesis is commonly practiced in the health and social sciences, our journal also welcomes contributions from other fields to enrich the methodologies employed in research synthesis across scientific disciplines. By bridging different disciplines, we aim to foster collaboration and cross-fertilization of ideas, ultimately enhancing the quality and effectiveness of research synthesis methods. Whether you are a researcher, practitioner, or stakeholder involved in research synthesis, our journal strives to offer valuable insights and practical guidance for your work.
期刊最新文献
Issue Information A tutorial on aggregating evidence from conceptual replication studies using the product Bayes factor Evolving use of the Cochrane Risk of Bias 2 tool in biomedical systematic reviews Exploring methodological approaches used in network meta-analysis of psychological interventions: A scoping review An evaluation of the performance of stopping rules in AI-aided screening for psychological meta-analytical research
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1