Shan Zhang, Chris Palaguachi, Marcin Pitera, Chris Davis Jaldi, Noah L. Schroeder, Anthony F. Botelho, Jessica R. Gladstone
{"title":"Semi-automating the Scoping Review Process: Is it Worthwhile? A Methodological Evaluation","authors":"Shan Zhang, Chris Palaguachi, Marcin Pitera, Chris Davis Jaldi, Noah L. Schroeder, Anthony F. Botelho, Jessica R. Gladstone","doi":"10.1007/s10648-024-09972-0","DOIUrl":null,"url":null,"abstract":"<p>Systematic reviews are a time-consuming yet effective approach to understanding research trends. While researchers have investigated how to speed up the process of screening studies for potential inclusion, few have focused on to what extent we can use algorithms to extract data instead of human coders. In this study, we explore to what extent analyses and algorithms can produce results similar to human data extraction during a scoping review—a type of systematic review aimed at understanding the nature of the field rather than the efficacy of an intervention—in the context of a never before analyzed sample of studies that were intended for a scoping review. Specifically, we tested five approaches: bibliometric analysis with VOSviewer, latent Dirichlet allocation (LDA) with bag of words, <i>k</i>-means clustering with TF-IDF, Sentence-BERT, or SPECTER, hierarchical clustering with Sentence-BERT, and BERTopic. Our results showed that topic modeling approaches (LDA/BERTopic) and <i>k</i>-means clustering identified specific, but often narrow research areas, leaving a substantial portion of the sample unclassified or in unclear topics. Meanwhile, bibliometric analysis and hierarchical clustering with SBERT were more informative for our purposes, identifying key author networks and categorizing studies into distinct themes as well as reflecting the relationships between themes, respectively. Overall, we highlight the capabilities and limitations of each method and discuss how these techniques can complement traditional human data extraction methods. We conclude that the analyses tested here likely cannot fully replace human data extraction in scoping reviews but serve as valuable supplements.</p>","PeriodicalId":48344,"journal":{"name":"Educational Psychology Review","volume":"80 1","pages":""},"PeriodicalIF":10.1000,"publicationDate":"2024-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Educational Psychology Review","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1007/s10648-024-09972-0","RegionNum":1,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, EDUCATIONAL","Score":null,"Total":0}
引用次数: 0
Abstract
Systematic reviews are a time-consuming yet effective approach to understanding research trends. While researchers have investigated how to speed up the process of screening studies for potential inclusion, few have focused on to what extent we can use algorithms to extract data instead of human coders. In this study, we explore to what extent analyses and algorithms can produce results similar to human data extraction during a scoping review—a type of systematic review aimed at understanding the nature of the field rather than the efficacy of an intervention—in the context of a never before analyzed sample of studies that were intended for a scoping review. Specifically, we tested five approaches: bibliometric analysis with VOSviewer, latent Dirichlet allocation (LDA) with bag of words, k-means clustering with TF-IDF, Sentence-BERT, or SPECTER, hierarchical clustering with Sentence-BERT, and BERTopic. Our results showed that topic modeling approaches (LDA/BERTopic) and k-means clustering identified specific, but often narrow research areas, leaving a substantial portion of the sample unclassified or in unclear topics. Meanwhile, bibliometric analysis and hierarchical clustering with SBERT were more informative for our purposes, identifying key author networks and categorizing studies into distinct themes as well as reflecting the relationships between themes, respectively. Overall, we highlight the capabilities and limitations of each method and discuss how these techniques can complement traditional human data extraction methods. We conclude that the analyses tested here likely cannot fully replace human data extraction in scoping reviews but serve as valuable supplements.
期刊介绍:
Educational Psychology Review aims to disseminate knowledge and promote dialogue within the field of educational psychology. It serves as a platform for the publication of various types of articles, including peer-reviewed integrative reviews, special thematic issues, reflections on previous research or new research directions, interviews, and research-based advice for practitioners. The journal caters to a diverse readership, ranging from generalists in educational psychology to experts in specific areas of the discipline. The content offers a comprehensive coverage of topics and provides in-depth information to meet the needs of both specialized researchers and practitioners.