Kaseba Chibwe, David Mantilla-Calderon, Fangqiong Ling
{"title":"Evaluating GPT Models for Automated Literature Screening in Wastewater-Based Epidemiology.","authors":"Kaseba Chibwe, David Mantilla-Calderon, Fangqiong Ling","doi":"10.1021/acsenvironau.4c00042","DOIUrl":null,"url":null,"abstract":"<p><p>Methods to quantitatively synthesize findings across multiple studies is an emerging need in wastewater-based epidemiology (WBE), where disease tracking through wastewater analysis is performed at broad geographical locations using various techniques to facilitate public health responses. Meta-analysis provides a rigorous statistical procedure for research synthesis, yet the manual process of screening large volumes of literature remains a hurdle for its application in timely evidence-based public health responses. Here, we evaluated the performance of GPT-3, GPT-3.5, and GPT-4 models in automated screening of publications for meta-analysis in the WBE literature. We show that the chat completion model in GPT-4 accurately differentiates papers that contain original data from those that did not with texts of the Abstract as the input at a Precision of 0.96 and Recall of 1.00, exceeding current quality standards for manual screening (Recall = 0.95) while costing less than $0.01 per paper. GPT models performed less accurately in detecting studies reporting relevant sampling location, highlighting the value of maintaining human intervention in AI-assisted literature screening. Importantly, we show that certain formulation and model choices generated nonsensical answers to the screening tasks, while others did not, urging the attention to robustness when employing AI-assisted literature screening. This study provided novel performance evaluation data on GPT models for document screening as a step in meta-analysis, suggesting AI-assisted literature screening a useful complementary technique to speed up research synthesis in WBE.</p>","PeriodicalId":29801,"journal":{"name":"ACS Environmental Au","volume":"5 1","pages":"61-68"},"PeriodicalIF":6.7000,"publicationDate":"2024-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11741058/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Environmental Au","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1021/acsenvironau.4c00042","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/15 0:00:00","PubModel":"eCollection","JCR":"Q1","JCRName":"ENGINEERING, ENVIRONMENTAL","Score":null,"Total":0}
引用次数: 0
Abstract
Methods to quantitatively synthesize findings across multiple studies is an emerging need in wastewater-based epidemiology (WBE), where disease tracking through wastewater analysis is performed at broad geographical locations using various techniques to facilitate public health responses. Meta-analysis provides a rigorous statistical procedure for research synthesis, yet the manual process of screening large volumes of literature remains a hurdle for its application in timely evidence-based public health responses. Here, we evaluated the performance of GPT-3, GPT-3.5, and GPT-4 models in automated screening of publications for meta-analysis in the WBE literature. We show that the chat completion model in GPT-4 accurately differentiates papers that contain original data from those that did not with texts of the Abstract as the input at a Precision of 0.96 and Recall of 1.00, exceeding current quality standards for manual screening (Recall = 0.95) while costing less than $0.01 per paper. GPT models performed less accurately in detecting studies reporting relevant sampling location, highlighting the value of maintaining human intervention in AI-assisted literature screening. Importantly, we show that certain formulation and model choices generated nonsensical answers to the screening tasks, while others did not, urging the attention to robustness when employing AI-assisted literature screening. This study provided novel performance evaluation data on GPT models for document screening as a step in meta-analysis, suggesting AI-assisted literature screening a useful complementary technique to speed up research synthesis in WBE.
期刊介绍:
ACS Environmental Au is an open access journal which publishes experimental research and theoretical results in all aspects of environmental science and technology both pure and applied. Short letters comprehensive articles reviews and perspectives are welcome in the following areas:Alternative EnergyAnthropogenic Impacts on Atmosphere Soil or WaterBiogeochemical CyclingBiomass or Wastes as ResourcesContaminants in Aquatic and Terrestrial EnvironmentsEnvironmental Data ScienceEcotoxicology and Public HealthEnergy and ClimateEnvironmental Modeling Processes and Measurement Methods and TechnologiesEnvironmental Nanotechnology and BiotechnologyGreen ChemistryGreen Manufacturing and EngineeringRisk assessment Regulatory Frameworks and Life-Cycle AssessmentsTreatment and Resource Recovery and Waste Management