Benjamin J McCloskey, Bruce A Cox, L. Champagne, Trevor J. Bihl
{"title":"Benefits of using blended generative adversarial network images to augment classification model training data sets","authors":"Benjamin J McCloskey, Bruce A Cox, L. Champagne, Trevor J. Bihl","doi":"10.1177/15485129231170225","DOIUrl":null,"url":null,"abstract":"Object detection algorithms have reached nearly superhuman levels within the last decade; however, these algorithms require large diverse training data sets to ensure their operational performance matches performance demonstrated during testing. The collection and human labeling of such data sets can be expensive and, in some cases, such as Intelligence, Surveillance and Reconnaissance of rare events it may not even be feasible. This research proposes a novel method for creating additional variability within the training data set by utilizing multiple models of generative adversarial networks producing both high- and low-quality synthetic images of vehicles and inserting those images alongside images of real vehicles into real backgrounds. This research demonstrates a 17.90% increase in mean absolute percentage error, on average, compared to the YOLOv4-Tiny Model trained on the original non-augmented training set as well as a 14.44% average improvement in the average intersection over union rate. In addition, our research adds to a small, but growing, body of literature indicating that the inclusion of low-quality images into training data sets is beneficial to the performance of computer vision models.","PeriodicalId":44661,"journal":{"name":"Journal of Defense Modeling and Simulation-Applications Methodology Technology-JDMS","volume":null,"pages":null},"PeriodicalIF":1.0000,"publicationDate":"2023-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Defense Modeling and Simulation-Applications Methodology Technology-JDMS","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1177/15485129231170225","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
Object detection algorithms have reached nearly superhuman levels within the last decade; however, these algorithms require large diverse training data sets to ensure their operational performance matches performance demonstrated during testing. The collection and human labeling of such data sets can be expensive and, in some cases, such as Intelligence, Surveillance and Reconnaissance of rare events it may not even be feasible. This research proposes a novel method for creating additional variability within the training data set by utilizing multiple models of generative adversarial networks producing both high- and low-quality synthetic images of vehicles and inserting those images alongside images of real vehicles into real backgrounds. This research demonstrates a 17.90% increase in mean absolute percentage error, on average, compared to the YOLOv4-Tiny Model trained on the original non-augmented training set as well as a 14.44% average improvement in the average intersection over union rate. In addition, our research adds to a small, but growing, body of literature indicating that the inclusion of low-quality images into training data sets is beneficial to the performance of computer vision models.