Jonathan Sato, Chelsea Mediavilla, C. Ward, S. Parameswaran
{"title":"SRC3: A Video Dataset for Evaluating Domain Mismatch","authors":"Jonathan Sato, Chelsea Mediavilla, C. Ward, S. Parameswaran","doi":"10.1109/AIPR47015.2019.9174589","DOIUrl":null,"url":null,"abstract":"In this paper we introduce new video datasets to investigate the gaps between synthetic and real imagery in object detection and depth estimation. Currently, synthetic image datasets with real-world counterparts largely focus on computer vision applications for autonomous driving in unconstrained environments. The high scene complexity of such datasets pose challenges for systematic studies of domain disparities. We aim to create a set of paired datasets to study the discrepancies between the two domains in a more controlled setting. To this end, we have created Synthetic-Real Counterpart 3 (SRC3), which contains multiple datasets with varying levels of scene and object complexity. These versatile datasets span multiple environments and consist of ground-truthed, real-world, and synthetic videos generated using a gaming engine. In addition to the dataset, we present an in-depth analysis and provide comparison benchmarks of these datasets using state-of-the-art detection algorithms. Our results show contrasting performance during cross-domain testing due to differences in image quality and statistics, indicating a need for domain adapted datasets and models.","PeriodicalId":167075,"journal":{"name":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","volume":"71 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AIPR47015.2019.9174589","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
In this paper we introduce new video datasets to investigate the gaps between synthetic and real imagery in object detection and depth estimation. Currently, synthetic image datasets with real-world counterparts largely focus on computer vision applications for autonomous driving in unconstrained environments. The high scene complexity of such datasets pose challenges for systematic studies of domain disparities. We aim to create a set of paired datasets to study the discrepancies between the two domains in a more controlled setting. To this end, we have created Synthetic-Real Counterpart 3 (SRC3), which contains multiple datasets with varying levels of scene and object complexity. These versatile datasets span multiple environments and consist of ground-truthed, real-world, and synthetic videos generated using a gaming engine. In addition to the dataset, we present an in-depth analysis and provide comparison benchmarks of these datasets using state-of-the-art detection algorithms. Our results show contrasting performance during cross-domain testing due to differences in image quality and statistics, indicating a need for domain adapted datasets and models.