{"title":"Performance Enhancement using Data Augmentation of Depth Estimation for Autonomous Driving","authors":"Jisang Yoo, Woomin Jun, Sungjin Lee","doi":"10.1109/ICCE59016.2024.10444235","DOIUrl":null,"url":null,"abstract":"For autonomous driving, various sensors such as cameras, LiDAR, and radar are required to accurately perceive the surrounding environment. These sensors provide information for tasks like object recognition, lane detection, path planning, and distance estimation. However, processing information from these multiple sensors for perception tasks demands significant costs, computational resources, and latency. These challenges hinder the practical implementation of real-time edge computing in autonomous driving systems. Consequently, research is actively exploring methods to perform perception using only cameras, particularly to alleviate the computational burden and cost associated with 3D point cloud data from LiDAR or radar sensors. In this study, we investigate techniques to optimize the performance of Monocular Depth Estimation (MDE) methods, which utilize a single camera to extract 3D information about the surrounding environment. We focus on enhancing accuracy through classical data augmentation techniques and synthetic data generation methods. Additionally, we explore the selection of an optimal loss function. Experimental results demonstrate that employing our proposed data augmentation approach reduces REL by approximately 3.9%, showcasing the potential of this method.","PeriodicalId":518694,"journal":{"name":"2024 IEEE International Conference on Consumer Electronics (ICCE)","volume":"2 5","pages":"1-7"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 IEEE International Conference on Consumer Electronics (ICCE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCE59016.2024.10444235","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
For autonomous driving, various sensors such as cameras, LiDAR, and radar are required to accurately perceive the surrounding environment. These sensors provide information for tasks like object recognition, lane detection, path planning, and distance estimation. However, processing information from these multiple sensors for perception tasks demands significant costs, computational resources, and latency. These challenges hinder the practical implementation of real-time edge computing in autonomous driving systems. Consequently, research is actively exploring methods to perform perception using only cameras, particularly to alleviate the computational burden and cost associated with 3D point cloud data from LiDAR or radar sensors. In this study, we investigate techniques to optimize the performance of Monocular Depth Estimation (MDE) methods, which utilize a single camera to extract 3D information about the surrounding environment. We focus on enhancing accuracy through classical data augmentation techniques and synthetic data generation methods. Additionally, we explore the selection of an optimal loss function. Experimental results demonstrate that employing our proposed data augmentation approach reduces REL by approximately 3.9%, showcasing the potential of this method.