{"title":"Unsupervised Scale Network for Monocular Relative Depth and Visual Odometry","authors":"Zhongyi Wang;Qijun Chen","doi":"10.1109/TIM.2024.3451584","DOIUrl":null,"url":null,"abstract":"With the rapid development of deep learning and computer vision, learning-based monocular depth estimation and visual odometry have achieved increasingly remarkable results. However, there are few studies on the scale ambiguity of monocular depth estimation and visual odometry in an unsupervised network framework. Therefore, this article is to solve this thorny problem. We propose a joint unsupervised network framework that can provide necessary information to each other between different tasks to meet the needs of multiple tasks. To address the issue of scale ambiguity for learning-based monocular depth estimation, we propose a novel ScaleNet, an unsupervised scale network that provides scale information for the relative depths predicted by monocular depth networks, thereby recovering the absolute depths. Meanwhile, we propose a pseudo ground-truth scale generator and constrain the scale network by scale loss. The experimental results show that our monocular depth estimation results are competitive, and the scale network can provide reliable scale information for monocular depth networks. To address the challenge of scale ambiguity in learning-based monocular visual odometry, we propose a solution based on scale and optical flow to obtain the absolute scale of translational vectors using our depth alignment method. The experimental results show that our monocular visual odometry achieves state-of-the-art performance. The extensive experiments on the KITTI dataset for different tasks demonstrate the effectiveness and generalization of our proposed method.","PeriodicalId":13341,"journal":{"name":"IEEE Transactions on Instrumentation and Measurement","volume":null,"pages":null},"PeriodicalIF":5.6000,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Instrumentation and Measurement","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10666785/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
With the rapid development of deep learning and computer vision, learning-based monocular depth estimation and visual odometry have achieved increasingly remarkable results. However, there are few studies on the scale ambiguity of monocular depth estimation and visual odometry in an unsupervised network framework. Therefore, this article is to solve this thorny problem. We propose a joint unsupervised network framework that can provide necessary information to each other between different tasks to meet the needs of multiple tasks. To address the issue of scale ambiguity for learning-based monocular depth estimation, we propose a novel ScaleNet, an unsupervised scale network that provides scale information for the relative depths predicted by monocular depth networks, thereby recovering the absolute depths. Meanwhile, we propose a pseudo ground-truth scale generator and constrain the scale network by scale loss. The experimental results show that our monocular depth estimation results are competitive, and the scale network can provide reliable scale information for monocular depth networks. To address the challenge of scale ambiguity in learning-based monocular visual odometry, we propose a solution based on scale and optical flow to obtain the absolute scale of translational vectors using our depth alignment method. The experimental results show that our monocular visual odometry achieves state-of-the-art performance. The extensive experiments on the KITTI dataset for different tasks demonstrate the effectiveness and generalization of our proposed method.
期刊介绍:
Papers are sought that address innovative solutions to the development and use of electrical and electronic instruments and equipment to measure, monitor and/or record physical phenomena for the purpose of advancing measurement science, methods, functionality and applications. The scope of these papers may encompass: (1) theory, methodology, and practice of measurement; (2) design, development and evaluation of instrumentation and measurement systems and components used in generating, acquiring, conditioning and processing signals; (3) analysis, representation, display, and preservation of the information obtained from a set of measurements; and (4) scientific and technical support to establishment and maintenance of technical standards in the field of Instrumentation and Measurement.