{"title":"InPosNet: Context Aware DNN for Visual SLAM","authors":"Anvaya Rai, B. Lall, Astha Zalani, Raghawendra Prakash Singh, Shikha Srivastava","doi":"10.1109/IRI58017.2023.00012","DOIUrl":null,"url":null,"abstract":"This paper introduces a novel approach to accurately localize a subject in indoor environments by using the scene images captured from the subject’s mobile phone camera. The objective of this work is to present a novel deep neural network (DNN), called InPosNet, that generates a concise representation of an indoor scene while being able to distinguish between their inherent symmetry. It also enables the user in real time distinction between the images of the same location but captured from different orientations, thereby enabling the user to detect the orientation along with position. A localization accuracy of less than 1 meter from ground truth is achieved and enumerated through the experimental results. The novel DNN presented in the work is motivated by MobileNetv3-Small [2], followed by PCA based feature space transformation. PCA helps in feature space dimensionality reduction and projection of query images onto an optimally dense subspace of the original latent feature space. The goal is to present a vision based system that will have the ability to be used for indoor positioning, without any need for additional infrastructure or external hardware.","PeriodicalId":290818,"journal":{"name":"2023 IEEE 24th International Conference on Information Reuse and Integration for Data Science (IRI)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 24th International Conference on Information Reuse and Integration for Data Science (IRI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IRI58017.2023.00012","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper introduces a novel approach to accurately localize a subject in indoor environments by using the scene images captured from the subject’s mobile phone camera. The objective of this work is to present a novel deep neural network (DNN), called InPosNet, that generates a concise representation of an indoor scene while being able to distinguish between their inherent symmetry. It also enables the user in real time distinction between the images of the same location but captured from different orientations, thereby enabling the user to detect the orientation along with position. A localization accuracy of less than 1 meter from ground truth is achieved and enumerated through the experimental results. The novel DNN presented in the work is motivated by MobileNetv3-Small [2], followed by PCA based feature space transformation. PCA helps in feature space dimensionality reduction and projection of query images onto an optimally dense subspace of the original latent feature space. The goal is to present a vision based system that will have the ability to be used for indoor positioning, without any need for additional infrastructure or external hardware.