{"title":"CPR-SLAM: RGB-D SLAM in dynamic environment using sub-point cloud correlations","authors":"Xinyi Yu, Wancai Zheng, Linlin Ou","doi":"10.1017/s0263574724000754","DOIUrl":null,"url":null,"abstract":"\n The early applications of Visual Simultaneous Localization and Mapping (VSLAM) technology were primarily focused on static environments, relying on the static nature of the environment for map construction and localization. However, in practical applications, we often encounter various dynamic environments, such as city streets, where moving objects are present. These dynamic objects can make it challenging for robots to accurately understand their own position. This paper proposes a real-time localization and mapping method tailored for dynamic environments to effectively deal with the interference of moving objects in such settings. Firstly, depth images are clustered, and they are subdivided into sub-point clouds to obtain clearer local information. Secondly, when processing regular frames, we fully exploit the structural invariance of static sub-point clouds and their relative relationships. Among these, the concept of the sub-point cloud is introduced as novel idea in this paper. By utilizing the results computed based on sub-poses, we can effectively quantify the disparities between regular frames and reference frames. This enables us to accurately detect dynamic areas within the regular frames. Furthermore, by refining the dynamic areas of keyframes using historical observation data, the robustness of the system is further enhanced. We conducted comprehensive experimental evaluations on challenging dynamic sequences from the TUM dataset and compared our approach with state-of-the-art dynamic VSLAM systems. The experimental results demonstrate that our method significantly enhances the accuracy and robustness of pose estimation. Additionally, we validated the effectiveness of the system in dynamic environments through real-world scenario tests.","PeriodicalId":49593,"journal":{"name":"Robotica","volume":null,"pages":null},"PeriodicalIF":1.9000,"publicationDate":"2024-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Robotica","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1017/s0263574724000754","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
The early applications of Visual Simultaneous Localization and Mapping (VSLAM) technology were primarily focused on static environments, relying on the static nature of the environment for map construction and localization. However, in practical applications, we often encounter various dynamic environments, such as city streets, where moving objects are present. These dynamic objects can make it challenging for robots to accurately understand their own position. This paper proposes a real-time localization and mapping method tailored for dynamic environments to effectively deal with the interference of moving objects in such settings. Firstly, depth images are clustered, and they are subdivided into sub-point clouds to obtain clearer local information. Secondly, when processing regular frames, we fully exploit the structural invariance of static sub-point clouds and their relative relationships. Among these, the concept of the sub-point cloud is introduced as novel idea in this paper. By utilizing the results computed based on sub-poses, we can effectively quantify the disparities between regular frames and reference frames. This enables us to accurately detect dynamic areas within the regular frames. Furthermore, by refining the dynamic areas of keyframes using historical observation data, the robustness of the system is further enhanced. We conducted comprehensive experimental evaluations on challenging dynamic sequences from the TUM dataset and compared our approach with state-of-the-art dynamic VSLAM systems. The experimental results demonstrate that our method significantly enhances the accuracy and robustness of pose estimation. Additionally, we validated the effectiveness of the system in dynamic environments through real-world scenario tests.
视觉同步定位与绘图(VSLAM)技术的早期应用主要集中在静态环境中,依靠环境的静态特性来构建地图和进行定位。然而,在实际应用中,我们经常会遇到各种动态环境,如城市街道,其中存在移动物体。这些动态物体会给机器人准确了解自身位置带来挑战。本文提出了一种专为动态环境量身定制的实时定位和绘图方法,以有效应对此类环境中移动物体的干扰。首先,对深度图像进行聚类,并将其细分为子点云,以获得更清晰的局部信息。其次,在处理常规帧时,我们充分利用了静态子点云的结构不变性及其相对关系。其中,子点云的概念是本文引入的新思路。利用基于子姿态计算出的结果,我们可以有效地量化常规帧和参考帧之间的差异。这样,我们就能准确检测出常规帧内的动态区域。此外,通过使用历史观测数据完善关键帧的动态区域,还能进一步增强系统的鲁棒性。我们对 TUM 数据集中具有挑战性的动态序列进行了全面的实验评估,并将我们的方法与最先进的动态 VSLAM 系统进行了比较。实验结果表明,我们的方法显著提高了姿势估计的准确性和鲁棒性。此外,我们还通过实际场景测试验证了系统在动态环境中的有效性。
期刊介绍:
Robotica is a forum for the multidisciplinary subject of robotics and encourages developments, applications and research in this important field of automation and robotics with regard to industry, health, education and economic and social aspects of relevance. Coverage includes activities in hostile environments, applications in the service and manufacturing industries, biological robotics, dynamics and kinematics involved in robot design and uses, on-line robots, robot task planning, rehabilitation robotics, sensory perception, software in the widest sense, particularly in respect of programming languages and links with CAD/CAM systems, telerobotics and various other areas. In addition, interest is focused on various Artificial Intelligence topics of theoretical and practical interest.