Dynamic scene SLAM algorithm based on semantic information and joint constraints of optical flow and geometry

Jinyan Li, Xiangde Liu, Yi Zhang, Yunchuan Hu
{"title":"Dynamic scene SLAM algorithm based on semantic information and joint constraints of optical flow and geometry","authors":"Jinyan Li, Xiangde Liu, Yi Zhang, Yunchuan Hu","doi":"10.1109/WCMEIM56910.2022.10021365","DOIUrl":null,"url":null,"abstract":"Traditional simultaneous localization and mapping (SALM) algorithms are based on static environments. If there are dynamic objects in the environment, it will cause inaccurate positioning or problems that cannot be located. In order to solve this problem, the method of SegNet lightweight neural network and sparse optical flow combined with multi-view geometry is proposed to eliminate dynamic feature points. Firstly, the SegNet network is used to obtain the mask of potential moving objects. Secondly, sparse optical flow and geometric methods detect dynamic feature points. Finally, the dynamic feature points detected by semantics, optical flow, and geometric methods are combined to reject the feature points. This method can improve the positioning accuracy of the SLAM system in a dynamic environment.","PeriodicalId":202270,"journal":{"name":"2022 5th World Conference on Mechanical Engineering and Intelligent Manufacturing (WCMEIM)","volume":"434 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 5th World Conference on Mechanical Engineering and Intelligent Manufacturing (WCMEIM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WCMEIM56910.2022.10021365","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Traditional simultaneous localization and mapping (SALM) algorithms are based on static environments. If there are dynamic objects in the environment, it will cause inaccurate positioning or problems that cannot be located. In order to solve this problem, the method of SegNet lightweight neural network and sparse optical flow combined with multi-view geometry is proposed to eliminate dynamic feature points. Firstly, the SegNet network is used to obtain the mask of potential moving objects. Secondly, sparse optical flow and geometric methods detect dynamic feature points. Finally, the dynamic feature points detected by semantics, optical flow, and geometric methods are combined to reject the feature points. This method can improve the positioning accuracy of the SLAM system in a dynamic environment.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于语义信息和光流和几何联合约束的动态场景SLAM算法
传统的同步定位与映射算法是基于静态环境的。如果环境中存在动态物体,则会造成定位不准确或无法定位的问题。为了解决这一问题,提出了SegNet轻量级神经网络和稀疏光流结合多视图几何的方法来消除动态特征点。首先,利用隔离网网络获取潜在运动目标的掩码;其次,利用稀疏光流和几何方法检测动态特征点;最后,结合语义、光流和几何方法检测的动态特征点进行特征点的剔除。该方法可以提高SLAM系统在动态环境下的定位精度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Design and Analysis of a Novel Soft Actuator with High Contraction Ratio Based on Nested Structure Design and Verification of Thermal Balance System for Electric Drive Transmission in Urban Public Transit Design and Experiment of a Novel Manipulator for Autonomous Harvesting Tomato Clusters Research on Young's Modulus Prediction Model of Particle Reinforced Composites The Liquid Rocket Engine Experiment Data Quality Improvement Based on 3σ-LMBP
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1