A New Approach to Lidar and Camera Fusion for Autonomous Driving

Seunghwan Bae, Dongun Han, Seongkeun Park
{"title":"A New Approach to Lidar and Camera Fusion for Autonomous Driving","authors":"Seunghwan Bae, Dongun Han, Seongkeun Park","doi":"10.1109/ICAIIC57133.2023.10066963","DOIUrl":null,"url":null,"abstract":"In this paper, we introduce an object detection model that combines a camera and a LiDAR sensor. In previous object detection studies have mainly focused on using one sensor, and mainly camera and LiDAR sensors were used. Research was mainly conducted in the direction of utilizing a single sensor, and typically cameras and LiDAR sensors were used. However, Camera and Li-DAR sensors have disadvantages such as being vulnerable to environmental changes or having sparse expressive power, so the method to improve them is needed for a stable cognitive system. In this paper, we propose the LiDAR Camera Fusion Network, a sensor fusion object detection model that uses the advantages of each sensor to improve the disadvantages of cameras and Li-DAR sensors. The sensor fusion object detector developed in this study has the feature of estimating the location of an object through LiDAR Clustering. Extraction speed is about 58 times faster than Selective search without prior learning, reducing the number of candidate regions from 2000 to 98, despite reducing the number of candidate regions, compared to existing methods, the ratio of the correct answer candidate areas among the total location candidate regions was 10 times larger. Due to the above characteristics, efficient learning and inference were possible compared to the existing method, and this model finally extracts the probability value of the object, the bounding box correction value, and the distance value from the object. Due to the characteristic of our research, we used KITTI data because LiDAR and image data were needed. As a result, we compare the results with object detection models that are often used in the object detection area.","PeriodicalId":105769,"journal":{"name":"2023 International Conference on Artificial Intelligence in Information and Communication (ICAIIC)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-02-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Artificial Intelligence in Information and Communication (ICAIIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAIIC57133.2023.10066963","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this paper, we introduce an object detection model that combines a camera and a LiDAR sensor. In previous object detection studies have mainly focused on using one sensor, and mainly camera and LiDAR sensors were used. Research was mainly conducted in the direction of utilizing a single sensor, and typically cameras and LiDAR sensors were used. However, Camera and Li-DAR sensors have disadvantages such as being vulnerable to environmental changes or having sparse expressive power, so the method to improve them is needed for a stable cognitive system. In this paper, we propose the LiDAR Camera Fusion Network, a sensor fusion object detection model that uses the advantages of each sensor to improve the disadvantages of cameras and Li-DAR sensors. The sensor fusion object detector developed in this study has the feature of estimating the location of an object through LiDAR Clustering. Extraction speed is about 58 times faster than Selective search without prior learning, reducing the number of candidate regions from 2000 to 98, despite reducing the number of candidate regions, compared to existing methods, the ratio of the correct answer candidate areas among the total location candidate regions was 10 times larger. Due to the above characteristics, efficient learning and inference were possible compared to the existing method, and this model finally extracts the probability value of the object, the bounding box correction value, and the distance value from the object. Due to the characteristic of our research, we used KITTI data because LiDAR and image data were needed. As a result, we compare the results with object detection models that are often used in the object detection area.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
自动驾驶激光雷达与摄像头融合的新方法
本文介绍了一种结合摄像头和激光雷达传感器的目标检测模型。以往的目标检测研究主要集中在单一传感器的使用上,主要是相机和激光雷达传感器。研究主要是在单一传感器的使用方向上进行的,通常使用相机和LiDAR传感器。然而,Camera和Li-DAR传感器存在易受环境变化影响或表达能力较弱的缺点,因此需要一种稳定的认知系统来改进它们。在本文中,我们提出了LiDAR相机融合网络,这是一种传感器融合目标检测模型,利用每个传感器的优点来改进相机和Li-DAR传感器的缺点。本研究开发的传感器融合目标检测器具有通过激光雷达聚类估计目标位置的特点。提取速度比无先验学习的选择性搜索快约58倍,候选区域数量从2000个减少到98个,尽管候选区域数量减少,但与现有方法相比,正确答案候选区域占总位置候选区域的比例增加了10倍。由于上述特点,与现有方法相比,该模型可以进行高效的学习和推理,并最终提取出目标的概率值、边界框校正值以及与目标的距离值。由于我们研究的特点,我们使用KITTI数据,因为需要激光雷达和图像数据。因此,我们将结果与目标检测领域常用的目标检测模型进行了比较。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Development of AI Educational Datasets Library Using Synthetic Dataset Generation Method Channel Access Control Instead of Random Backoff Algorithm Illegal 3D Content Distribution Tracking System based on DNN Forensic Watermarking Deep Learning-based Spectral Efficiency Maximization in Massive MIMO-NOMA Systems with STAR-RIS Data Pipeline Design for Dangerous Driving Behavior Detection System
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1