通过融合二维图像和激光雷达并行处理三维物体识别,实现自动驾驶

Heuijee Yun, Daejin Park
{"title":"通过融合二维图像和激光雷达并行处理三维物体识别,实现自动驾驶","authors":"Heuijee Yun, Daejin Park","doi":"10.1109/ICEIC61013.2024.10457186","DOIUrl":null,"url":null,"abstract":"At the moment, autonomous driving requires a lot of sensors: cameras, lidar, etc. It takes a lot of time and resources to process all the input data from these sensors. In this paper, we reduce the processing time and resources of lidar and camera data by parallelizing the input data of autonomous vehicles. Cameras mounted on autonomous vehicles are often wide-angle or have multiple angles of view. These multiple camera inputs are flattened and processed in parallel, and then YOLO is used to combine the 3D data from the lidar with the 2D inputs from the camera. By combining cameras from multiple angles and processing them in parallel, except where they overlap, you can reduce the time it would take to process each image serially. This algorithm is also highly scalable as it can be applied to a single camera rather than multiple camera sensors. Experiments were conducted using KITTY and YOLO with labelled 3D lidar data and 2D image data. The FPS is 7.98, which is fast, and the parallel processing reduces the time by about 1.4 times.","PeriodicalId":518726,"journal":{"name":"2024 International Conference on Electronics, Information, and Communication (ICEIC)","volume":"369 1","pages":"1-2"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Parallel Processing of 3D Object Recognition by Fusion of 2D Images and LiDAR for Autonomous Driving\",\"authors\":\"Heuijee Yun, Daejin Park\",\"doi\":\"10.1109/ICEIC61013.2024.10457186\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"At the moment, autonomous driving requires a lot of sensors: cameras, lidar, etc. It takes a lot of time and resources to process all the input data from these sensors. In this paper, we reduce the processing time and resources of lidar and camera data by parallelizing the input data of autonomous vehicles. Cameras mounted on autonomous vehicles are often wide-angle or have multiple angles of view. These multiple camera inputs are flattened and processed in parallel, and then YOLO is used to combine the 3D data from the lidar with the 2D inputs from the camera. By combining cameras from multiple angles and processing them in parallel, except where they overlap, you can reduce the time it would take to process each image serially. This algorithm is also highly scalable as it can be applied to a single camera rather than multiple camera sensors. Experiments were conducted using KITTY and YOLO with labelled 3D lidar data and 2D image data. The FPS is 7.98, which is fast, and the parallel processing reduces the time by about 1.4 times.\",\"PeriodicalId\":518726,\"journal\":{\"name\":\"2024 International Conference on Electronics, Information, and Communication (ICEIC)\",\"volume\":\"369 1\",\"pages\":\"1-2\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2024 International Conference on Electronics, Information, and Communication (ICEIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICEIC61013.2024.10457186\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 International Conference on Electronics, Information, and Communication (ICEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICEIC61013.2024.10457186","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

目前,自动驾驶需要大量传感器:摄像头、激光雷达等。处理这些传感器的所有输入数据需要耗费大量时间和资源。在本文中,我们通过并行处理自动驾驶车辆的输入数据,减少了激光雷达和摄像头数据的处理时间和资源。自动驾驶车辆上安装的摄像头通常是广角或多视角的。这些多角度的摄像头输入数据会被扁平化并行处理,然后使用 YOLO 将激光雷达的三维数据与摄像头的二维输入数据结合起来。通过将多角度的相机组合起来并行处理(重叠部分除外),可以减少串行处理每幅图像所需的时间。这种算法还具有很强的可扩展性,因为它可以应用于单个摄像头,而不是多个摄像头传感器。我们使用 KITTY 和 YOLO 对标有三维激光雷达数据和二维图像数据进行了实验。FPS 为 7.98,速度很快,并行处理时间缩短了约 1.4 倍。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Parallel Processing of 3D Object Recognition by Fusion of 2D Images and LiDAR for Autonomous Driving
At the moment, autonomous driving requires a lot of sensors: cameras, lidar, etc. It takes a lot of time and resources to process all the input data from these sensors. In this paper, we reduce the processing time and resources of lidar and camera data by parallelizing the input data of autonomous vehicles. Cameras mounted on autonomous vehicles are often wide-angle or have multiple angles of view. These multiple camera inputs are flattened and processed in parallel, and then YOLO is used to combine the 3D data from the lidar with the 2D inputs from the camera. By combining cameras from multiple angles and processing them in parallel, except where they overlap, you can reduce the time it would take to process each image serially. This algorithm is also highly scalable as it can be applied to a single camera rather than multiple camera sensors. Experiments were conducted using KITTY and YOLO with labelled 3D lidar data and 2D image data. The FPS is 7.98, which is fast, and the parallel processing reduces the time by about 1.4 times.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Study on Improving the Durability of Shaded Pole Induction Motors Used for Refrigerator Fans New Approximate 4:2 Compressor for High Accuracy and Small Area Using MUX Logic A Study on the UWB/Encoder/IMU Sensor Fusion Position Estimation System for the Development of Driving Assistance Technology in Autonomous Driving Wheelchairs DDANet: Dilated Deformable Attention Network for Dynamic Scene Deblurring NIR to LWIR Image Translation for Generating LWIR Image Datasets
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1