Guoyong Qian, Dongbo Xie, Dawei Bi, Qi Wang, Liqing Chen, Hai Wang
{"title":"Lightweight environment sensing algorithm for intelligent driving based on improved YOLOv7","authors":"Guoyong Qian, Dongbo Xie, Dawei Bi, Qi Wang, Liqing Chen, Hai Wang","doi":"10.1049/cth2.12704","DOIUrl":null,"url":null,"abstract":"Accurately and quickly detecting obstacles ahead is a prerequisite for intelligent driving. The combined detection scheme of light detection and ranging (LiDAR) and the camera is far more capable of coping with complex road conditions than a single sensor. However, immediately afterward, ensuring the real‐time performance of the sensing algorithms through a significantly increased amount of computation has become a new challenge. For this purpose, the paper introduces an improved dynamic obstacle detection algorithm based on YOLOv7 (You Only Look Once version 7) to overcome the drawbacks of slow and unstable detection of traditional methods. Concretely, Mobilenetv3 supplants the backbone network utilized in the original YOLOv7 architecture, thereby achieving a reduction in computational overhead. It integrates a specialized layer for the detection of small‐scale targets and incorporates a convolutional block attention module to enhance detection efficacy for diminutive obstacles. Furthermore, the framework adopts the Efficient Intersection over Union Loss function, which is specifically designed to mitigate the issue of mutual occlusion among detected objects. On a dataset consisting of 27,362 labelled KITTI data samples, the improved YOLOv7 algorithm achieves 92.6% mean average precision and 82 frames per second, which reduces the Model_size by 85.9% and loses only 1.5% accuracy compared with the traditional YOLOv7 algorithm. In addition, this paper builds a virtual scene to test the improved algorithm and fuses LiDAR and camera data. Experimental results conducted on a test vehicle equipped with a camera and LiDAR sensor demonstrate the effectiveness and significant performance of the method. The improved obstacle detection algorithm proposed in this research can significantly reduce the computational cost of the environment perception task, meet the requirements of real‐world applications, and is crucial for achieving safer and smarter driving.","PeriodicalId":502998,"journal":{"name":"IET Control Theory & Applications","volume":" 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Control Theory & Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1049/cth2.12704","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Accurately and quickly detecting obstacles ahead is a prerequisite for intelligent driving. The combined detection scheme of light detection and ranging (LiDAR) and the camera is far more capable of coping with complex road conditions than a single sensor. However, immediately afterward, ensuring the real‐time performance of the sensing algorithms through a significantly increased amount of computation has become a new challenge. For this purpose, the paper introduces an improved dynamic obstacle detection algorithm based on YOLOv7 (You Only Look Once version 7) to overcome the drawbacks of slow and unstable detection of traditional methods. Concretely, Mobilenetv3 supplants the backbone network utilized in the original YOLOv7 architecture, thereby achieving a reduction in computational overhead. It integrates a specialized layer for the detection of small‐scale targets and incorporates a convolutional block attention module to enhance detection efficacy for diminutive obstacles. Furthermore, the framework adopts the Efficient Intersection over Union Loss function, which is specifically designed to mitigate the issue of mutual occlusion among detected objects. On a dataset consisting of 27,362 labelled KITTI data samples, the improved YOLOv7 algorithm achieves 92.6% mean average precision and 82 frames per second, which reduces the Model_size by 85.9% and loses only 1.5% accuracy compared with the traditional YOLOv7 algorithm. In addition, this paper builds a virtual scene to test the improved algorithm and fuses LiDAR and camera data. Experimental results conducted on a test vehicle equipped with a camera and LiDAR sensor demonstrate the effectiveness and significant performance of the method. The improved obstacle detection algorithm proposed in this research can significantly reduce the computational cost of the environment perception task, meet the requirements of real‐world applications, and is crucial for achieving safer and smarter driving.
准确、快速地探测前方障碍物是智能驾驶的先决条件。光探测与测距(LiDAR)和摄像头的组合探测方案远比单一传感器更能应对复杂的路况。然而,紧接着,通过大幅增加计算量来确保传感算法的实时性能就成了新的挑战。为此,本文介绍了一种基于 YOLOv7(You Only Look Once version 7)的改进型动态障碍物检测算法,以克服传统方法检测速度慢和不稳定的缺点。具体来说,Mobilenetv3 取代了原有 YOLOv7 架构中使用的主干网络,从而减少了计算开销。它集成了一个专门用于检测小型目标的层,并加入了一个卷积块注意力模块,以提高对小型障碍物的检测效率。此外,该框架还采用了 "Efficient Intersection over Union Loss "函数,该函数专门用于缓解检测对象之间的相互遮挡问题。在由 27,362 个带标签的 KITTI 数据样本组成的数据集上,改进后的 YOLOv7 算法达到了 92.6% 的平均精度和 82 帧/秒的速度,与传统的 YOLOv7 算法相比,模型大小减少了 85.9%,精度仅降低了 1.5%。此外,本文还建立了一个虚拟场景来测试改进算法,并融合了激光雷达和摄像头数据。在装有摄像头和激光雷达传感器的测试车辆上进行的实验结果证明了该方法的有效性和显著性能。本研究提出的改进型障碍物检测算法能显著降低环境感知任务的计算成本,满足实际应用的要求,对实现更安全、更智能的驾驶至关重要。