X. Zhang, Yansheng Gong, Zhiwei Li, Xuan Liu, Shuyue Pan, Jun Li
{"title":"多模态注意力引导实时车道检测","authors":"X. Zhang, Yansheng Gong, Zhiwei Li, Xuan Liu, Shuyue Pan, Jun Li","doi":"10.1109/ICARM52023.2021.9536157","DOIUrl":null,"url":null,"abstract":"Multimodal data fusion is becoming a trend for the field of autonomous driving, especially for lane detection. In the process of driving, sensors often encounter problems such as modality imbalance, changing illumination and so on. Therefore, it is worthwhile to study the problems of applying multimodal fusion for lane detection and modality imbalance in the fusion process. In this paper, we propose a novel multimodal model for lane detection, in which attention mechanism is embedded into network to balance multimodal feature fusion and to improve detection capability. In addition, we use multi-frame input and long short-term memory (LSTM) network to solve the shadow interference, vehicles occlusion and mark degradation. At the same time, the network can be applied to the task of lane detection. In order to verify the effect of multimodal application and attention mechanism on fusion, we have designed adequate experiments on processed continuous scene KITTI dataset. The results show that precision increases by about 15% when LiDAR is added compared with RGB only. Besides, attention mechanism obviously improves the performance of multi-modal detection by balancing multi-modal features.","PeriodicalId":367307,"journal":{"name":"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)","volume":"56 2","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Multi-Modal Attention Guided Real-Time Lane Detection\",\"authors\":\"X. Zhang, Yansheng Gong, Zhiwei Li, Xuan Liu, Shuyue Pan, Jun Li\",\"doi\":\"10.1109/ICARM52023.2021.9536157\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multimodal data fusion is becoming a trend for the field of autonomous driving, especially for lane detection. In the process of driving, sensors often encounter problems such as modality imbalance, changing illumination and so on. Therefore, it is worthwhile to study the problems of applying multimodal fusion for lane detection and modality imbalance in the fusion process. In this paper, we propose a novel multimodal model for lane detection, in which attention mechanism is embedded into network to balance multimodal feature fusion and to improve detection capability. In addition, we use multi-frame input and long short-term memory (LSTM) network to solve the shadow interference, vehicles occlusion and mark degradation. At the same time, the network can be applied to the task of lane detection. In order to verify the effect of multimodal application and attention mechanism on fusion, we have designed adequate experiments on processed continuous scene KITTI dataset. The results show that precision increases by about 15% when LiDAR is added compared with RGB only. Besides, attention mechanism obviously improves the performance of multi-modal detection by balancing multi-modal features.\",\"PeriodicalId\":367307,\"journal\":{\"name\":\"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)\",\"volume\":\"56 2\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICARM52023.2021.9536157\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICARM52023.2021.9536157","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-Modal Attention Guided Real-Time Lane Detection
Multimodal data fusion is becoming a trend for the field of autonomous driving, especially for lane detection. In the process of driving, sensors often encounter problems such as modality imbalance, changing illumination and so on. Therefore, it is worthwhile to study the problems of applying multimodal fusion for lane detection and modality imbalance in the fusion process. In this paper, we propose a novel multimodal model for lane detection, in which attention mechanism is embedded into network to balance multimodal feature fusion and to improve detection capability. In addition, we use multi-frame input and long short-term memory (LSTM) network to solve the shadow interference, vehicles occlusion and mark degradation. At the same time, the network can be applied to the task of lane detection. In order to verify the effect of multimodal application and attention mechanism on fusion, we have designed adequate experiments on processed continuous scene KITTI dataset. The results show that precision increases by about 15% when LiDAR is added compared with RGB only. Besides, attention mechanism obviously improves the performance of multi-modal detection by balancing multi-modal features.