用于可解释自动驾驶系统的图像转换器的开发和测试

Jiqian Dong;Sikai Chen;Mohammad Miralinaghi;Tiantian Chen;Samuel Labi
{"title":"用于可解释自动驾驶系统的图像转换器的开发和测试","authors":"Jiqian Dong;Sikai Chen;Mohammad Miralinaghi;Tiantian Chen;Samuel Labi","doi":"10.1108/JICV-06-2022-0021","DOIUrl":null,"url":null,"abstract":"Purpose - Perception has been identified as the main cause underlying most autonomous vehicle related accidents. As the key technology in perception, deep learning (DL) based computer vision models are generally considered to be black boxes due to poor interpretability. These have exacerbated user distrust and further forestalled their widespread deployment in practical usage. This paper aims to develop explainable DL models for autonomous driving by jointly predicting potential driving actions with corresponding explanations. The explainable DL models can not only boost user trust in autonomy but also serve as a diagnostic approach to identify any model deficiencies or limitations during the system development phase. Design/methodology/approach - This paper proposes an explainable end-to-end autonomous driving system based on \"Transformer,\" a state-of-the-art self-attention (SA) based model. The model maps visual features from images collected by onboard cameras to guide potential driving actions with corresponding explanations, and aims to achieve soft attention over the image's global features. Findings - The results demonstrate the efficacy of the proposed model as it exhibits superior performance (in terms of correct prediction of actions and explanations) compared to the benchmark model by a significant margin with much lower computational cost on a public data set (BDD-OIA). From the ablation studies, the proposed SA module also outperforms other attention mechanisms in feature fusion and can generate meaningful representations for downstream prediction. Originality/value - In the contexts of situational awareness and driver assistance, the proposed model can perform as a driving alarm system for both human-driven vehicles and autonomous vehicles because it is capable of quickly understanding/characterizing the environment and identifying any infeasible driving actions. In addition, the extra explanation head of the proposed model provides an extra channel for sanity checks to guarantee that the model learns the ideal causal relationships. This provision is critical in the development of autonomous systems.","PeriodicalId":100793,"journal":{"name":"Journal of Intelligent and Connected Vehicles","volume":"5 3","pages":"235-249"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/iel7/9944931/10004521/10004532.pdf","citationCount":"10","resultStr":"{\"title\":\"Development and testing of an image transformer for explainable autonomous driving systems\",\"authors\":\"Jiqian Dong;Sikai Chen;Mohammad Miralinaghi;Tiantian Chen;Samuel Labi\",\"doi\":\"10.1108/JICV-06-2022-0021\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Purpose - Perception has been identified as the main cause underlying most autonomous vehicle related accidents. As the key technology in perception, deep learning (DL) based computer vision models are generally considered to be black boxes due to poor interpretability. These have exacerbated user distrust and further forestalled their widespread deployment in practical usage. This paper aims to develop explainable DL models for autonomous driving by jointly predicting potential driving actions with corresponding explanations. The explainable DL models can not only boost user trust in autonomy but also serve as a diagnostic approach to identify any model deficiencies or limitations during the system development phase. Design/methodology/approach - This paper proposes an explainable end-to-end autonomous driving system based on \\\"Transformer,\\\" a state-of-the-art self-attention (SA) based model. The model maps visual features from images collected by onboard cameras to guide potential driving actions with corresponding explanations, and aims to achieve soft attention over the image's global features. Findings - The results demonstrate the efficacy of the proposed model as it exhibits superior performance (in terms of correct prediction of actions and explanations) compared to the benchmark model by a significant margin with much lower computational cost on a public data set (BDD-OIA). From the ablation studies, the proposed SA module also outperforms other attention mechanisms in feature fusion and can generate meaningful representations for downstream prediction. Originality/value - In the contexts of situational awareness and driver assistance, the proposed model can perform as a driving alarm system for both human-driven vehicles and autonomous vehicles because it is capable of quickly understanding/characterizing the environment and identifying any infeasible driving actions. In addition, the extra explanation head of the proposed model provides an extra channel for sanity checks to guarantee that the model learns the ideal causal relationships. This provision is critical in the development of autonomous systems.\",\"PeriodicalId\":100793,\"journal\":{\"name\":\"Journal of Intelligent and Connected Vehicles\",\"volume\":\"5 3\",\"pages\":\"235-249\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://ieeexplore.ieee.org/iel7/9944931/10004521/10004532.pdf\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Intelligent and Connected Vehicles\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10004532/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Intelligent and Connected Vehicles","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10004532/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10

摘要

目的-感知已被确定为大多数自动驾驶汽车相关事故的主要原因。作为感知领域的关键技术,基于深度学习的计算机视觉模型由于其可解释性差,通常被认为是黑匣子。这些加剧了用户的不信任,并进一步阻止了它们在实际使用中的广泛部署。本文旨在通过联合预测潜在的驾驶行为和相应的解释,开发可解释的自动驾驶DL模型。可解释的DL模型不仅可以提高用户对自主性的信任,还可以作为一种诊断方法,在系统开发阶段识别任何模型缺陷或限制。设计/方法/方法-本文提出了一种基于“Transformer”的可解释的端到端自动驾驶系统,这是一种最先进的基于自注意(SA)的模型。该模型从车载摄像头收集的图像中映射视觉特征,以通过相应的解释来指导潜在的驾驶行为,并旨在实现对图像全局特征的软关注。研究结果-结果证明了所提出的模型的有效性,因为与基准模型相比,该模型在公共数据集(BDD-OIA)上表现出显著的优越性能(在行动和解释的正确预测方面),计算成本低得多。从消融研究来看,所提出的SA模块在特征融合方面也优于其他注意力机制,并且可以为下游预测生成有意义的表示。独创性/价值-在情景感知和驾驶员辅助的背景下,所提出的模型可以作为人类驾驶车辆和自动驾驶车辆的驾驶警报系统,因为它能够快速了解/表征环境并识别任何不可行的驾驶行为。此外,所提出的模型的额外解释头为健全性检查提供了额外的通道,以确保模型学习理想的因果关系。这一规定对自主系统的发展至关重要。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Development and testing of an image transformer for explainable autonomous driving systems
Purpose - Perception has been identified as the main cause underlying most autonomous vehicle related accidents. As the key technology in perception, deep learning (DL) based computer vision models are generally considered to be black boxes due to poor interpretability. These have exacerbated user distrust and further forestalled their widespread deployment in practical usage. This paper aims to develop explainable DL models for autonomous driving by jointly predicting potential driving actions with corresponding explanations. The explainable DL models can not only boost user trust in autonomy but also serve as a diagnostic approach to identify any model deficiencies or limitations during the system development phase. Design/methodology/approach - This paper proposes an explainable end-to-end autonomous driving system based on "Transformer," a state-of-the-art self-attention (SA) based model. The model maps visual features from images collected by onboard cameras to guide potential driving actions with corresponding explanations, and aims to achieve soft attention over the image's global features. Findings - The results demonstrate the efficacy of the proposed model as it exhibits superior performance (in terms of correct prediction of actions and explanations) compared to the benchmark model by a significant margin with much lower computational cost on a public data set (BDD-OIA). From the ablation studies, the proposed SA module also outperforms other attention mechanisms in feature fusion and can generate meaningful representations for downstream prediction. Originality/value - In the contexts of situational awareness and driver assistance, the proposed model can perform as a driving alarm system for both human-driven vehicles and autonomous vehicles because it is capable of quickly understanding/characterizing the environment and identifying any infeasible driving actions. In addition, the extra explanation head of the proposed model provides an extra channel for sanity checks to guarantee that the model learns the ideal causal relationships. This provision is critical in the development of autonomous systems.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
7.10
自引率
0.00%
发文量
0
期刊最新文献
Front Cover Contents Advancements and Prospects in Multisensor Fusion for Autonomous Driving Extracting Networkwide Road Segment Location, Direction, and Turning Movement Rules From Global Positioning System Vehicle Trajectory Data for Macrosimulation Decision Making and Control of Autonomous Vehicles Under the Condition of Front Vehicle Sideslip
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1