Monitoring perception reliability in autonomous driving: Distributional shift detection for estimating the impact of input data on prediction accuracy

F. Hell, Gereon Hinz, Feng Liu, Sakshi Goyal, Ke Pei, T. Lytvynenko, Alois Knoll, Yiqiang Chen
{"title":"Monitoring perception reliability in autonomous driving: Distributional shift detection for estimating the impact of input data on prediction accuracy","authors":"F. Hell, Gereon Hinz, Feng Liu, Sakshi Goyal, Ke Pei, T. Lytvynenko, Alois Knoll, Yiqiang Chen","doi":"10.1145/3488904.3493382","DOIUrl":null,"url":null,"abstract":"Deep neural networks are at the heart of safety-critical applications such as autonomous driving. Distributional shift is a typical problem in predictive modeling, when the feature distribution of inputs and outputs varies between the training and test stages. When used on data different from the training distribution, neural networks provide little or no performance guarantees on such out-of-distribution (OOD) inputs. Monitoring distributional shift can help assess reliability of neural network predictions with the purpose of predicting potential safety-critical contexts. With our research, we evaluate state of the art OOD detection methods on autonomous driving camera data, while also demonstrating the influence of OOD data on the prediction reliability of neural networks. We evaluate three different OOD detection methods: As a baseline method we employ a variational autoencoder (VAE) trained on the similar data as the perception network (depth estimation) and use a reconstruction error based out of distribution measure. As a second approach, we choose to evaluate a method termed Likelihood Regret, which has been shown to be an efficient likelihood based OOD measure for VAEs. As a third approach, we evaluate another recently introduced method based on generative modelling termed SSD, which uses self-supervised representation learning followed by a distance based detection in the feature space, to calculate the outlier score. We compare all 3 methods and evaluate them concurrently with the error of an depth estimation network. Results show that while the reconstruction error based OOD metric is not able to differentiate between in and out of distribution data across all scenarios, the likelihood regret based OOD metric as well as the SSD outlier score perform fairly well in OOD detection. Their metrics are also highly correlated with perception error, rendering them promising candidates for an autonomous driving system reliability monitor.","PeriodicalId":332312,"journal":{"name":"Proceedings of the 5th ACM Computer Science in Cars Symposium","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 5th ACM Computer Science in Cars Symposium","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3488904.3493382","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

Deep neural networks are at the heart of safety-critical applications such as autonomous driving. Distributional shift is a typical problem in predictive modeling, when the feature distribution of inputs and outputs varies between the training and test stages. When used on data different from the training distribution, neural networks provide little or no performance guarantees on such out-of-distribution (OOD) inputs. Monitoring distributional shift can help assess reliability of neural network predictions with the purpose of predicting potential safety-critical contexts. With our research, we evaluate state of the art OOD detection methods on autonomous driving camera data, while also demonstrating the influence of OOD data on the prediction reliability of neural networks. We evaluate three different OOD detection methods: As a baseline method we employ a variational autoencoder (VAE) trained on the similar data as the perception network (depth estimation) and use a reconstruction error based out of distribution measure. As a second approach, we choose to evaluate a method termed Likelihood Regret, which has been shown to be an efficient likelihood based OOD measure for VAEs. As a third approach, we evaluate another recently introduced method based on generative modelling termed SSD, which uses self-supervised representation learning followed by a distance based detection in the feature space, to calculate the outlier score. We compare all 3 methods and evaluate them concurrently with the error of an depth estimation network. Results show that while the reconstruction error based OOD metric is not able to differentiate between in and out of distribution data across all scenarios, the likelihood regret based OOD metric as well as the SSD outlier score perform fairly well in OOD detection. Their metrics are also highly correlated with perception error, rendering them promising candidates for an autonomous driving system reliability monitor.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
深度神经网络是自动驾驶等安全关键应用的核心。分布移位是预测建模中的一个典型问题,当输入和输出的特征分布在训练和测试阶段之间发生变化时。当用于与训练分布不同的数据时,神经网络对这种偏离分布(OOD)的输入提供很少或根本没有性能保证。监测分布位移可以帮助评估神经网络预测的可靠性,从而预测潜在的安全关键环境。通过我们的研究,我们评估了自动驾驶摄像头数据上最先进的OOD检测方法,同时也证明了OOD数据对神经网络预测可靠性的影响。我们评估了三种不同的OOD检测方法:作为基线方法,我们使用与感知网络(深度估计)相似的数据训练的变分自编码器(VAE),并使用基于分布度量的重建误差。作为第二种方法,我们选择评估一种称为可能性后悔的方法,该方法已被证明是一种有效的基于可能性的面向对象评价方法。作为第三种方法,我们评估了另一种最近引入的基于生成建模的方法,称为SSD,它使用自监督表示学习,然后在特征空间中进行基于距离的检测,来计算异常值得分。我们比较了这三种方法,并结合深度估计网络的误差对它们进行了评价。结果表明,虽然基于重建误差的OOD度量不能在所有场景中区分分布内和分布外数据,但基于可能性后悔的OOD度量以及SSD异常值得分在OOD检测中表现相当好。它们的指标也与感知误差高度相关,使它们成为自动驾驶系统可靠性监测器的有希望的候选者。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Comparison of De-Identification Techniques for Privacy Preserving Data Analysis in Vehicular Data Sharing Improved Sensor Model for Realistic Synthetic Data Generation Proposing HEAVENS 2.0 – an automotive risk assessment model Following the White Rabbit: Integrity Verification Based on Risk Analysis Results Real-time Uncertainty Estimation Based On Intermediate Layer Variational Inference
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1