Interpretable AI Agent Through Nonlinear Decision Trees for Lane Change Problem

Abhiroop Ghosh, Yashesh D. Dhebar, Ritam Guha, K. Deb, S. Nageshrao, Ling Zhu, E. Tseng, Dimitar Filev
{"title":"Interpretable AI Agent Through Nonlinear Decision Trees for Lane Change Problem","authors":"Abhiroop Ghosh, Yashesh D. Dhebar, Ritam Guha, K. Deb, S. Nageshrao, Ling Zhu, E. Tseng, Dimitar Filev","doi":"10.1109/SSCI50451.2021.9659552","DOIUrl":null,"url":null,"abstract":"The recent years have witnessed a surge in application of deep neural networks (DNNs) and reinforcement learning (RL) methods to various autonomous control systems and game playing problems. While they are capable of learning from real-world data and produce adequate actions to various state conditions, their internal complexity does not allow an easy way to provide an explanation for their actions. In this paper, we generate state-action pair data from a trained DNN/RL system and employ a previously proposed nonlinear decision tree (NLDT) framework to decipher hidden simplistic rule sets that interpret the working of DNN/RL systems. The complexity of the rule sets are controllable by the user. In essence, the inherent bilevel optimization procedure that finds the NLDTs is capable of reducing the complexities of the state-action logic to a minimalist and intrepretable level. Demonstrating the working principle of the NLDT method to a revised mountain car control problem, this paper applies the methodology to the lane changing problem involving six critical cars in front and rear in left, middle, and right lanes of a pilot car. NLDTs are derived to have simplistic relationships of 12 decision variables involving relative distances and velocities of the six critical cars. The derived analytical decision rules are then simplified further by using a symbolic analysis tool to provide English-like interpretation of the lane change problem. This study makes a scratch to the issue of interpretability of modern machine learning based tools and it now deserves further attention and applications to make the overall approach more integrated and effective.","PeriodicalId":255763,"journal":{"name":"2021 IEEE Symposium Series on Computational Intelligence (SSCI)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE Symposium Series on Computational Intelligence (SSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SSCI50451.2021.9659552","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

The recent years have witnessed a surge in application of deep neural networks (DNNs) and reinforcement learning (RL) methods to various autonomous control systems and game playing problems. While they are capable of learning from real-world data and produce adequate actions to various state conditions, their internal complexity does not allow an easy way to provide an explanation for their actions. In this paper, we generate state-action pair data from a trained DNN/RL system and employ a previously proposed nonlinear decision tree (NLDT) framework to decipher hidden simplistic rule sets that interpret the working of DNN/RL systems. The complexity of the rule sets are controllable by the user. In essence, the inherent bilevel optimization procedure that finds the NLDTs is capable of reducing the complexities of the state-action logic to a minimalist and intrepretable level. Demonstrating the working principle of the NLDT method to a revised mountain car control problem, this paper applies the methodology to the lane changing problem involving six critical cars in front and rear in left, middle, and right lanes of a pilot car. NLDTs are derived to have simplistic relationships of 12 decision variables involving relative distances and velocities of the six critical cars. The derived analytical decision rules are then simplified further by using a symbolic analysis tool to provide English-like interpretation of the lane change problem. This study makes a scratch to the issue of interpretability of modern machine learning based tools and it now deserves further attention and applications to make the overall approach more integrated and effective.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于非线性决策树的可解释AI智能体变道问题研究
近年来,深度神经网络(dnn)和强化学习(RL)方法在各种自主控制系统和游戏问题中的应用激增。虽然它们能够从现实世界的数据中学习,并针对各种状态条件产生适当的动作,但它们内部的复杂性不允许一种简单的方法来解释它们的动作。在本文中,我们从训练有素的DNN/RL系统中生成状态-动作对数据,并采用先前提出的非线性决策树(NLDT)框架来破译解释DNN/RL系统工作的隐藏简化规则集。规则集的复杂度由用户控制。本质上,发现nldt的固有双层优化过程能够将状态-行为逻辑的复杂性降低到最低限度和可解释的水平。本文将NLDT方法的工作原理应用于一个修正的山地车控制问题,并将该方法应用于一辆试验车左、中、右车道前后6辆关键车的变道问题。NLDTs推导为涉及6辆关键汽车的相对距离和速度的12个决策变量的简单关系。然后,通过使用符号分析工具进一步简化导出的分析决策规则,以提供对变道问题的类似英语的解释。本研究触及了基于现代机器学习工具的可解释性问题,值得进一步关注和应用,以使整体方法更加集成和有效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Voice Dialog System for Simulated Patient Robot and Detection of Interviewer Nodding Deep Learning Approaches to Remaining Useful Life Prediction: A Survey Evaluation of Graph Convolutions for Spatio-Temporal Predictions of EV-Charge Availability Balanced K-means using Quantum annealing A Study of Transfer Learning in a Generation Constructive Hyper-Heuristic for One Dimensional Bin Packing
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1