Intelligent Set Speed Estimation for Vehicle Longitudinal Control with Deep Reinforcement Learning

Tobias Eichenlaub, S. Rinderknecht
{"title":"Intelligent Set Speed Estimation for Vehicle Longitudinal Control with Deep Reinforcement Learning","authors":"Tobias Eichenlaub, S. Rinderknecht","doi":"10.1145/3459104.3459123","DOIUrl":null,"url":null,"abstract":"Besides the goal of reducing driving tasks, modern longitudinal control systems also aim to improve fuel efficiency and driver comfort. Most of the vehicles use Adaptive Cruise Control (ACC) systems that track constant set speeds and set headways which makes the trajectory of the vehicle in headway mode highly dependent on the trajectory of a preceding vehicle. Hence, this might lead to increased consumptions in dense traffic situations or when the leader has a less careful driving style. In this work, a method based on Deep Reinforcement Learning (DRL) is presented that finds a control strategy by estimating an intelligent variable set speed based on the system state. Additional control objectives, such as minimizing consumption, are considered explicitly through the feedback in a reward function. A DRL framework is set up that enables the training of a neural set speed estimator for vehicle longitudinal control in a simulative environment. The Deep Deterministic Policy Gradient algorithm is used for the training of the agent. Training is carried out on a simple test track to teach the basic concepts of the control objective to the DRL agent. The learned behavior is then examined in a more complex, stochastic microscopic traffic simulation of the city center of Darmstadt and is compared to a conventional ACC algorithm. The analysis shows that the DRL controller is capable of finding fuel efficient trajectories which are less dependent on the preceding vehicle and is able to generalize to more complex traffic environments, but still shows some unexpected behavior in certain situations. The combination of DRL and conventional models to build up on the existing engineering knowledge is therefore expected to yield promising results in the future.","PeriodicalId":142284,"journal":{"name":"2021 International Symposium on Electrical, Electronics and Information Engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Symposium on Electrical, Electronics and Information Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3459104.3459123","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Besides the goal of reducing driving tasks, modern longitudinal control systems also aim to improve fuel efficiency and driver comfort. Most of the vehicles use Adaptive Cruise Control (ACC) systems that track constant set speeds and set headways which makes the trajectory of the vehicle in headway mode highly dependent on the trajectory of a preceding vehicle. Hence, this might lead to increased consumptions in dense traffic situations or when the leader has a less careful driving style. In this work, a method based on Deep Reinforcement Learning (DRL) is presented that finds a control strategy by estimating an intelligent variable set speed based on the system state. Additional control objectives, such as minimizing consumption, are considered explicitly through the feedback in a reward function. A DRL framework is set up that enables the training of a neural set speed estimator for vehicle longitudinal control in a simulative environment. The Deep Deterministic Policy Gradient algorithm is used for the training of the agent. Training is carried out on a simple test track to teach the basic concepts of the control objective to the DRL agent. The learned behavior is then examined in a more complex, stochastic microscopic traffic simulation of the city center of Darmstadt and is compared to a conventional ACC algorithm. The analysis shows that the DRL controller is capable of finding fuel efficient trajectories which are less dependent on the preceding vehicle and is able to generalize to more complex traffic environments, but still shows some unexpected behavior in certain situations. The combination of DRL and conventional models to build up on the existing engineering knowledge is therefore expected to yield promising results in the future.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于深度强化学习的车辆纵向控制智能设定速度估计
除了减少驾驶任务的目标外,现代纵向控制系统还旨在提高燃油效率和驾驶员舒适度。大多数车辆使用自适应巡航控制(ACC)系统,该系统跟踪恒定的设定速度和设定车头距,这使得车辆在车头距模式下的轨迹高度依赖于前车的轨迹。因此,这可能会导致在交通密集的情况下,或者当领导者的驾驶风格不太小心时,消费增加。本文提出了一种基于深度强化学习(DRL)的方法,该方法通过估计基于系统状态的智能可变集速度来找到控制策略。额外的控制目标,如最小化消耗,是通过奖励函数的反馈明确考虑的。建立了一个DRL框架,用于在模拟环境中训练用于车辆纵向控制的神经集速度估计器。深度确定性策略梯度算法用于智能体的训练。在简单的测试轨道上进行训练,向DRL agent传授控制目标的基本概念。然后在达姆施塔特市中心的一个更复杂、随机的微观交通模拟中检查学习行为,并与传统的ACC算法进行比较。分析表明,DRL控制器能够找到对前车依赖较小的节油轨迹,并且能够推广到更复杂的交通环境中,但在某些情况下仍然会出现一些意外行为。因此,在现有工程知识的基础上,结合DRL和传统模型有望在未来产生有希望的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Exploring the Integration of Blockchain Technology and IoT in a Smart University Application Architecture 3D Moving Rigid Body Localization in the Presence of Anchor Position Errors RANS/LES Simulation of Low-Frequency Flow Oscillations on a NACA0012 Airfoil Near Stall Tuning Language Representation Models for Classification of Turkish News Improving Consumer Experience for Medical Information Using Text Analytics
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1