混合动力储能系统电动汽车的仿真强化学习能源管理

IF 10.1 1区 工程技术 Q1 ENERGY & FUELS Applied Energy Pub Date : 2024-11-11 DOI:10.1016/j.apenergy.2024.124832
Weirong Liu , Pengfei Yao , Yue Wu , Lijun Duan , Heng Li , Jun Peng
{"title":"混合动力储能系统电动汽车的仿真强化学习能源管理","authors":"Weirong Liu ,&nbsp;Pengfei Yao ,&nbsp;Yue Wu ,&nbsp;Lijun Duan ,&nbsp;Heng Li ,&nbsp;Jun Peng","doi":"10.1016/j.apenergy.2024.124832","DOIUrl":null,"url":null,"abstract":"<div><div>Deep reinforcement learning has become a promising method for the energy management of electric vehicles. However, deep reinforcement learning relies on a large amount of trial-and-error training to acquire near-optimal performance. An adversarial imitation reinforcement learning energy management strategy is proposed for electric vehicles with hybrid energy storage system to minimize the cost of battery capacity loss. Firstly, the reinforcement learning exploration is guided by expert knowledge, which is generated by dynamic programming under various standard driving conditions. The expert knowledge is represented as the optimal power allocation mapping. Secondly, at the early imitation stage, the action of the reinforcement learning agent approaches the optimal power allocation mapping rapidly by using adversarial networks. Thirdly, a dynamic imitation weight is developed according to the Discriminator of adversarial networks, making the agent transit to self-explore the near-optimal power allocation under online driving conditions. Results demonstrate that the proposed strategy can accelerate the training by 42.60% while enhancing the reward by 15.79% compared with traditional reinforcement learning. Under different test driving cycles, the proposed method can further reduce the battery capacity loss cost by 5.1%–12.4%.</div></div>","PeriodicalId":246,"journal":{"name":"Applied Energy","volume":"378 ","pages":"Article 124832"},"PeriodicalIF":10.1000,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Imitation reinforcement learning energy management for electric vehicles with hybrid energy storage system\",\"authors\":\"Weirong Liu ,&nbsp;Pengfei Yao ,&nbsp;Yue Wu ,&nbsp;Lijun Duan ,&nbsp;Heng Li ,&nbsp;Jun Peng\",\"doi\":\"10.1016/j.apenergy.2024.124832\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Deep reinforcement learning has become a promising method for the energy management of electric vehicles. However, deep reinforcement learning relies on a large amount of trial-and-error training to acquire near-optimal performance. An adversarial imitation reinforcement learning energy management strategy is proposed for electric vehicles with hybrid energy storage system to minimize the cost of battery capacity loss. Firstly, the reinforcement learning exploration is guided by expert knowledge, which is generated by dynamic programming under various standard driving conditions. The expert knowledge is represented as the optimal power allocation mapping. Secondly, at the early imitation stage, the action of the reinforcement learning agent approaches the optimal power allocation mapping rapidly by using adversarial networks. Thirdly, a dynamic imitation weight is developed according to the Discriminator of adversarial networks, making the agent transit to self-explore the near-optimal power allocation under online driving conditions. Results demonstrate that the proposed strategy can accelerate the training by 42.60% while enhancing the reward by 15.79% compared with traditional reinforcement learning. Under different test driving cycles, the proposed method can further reduce the battery capacity loss cost by 5.1%–12.4%.</div></div>\",\"PeriodicalId\":246,\"journal\":{\"name\":\"Applied Energy\",\"volume\":\"378 \",\"pages\":\"Article 124832\"},\"PeriodicalIF\":10.1000,\"publicationDate\":\"2024-11-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Energy\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306261924022153\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENERGY & FUELS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Energy","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306261924022153","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENERGY & FUELS","Score":null,"Total":0}
引用次数: 0

摘要

深度强化学习已成为电动汽车能源管理的一种前景广阔的方法。然而,深度强化学习需要依赖大量的试错训练才能获得接近最优的性能。本文提出了一种对抗式模仿强化学习能源管理策略,适用于采用混合储能系统的电动汽车,以最大限度地降低电池容量损失的成本。首先,强化学习探索由专家知识指导,专家知识是在各种标准驾驶条件下通过动态编程生成的。专家知识表现为最优功率分配映射。其次,在早期模仿阶段,强化学习代理的行动通过对抗网络迅速接近最优功率分配映射。第三,根据对抗网络的判别器(Discriminator)开发动态模仿权重,使代理能够在在线驾驶条件下自我探索接近最优的功率分配。结果表明,与传统的强化学习相比,所提出的策略可以加快 42.60% 的训练速度,同时提高 15.79% 的奖励。在不同的测试驾驶周期下,所提出的方法可进一步降低电池容量损耗成本 5.1%-12.4%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Imitation reinforcement learning energy management for electric vehicles with hybrid energy storage system
Deep reinforcement learning has become a promising method for the energy management of electric vehicles. However, deep reinforcement learning relies on a large amount of trial-and-error training to acquire near-optimal performance. An adversarial imitation reinforcement learning energy management strategy is proposed for electric vehicles with hybrid energy storage system to minimize the cost of battery capacity loss. Firstly, the reinforcement learning exploration is guided by expert knowledge, which is generated by dynamic programming under various standard driving conditions. The expert knowledge is represented as the optimal power allocation mapping. Secondly, at the early imitation stage, the action of the reinforcement learning agent approaches the optimal power allocation mapping rapidly by using adversarial networks. Thirdly, a dynamic imitation weight is developed according to the Discriminator of adversarial networks, making the agent transit to self-explore the near-optimal power allocation under online driving conditions. Results demonstrate that the proposed strategy can accelerate the training by 42.60% while enhancing the reward by 15.79% compared with traditional reinforcement learning. Under different test driving cycles, the proposed method can further reduce the battery capacity loss cost by 5.1%–12.4%.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Applied Energy
Applied Energy 工程技术-工程:化工
CiteScore
21.20
自引率
10.70%
发文量
1830
审稿时长
41 days
期刊介绍: Applied Energy serves as a platform for sharing innovations, research, development, and demonstrations in energy conversion, conservation, and sustainable energy systems. The journal covers topics such as optimal energy resource use, environmental pollutant mitigation, and energy process analysis. It welcomes original papers, review articles, technical notes, and letters to the editor. Authors are encouraged to submit manuscripts that bridge the gap between research, development, and implementation. The journal addresses a wide spectrum of topics, including fossil and renewable energy technologies, energy economics, and environmental impacts. Applied Energy also explores modeling and forecasting, conservation strategies, and the social and economic implications of energy policies, including climate change mitigation. It is complemented by the open-access journal Advances in Applied Energy.
期刊最新文献
Boosting the power density of direct borohydride fuel cells to >600 mW cm−2 by cathode water management Editorial Board A distributed thermal-pressure coupling model of large-format lithium iron phosphate battery thermal runaway Optimization and parametric analysis of a novel design of Savonius hydrokinetic turbine using artificial neural network Delay-tolerant hierarchical distributed control for DC microgrid clusters considering microgrid autonomy
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1