MPC-Guided Deep Reinforcement Learning for Optimal Charging of Lithium-Ion Battery With Uncertainty

IF 8.3 1区 工程技术 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC IEEE Transactions on Transportation Electrification Pub Date : 2024-09-17 DOI:10.1109/TTE.2024.3462769
Zhipeng Zhu;Guangzhong Dong;Yunjiang Lou;Li Sun;Jincheng Yu;Liangcai Wu;Jingwen Wei
{"title":"MPC-Guided Deep Reinforcement Learning for Optimal Charging of Lithium-Ion Battery With Uncertainty","authors":"Zhipeng Zhu;Guangzhong Dong;Yunjiang Lou;Li Sun;Jincheng Yu;Liangcai Wu;Jingwen Wei","doi":"10.1109/TTE.2024.3462769","DOIUrl":null,"url":null,"abstract":"Ensuring the safe and fast charging of lithium-ion battery (LIB) is a pivotal technology that plays a key role in advancing the wide application of electric vehicles (EVs). Currently, the majority of model-based charging methods are developed for deterministic models, lacking consideration for strategy failure and battery safety issues caused by model or data uncertainty. Learning-based charging methods can address this issue due to their strong adaptability. However, training appropriate strategies requires a mass of iterative interaction. In this article, a model predictive control (MPC)-guided deep reinforcement learning (DRL) charging scheme is proposed to address the control challenges resulting from model uncertainty or additional disturbances. By integrating the advantages of both MPC and DRL, the scheme can not only solve the problem of performance degradation caused by uncertainty in model-based methods, but also reduce the search space of DRL to improve the sample efficiency of learning-based methods. The proposed strategy is compared with state-of-the-art standalone MPC and DRL controllers. Results show that the MPC-based controller inevitably violates constraints, while controllers under DRL framework successfully reduce the voltage violation rate from 34.28% to 0%. Compared to the standalone DRL controller, the proposed strategy converges approximately 60% faster. The average charging time is reduced by 1.96, 2.23, and 0.36 min after 500, 1000, and 1500 training episodes, respectively. Additionally, the proposed strategy ensures a safer training process.","PeriodicalId":56269,"journal":{"name":"IEEE Transactions on Transportation Electrification","volume":"11 1","pages":"4408-4419"},"PeriodicalIF":8.3000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Transportation Electrification","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10681475/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Ensuring the safe and fast charging of lithium-ion battery (LIB) is a pivotal technology that plays a key role in advancing the wide application of electric vehicles (EVs). Currently, the majority of model-based charging methods are developed for deterministic models, lacking consideration for strategy failure and battery safety issues caused by model or data uncertainty. Learning-based charging methods can address this issue due to their strong adaptability. However, training appropriate strategies requires a mass of iterative interaction. In this article, a model predictive control (MPC)-guided deep reinforcement learning (DRL) charging scheme is proposed to address the control challenges resulting from model uncertainty or additional disturbances. By integrating the advantages of both MPC and DRL, the scheme can not only solve the problem of performance degradation caused by uncertainty in model-based methods, but also reduce the search space of DRL to improve the sample efficiency of learning-based methods. The proposed strategy is compared with state-of-the-art standalone MPC and DRL controllers. Results show that the MPC-based controller inevitably violates constraints, while controllers under DRL framework successfully reduce the voltage violation rate from 34.28% to 0%. Compared to the standalone DRL controller, the proposed strategy converges approximately 60% faster. The average charging time is reduced by 1.96, 2.23, and 0.36 min after 500, 1000, and 1500 training episodes, respectively. Additionally, the proposed strategy ensures a safer training process.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
以 MPC 为指导的深度强化学习实现锂离子电池的不确定性优化充电
保证锂离子电池的安全快速充电是推动电动汽车广泛应用的关键技术。目前,大多数基于模型的充电方法都是针对确定性模型开发的,缺乏对模型或数据不确定性导致的策略失效和电池安全问题的考虑。基于学习的收费方法具有较强的适应性,可以解决这一问题。然而,训练合适的策略需要大量的迭代交互。本文提出了一种模型预测控制(MPC)引导的深度强化学习(DRL)充电方案,以解决模型不确定性或额外干扰带来的控制挑战。该方案综合了MPC和DRL的优点,既解决了基于模型的方法由于不确定性导致的性能下降问题,又缩小了DRL的搜索空间,提高了基于学习的方法的样本效率。提出的策略与最先进的独立MPC和DRL控制器进行了比较。结果表明,基于mpc的控制器不可避免地违反约束,而DRL框架下的控制器成功地将电压违反率从34.28%降低到0%。与独立DRL控制器相比,该策略的收敛速度提高了约60%。在训练500、1000和1500集后,平均充电时间分别减少1.96、2.23和0.36分钟。此外,拟议的战略确保了更安全的培训过程。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Transactions on Transportation Electrification
IEEE Transactions on Transportation Electrification Engineering-Electrical and Electronic Engineering
CiteScore
12.20
自引率
15.70%
发文量
449
期刊介绍: IEEE Transactions on Transportation Electrification is focused on components, sub-systems, systems, standards, and grid interface technologies related to power and energy conversion, propulsion, and actuation for all types of electrified vehicles including on-road, off-road, off-highway, and rail vehicles, airplanes, and ships.
期刊最新文献
Modeling, Evaluation, and Comparison of Battery Equalizers Based on the Conservation of Energy Asymptotic Homogenization Method to Estimate Equivalent Thermal Conductivity of Electrical Machine Windings Nonlinear Bounded Error Compensation for Flux based Encoderless Controller of PMSMs Design of a Wound Rotor Brushless Doubly-Fed Machine with Even Pole-Pair Ratio Using Superposition Method High-Speed Machine Design for Electric Turbochargers Based on General Air-gap Field Modulation Theory
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1