残差深度强化学习与基于模型的逆变器电压-电压控制优化

IF 8.6 1区 工程技术 Q1 ENERGY & FUELS IEEE Transactions on Sustainable Energy Pub Date : 2024-09-03 DOI:10.1109/TSTE.2024.3454080
Qiong Liu;Ye Guo;Lirong Deng;Haotian Liu;Dongyu Li;Hongbin Sun
{"title":"残差深度强化学习与基于模型的逆变器电压-电压控制优化","authors":"Qiong Liu;Ye Guo;Lirong Deng;Haotian Liu;Dongyu Li;Hongbin Sun","doi":"10.1109/TSTE.2024.3454080","DOIUrl":null,"url":null,"abstract":"A residual deep reinforcement learning (RDRL) based on an approximate-model-driven optimization approach is proposed for inverter-based volt-var control (IB-VVC) in active distribution networks. A modified Markov decision process is introduced to formulate the model-based and RDRL-based IB-VVC simultaneously, and then RDRL learns a residual action based on the action of the model-based approach with an approximate model. It inherits the control capability of the approximate-model-based optimization and enhances the policy optimization capability by residual policy learning. Since the approximate model acquired by operators is generally relatively reliable, the action solved by model-based optimization approaches is not far away from the optimal one. This allows RDRL to search for the residual action in a smaller residual action space, which further improves the approximation accuracy of the critic and reduces the search difficulties of the actor. Simulations demonstrate that RDRL improves the optimization performance considerably throughout the learning stage and verifies their three rationales for superior performance point-by-point on 69 and 141 bus balanced distribution networks.","PeriodicalId":452,"journal":{"name":"IEEE Transactions on Sustainable Energy","volume":"16 1","pages":"269-283"},"PeriodicalIF":8.6000,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Residual Deep Reinforcement Learning With Model-Based Optimization for Inverter-Based Volt-Var Control\",\"authors\":\"Qiong Liu;Ye Guo;Lirong Deng;Haotian Liu;Dongyu Li;Hongbin Sun\",\"doi\":\"10.1109/TSTE.2024.3454080\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A residual deep reinforcement learning (RDRL) based on an approximate-model-driven optimization approach is proposed for inverter-based volt-var control (IB-VVC) in active distribution networks. A modified Markov decision process is introduced to formulate the model-based and RDRL-based IB-VVC simultaneously, and then RDRL learns a residual action based on the action of the model-based approach with an approximate model. It inherits the control capability of the approximate-model-based optimization and enhances the policy optimization capability by residual policy learning. Since the approximate model acquired by operators is generally relatively reliable, the action solved by model-based optimization approaches is not far away from the optimal one. This allows RDRL to search for the residual action in a smaller residual action space, which further improves the approximation accuracy of the critic and reduces the search difficulties of the actor. Simulations demonstrate that RDRL improves the optimization performance considerably throughout the learning stage and verifies their three rationales for superior performance point-by-point on 69 and 141 bus balanced distribution networks.\",\"PeriodicalId\":452,\"journal\":{\"name\":\"IEEE Transactions on Sustainable Energy\",\"volume\":\"16 1\",\"pages\":\"269-283\"},\"PeriodicalIF\":8.6000,\"publicationDate\":\"2024-09-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Sustainable Energy\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10663699/\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENERGY & FUELS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Sustainable Energy","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10663699/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENERGY & FUELS","Score":null,"Total":0}
引用次数: 0

摘要

提出了一种基于近似模型驱动优化的残差深度强化学习(RDRL)方法,用于主动配电网中基于逆变器的电压无功控制(IB-VVC)。引入改进的马尔可夫决策过程,将基于模型的IB-VVC和基于RDRL的IB-VVC同时制定,RDRL在基于模型的方法的动作基础上,用近似模型学习残差动作。它继承了基于近似模型优化的控制能力,并通过残差策略学习增强了策略优化能力。由于算子获得的近似模型一般是比较可靠的,所以基于模型的优化方法所解出的动作离最优动作也不远。这使得RDRL可以在更小的残余动作空间中搜索残余动作,进一步提高了评论家的近似精度,降低了行动者的搜索难度。仿真结果表明,RDRL在整个学习阶段显著提高了优化性能,并在69和141总线均衡配电网上逐点验证了其优越性能的三个基本原理。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Residual Deep Reinforcement Learning With Model-Based Optimization for Inverter-Based Volt-Var Control
A residual deep reinforcement learning (RDRL) based on an approximate-model-driven optimization approach is proposed for inverter-based volt-var control (IB-VVC) in active distribution networks. A modified Markov decision process is introduced to formulate the model-based and RDRL-based IB-VVC simultaneously, and then RDRL learns a residual action based on the action of the model-based approach with an approximate model. It inherits the control capability of the approximate-model-based optimization and enhances the policy optimization capability by residual policy learning. Since the approximate model acquired by operators is generally relatively reliable, the action solved by model-based optimization approaches is not far away from the optimal one. This allows RDRL to search for the residual action in a smaller residual action space, which further improves the approximation accuracy of the critic and reduces the search difficulties of the actor. Simulations demonstrate that RDRL improves the optimization performance considerably throughout the learning stage and verifies their three rationales for superior performance point-by-point on 69 and 141 bus balanced distribution networks.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
IEEE Transactions on Sustainable Energy
IEEE Transactions on Sustainable Energy ENERGY & FUELS-ENGINEERING, ELECTRICAL & ELECTRONIC
CiteScore
21.40
自引率
5.70%
发文量
215
审稿时长
5 months
期刊介绍: The IEEE Transactions on Sustainable Energy serves as a pivotal platform for sharing groundbreaking research findings on sustainable energy systems, with a focus on their seamless integration into power transmission and/or distribution grids. The journal showcases original research spanning the design, implementation, grid-integration, and control of sustainable energy technologies and systems. Additionally, the Transactions warmly welcomes manuscripts addressing the design, implementation, and evaluation of power systems influenced by sustainable energy systems and devices.
期刊最新文献
Table of Contents IEEE Collabratec Get Published in the New IEEE Open Access Journal of Power and Energy Share Your Preprint Research with the World! IEEE Transactions on Sustainable Energy Information for Authors
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1