Reinforcement learning based two‐timescale energy management for energy hub

Jinfan Chen, C. Mao, Guanglin Sha, Wanxing Sheng, Hua Fan, Dan Wang, Shushan Qiu, Yunzhao Wu, Yao Zhang
{"title":"Reinforcement learning based two‐timescale energy management for energy hub","authors":"Jinfan Chen, C. Mao, Guanglin Sha, Wanxing Sheng, Hua Fan, Dan Wang, Shushan Qiu, Yunzhao Wu, Yao Zhang","doi":"10.1049/rpg2.12911","DOIUrl":null,"url":null,"abstract":"Maintaining energy balance and economical operation is significant for energy hub (EH) which serves as the central component. Implementing real‐time regulation for heating and cooling equipment within the EH is challenging due to their slow response time in response to the stochastic fluctuation in renewable energy sources and demands while the opposite is true for electric energy storage equipment (EST), a conventional single timescale energy management strategy is no longer sufficient to take into account the operating characteristics of all equipment. With this motivation, this study proposes a deep reinforcement learning based two‐timescale energy management strategy for EH, which controls heating & cooling equipment on a long timescale of 1 h, and EST on a short timescale of 15 min. The actions of the EST are modelled as discrete to reduce the action spaces, and the discrete‐continuous hybrid action sequential TD3 model is proposed to address the problem of handling both discrete and continuous actions in long timescale policy. A joint training approach based on the centralized training framework is proposed to learn multiple levels of policies in parallel. The case studies demonstrate that the proposed strategy reduces the economic cost and carbon emissions by 1%, and 0.5% compared to the single time‐scale strategy respectively.","PeriodicalId":507938,"journal":{"name":"IET Renewable Power Generation","volume":"45 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Renewable Power Generation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1049/rpg2.12911","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Maintaining energy balance and economical operation is significant for energy hub (EH) which serves as the central component. Implementing real‐time regulation for heating and cooling equipment within the EH is challenging due to their slow response time in response to the stochastic fluctuation in renewable energy sources and demands while the opposite is true for electric energy storage equipment (EST), a conventional single timescale energy management strategy is no longer sufficient to take into account the operating characteristics of all equipment. With this motivation, this study proposes a deep reinforcement learning based two‐timescale energy management strategy for EH, which controls heating & cooling equipment on a long timescale of 1 h, and EST on a short timescale of 15 min. The actions of the EST are modelled as discrete to reduce the action spaces, and the discrete‐continuous hybrid action sequential TD3 model is proposed to address the problem of handling both discrete and continuous actions in long timescale policy. A joint training approach based on the centralized training framework is proposed to learn multiple levels of policies in parallel. The case studies demonstrate that the proposed strategy reduces the economic cost and carbon emissions by 1%, and 0.5% compared to the single time‐scale strategy respectively.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于强化学习的能源枢纽双时标能源管理
保持能源平衡和经济运行对于作为核心部件的能源枢纽(EH)来说意义重大。对 EH 内的加热和冷却设备实施实时调节具有挑战性,因为它们对可再生能源和需求的随机波动响应时间较慢,而对电力储能设备(EST)则恰恰相反,传统的单一时间尺度能源管理策略已不足以考虑所有设备的运行特性。基于这一动机,本研究提出了一种基于深度强化学习的双时间尺度 EH 能源管理策略,即在 1 小时的长时间尺度上控制加热和冷却设备,在 15 分钟的短时间尺度上控制 EST。为减少行动空间,将 EST 的行动建模为离散行动,并提出了离散-连续混合行动序列 TD3 模型,以解决在长时间尺度策略中同时处理离散和连续行动的问题。提出了一种基于集中训练框架的联合训练方法,以并行学习多层次的策略。案例研究表明,与单一时间尺度策略相比,所提出的策略分别降低了 1%和 0.5%的经济成本和碳排放量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Two‐stage power control method of rotary power flow controller for flexible interconnection of distribution network A smooth method for primary frequency of wind turbine based on variable power point tracking and energy storage cooperative control Estimation of maximum non‐synchronous generation of renewable energy in the South Korea power system based on the minimum level of inertia Reinforcement learning based two‐timescale energy management for energy hub A novel deep learning technique to detect electricity theft in smart grids using AlexNet
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1