Anastasia Yu. Zhadan, Haitao Wu, Pavel S. Kudin, Yuyi Zhang, Ovanes L. Petrosian
{"title":"基于深度强化学习和数值优化方法的可再生能源微电网控制","authors":"Anastasia Yu. Zhadan, Haitao Wu, Pavel S. Kudin, Yuyi Zhang, Ovanes L. Petrosian","doi":"10.21638/11701/spbu10.2023.307","DOIUrl":null,"url":null,"abstract":"Optimal scheduling of battery energy storage system plays crucial part in dis- tributed energy system. As a data driven method, deep reinforcement learning does not require system knowledge of dynamic system, present optimal solution for nonlinear optimization problem. In this research, financial cost of energy con- sumption reduced by scheduling battery energy using deep reinforcement learning method (RL). Reinforcement learning can adapt to equipment parameter changes and noise in the data, while mixed-integer linear programming (MILP) requires high accuracy in forecasting power generation and demand, accurate equipment parameters to achieve good performance, and high computational cost for large- scale industrial applications. Based on this, it can be assumed that deep RL based solution is capable of outperform classic deterministic optimization model MILP. This study compares four state-of-the-art RL algorithms for the battery power plant control problem: PPO, A2C, SAC, TD3. According to the simulation results, TD3 shows the best results, outperforming MILP by 5% in cost savings, and the time to solve the problem is reduced by about a factor of three.","PeriodicalId":477285,"journal":{"name":"Вестник Санкт-Петербургского университета","volume":"52 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Microgrid control for renewable energy sources based on deep reinforcement learning and numerical optimization approaches\",\"authors\":\"Anastasia Yu. Zhadan, Haitao Wu, Pavel S. Kudin, Yuyi Zhang, Ovanes L. Petrosian\",\"doi\":\"10.21638/11701/spbu10.2023.307\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Optimal scheduling of battery energy storage system plays crucial part in dis- tributed energy system. As a data driven method, deep reinforcement learning does not require system knowledge of dynamic system, present optimal solution for nonlinear optimization problem. In this research, financial cost of energy con- sumption reduced by scheduling battery energy using deep reinforcement learning method (RL). Reinforcement learning can adapt to equipment parameter changes and noise in the data, while mixed-integer linear programming (MILP) requires high accuracy in forecasting power generation and demand, accurate equipment parameters to achieve good performance, and high computational cost for large- scale industrial applications. Based on this, it can be assumed that deep RL based solution is capable of outperform classic deterministic optimization model MILP. This study compares four state-of-the-art RL algorithms for the battery power plant control problem: PPO, A2C, SAC, TD3. According to the simulation results, TD3 shows the best results, outperforming MILP by 5% in cost savings, and the time to solve the problem is reduced by about a factor of three.\",\"PeriodicalId\":477285,\"journal\":{\"name\":\"Вестник Санкт-Петербургского университета\",\"volume\":\"52 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Вестник Санкт-Петербургского университета\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.21638/11701/spbu10.2023.307\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Вестник Санкт-Петербургского университета","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21638/11701/spbu10.2023.307","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Microgrid control for renewable energy sources based on deep reinforcement learning and numerical optimization approaches
Optimal scheduling of battery energy storage system plays crucial part in dis- tributed energy system. As a data driven method, deep reinforcement learning does not require system knowledge of dynamic system, present optimal solution for nonlinear optimization problem. In this research, financial cost of energy con- sumption reduced by scheduling battery energy using deep reinforcement learning method (RL). Reinforcement learning can adapt to equipment parameter changes and noise in the data, while mixed-integer linear programming (MILP) requires high accuracy in forecasting power generation and demand, accurate equipment parameters to achieve good performance, and high computational cost for large- scale industrial applications. Based on this, it can be assumed that deep RL based solution is capable of outperform classic deterministic optimization model MILP. This study compares four state-of-the-art RL algorithms for the battery power plant control problem: PPO, A2C, SAC, TD3. According to the simulation results, TD3 shows the best results, outperforming MILP by 5% in cost savings, and the time to solve the problem is reduced by about a factor of three.