Ruien Bian, Xiuchen Jiang, Guoying Zhao, Yadong Liu, Zhou Dai
{"title":"基于多代理强化学习方法的电动汽车可扩展协调能源管理","authors":"Ruien Bian, Xiuchen Jiang, Guoying Zhao, Yadong Liu, Zhou Dai","doi":"10.1155/2024/7765710","DOIUrl":null,"url":null,"abstract":"<div>\n <p>The electric vehicle (EV) has been popular in recent years, which also brings huge challenges to the distribution network due to its energy instability. In order to consider the economic factors of dispatching these distributed renewable resources, the voltage variation is also important. A novel model-free method is put forward for collaborative management of EV resources of aggregators in the distribution network. The economic costs and physical network constraints for this energy management issue are considered at the same time. A Multiagent Deep Deterministic Policy Gradient (MADDPG) algorithm is applied to learn the cooperative energy control strategies. A transfer learning technique is used to fine-tune the trained policy when more aggregators join in the network. The proposed method can achieve close results to the traditional optimization methods, while it takes less than one second to take control actions, making it is more suitable for real-time online energy management. Compared to other advanced reinforcement learning (RL) models, numerical simulations conducted on IEEE test cases greatly illustrate the effectiveness and superiority of the proposed method.</p>\n </div>","PeriodicalId":51293,"journal":{"name":"International Transactions on Electrical Energy Systems","volume":null,"pages":null},"PeriodicalIF":1.9000,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/7765710","citationCount":"0","resultStr":"{\"title\":\"A Scalable and Coordinated Energy Management for Electric Vehicles Based on Multiagent Reinforcement Learning Method\",\"authors\":\"Ruien Bian, Xiuchen Jiang, Guoying Zhao, Yadong Liu, Zhou Dai\",\"doi\":\"10.1155/2024/7765710\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n <p>The electric vehicle (EV) has been popular in recent years, which also brings huge challenges to the distribution network due to its energy instability. In order to consider the economic factors of dispatching these distributed renewable resources, the voltage variation is also important. A novel model-free method is put forward for collaborative management of EV resources of aggregators in the distribution network. The economic costs and physical network constraints for this energy management issue are considered at the same time. A Multiagent Deep Deterministic Policy Gradient (MADDPG) algorithm is applied to learn the cooperative energy control strategies. A transfer learning technique is used to fine-tune the trained policy when more aggregators join in the network. The proposed method can achieve close results to the traditional optimization methods, while it takes less than one second to take control actions, making it is more suitable for real-time online energy management. Compared to other advanced reinforcement learning (RL) models, numerical simulations conducted on IEEE test cases greatly illustrate the effectiveness and superiority of the proposed method.</p>\\n </div>\",\"PeriodicalId\":51293,\"journal\":{\"name\":\"International Transactions on Electrical Energy Systems\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":1.9000,\"publicationDate\":\"2024-10-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/7765710\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Transactions on Electrical Energy Systems\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1155/2024/7765710\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Transactions on Electrical Energy Systems","FirstCategoryId":"5","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1155/2024/7765710","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
A Scalable and Coordinated Energy Management for Electric Vehicles Based on Multiagent Reinforcement Learning Method
The electric vehicle (EV) has been popular in recent years, which also brings huge challenges to the distribution network due to its energy instability. In order to consider the economic factors of dispatching these distributed renewable resources, the voltage variation is also important. A novel model-free method is put forward for collaborative management of EV resources of aggregators in the distribution network. The economic costs and physical network constraints for this energy management issue are considered at the same time. A Multiagent Deep Deterministic Policy Gradient (MADDPG) algorithm is applied to learn the cooperative energy control strategies. A transfer learning technique is used to fine-tune the trained policy when more aggregators join in the network. The proposed method can achieve close results to the traditional optimization methods, while it takes less than one second to take control actions, making it is more suitable for real-time online energy management. Compared to other advanced reinforcement learning (RL) models, numerical simulations conducted on IEEE test cases greatly illustrate the effectiveness and superiority of the proposed method.
期刊介绍:
International Transactions on Electrical Energy Systems publishes original research results on key advances in the generation, transmission, and distribution of electrical energy systems. Of particular interest are submissions concerning the modeling, analysis, optimization and control of advanced electric power systems.
Manuscripts on topics of economics, finance, policies, insulation materials, low-voltage power electronics, plasmas, and magnetics will generally not be considered for review.