{"title":"使用 Q-Learning 和深度 Q-Learning 的在线家庭能源管理系统","authors":"Hasan İzmitligil , Abdurrahman Karamancıoğlu","doi":"10.1016/j.suscom.2024.101005","DOIUrl":null,"url":null,"abstract":"<div><p>The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.</p></div>","PeriodicalId":48686,"journal":{"name":"Sustainable Computing-Informatics & Systems","volume":"43 ","pages":"Article 101005"},"PeriodicalIF":3.8000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Online Home Energy Management System using Q-Learning and Deep Q-Learning\",\"authors\":\"Hasan İzmitligil , Abdurrahman Karamancıoğlu\",\"doi\":\"10.1016/j.suscom.2024.101005\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.</p></div>\",\"PeriodicalId\":48686,\"journal\":{\"name\":\"Sustainable Computing-Informatics & Systems\",\"volume\":\"43 \",\"pages\":\"Article 101005\"},\"PeriodicalIF\":3.8000,\"publicationDate\":\"2024-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Sustainable Computing-Informatics & Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2210537924000507\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Sustainable Computing-Informatics & Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2210537924000507","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0
摘要
由于通信技术和智能计量基础设施的进步,家庭能源管理系统的用户可以安排自己的实时能源消耗。本文提出了一种数据驱动策略,即在线家庭能源管理系统(ON-HEM),它使用强化学习算法(Q-Learning 和 Deep Q-Learning)来控制智能家居系统的最佳能耗。拟议的系统由电力资源(电网、光伏)、通信网络和电器组成,其代理分为四组:可延期、不可延期、功率水平可控和电动汽车。该系统降低了电费成本和高峰需求,同时考虑到了用户对真实数据不满的成本。我们使用 PyCharm 专业版软件对拟议的 ON-HEM 进行了仿真,考虑了不同的定价方法(实时定价和使用时间定价)以及 Q-Learning 和 Deep Q-Learning (DQL) 算法。研究结果表明,DQL 比 Q-Learning 更优越,而且建议的 ON-HEM 在降低高峰需求、电费和客户不满成本方面也很有效。通过使用 IBM SPSS 统计软件将基于模拟的研究结果与实际数据相结合,验证了所建议系统的效率和可靠性。
An Online Home Energy Management System using Q-Learning and Deep Q-Learning
The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.
期刊介绍:
Sustainable computing is a rapidly expanding research area spanning the fields of computer science and engineering, electrical engineering as well as other engineering disciplines. The aim of Sustainable Computing: Informatics and Systems (SUSCOM) is to publish the myriad research findings related to energy-aware and thermal-aware management of computing resource. Equally important is a spectrum of related research issues such as applications of computing that can have ecological and societal impacts. SUSCOM publishes original and timely research papers and survey articles in current areas of power, energy, temperature, and environment related research areas of current importance to readers. SUSCOM has an editorial board comprising prominent researchers from around the world and selects competitively evaluated peer-reviewed papers.