使用 Q-Learning 和深度 Q-Learning 的在线家庭能源管理系统

IF 3.8 3区 计算机科学 Q1 COMPUTER SCIENCE, HARDWARE & ARCHITECTURE Sustainable Computing-Informatics & Systems Pub Date : 2024-06-01 DOI:10.1016/j.suscom.2024.101005
Hasan İzmitligil , Abdurrahman Karamancıoğlu
{"title":"使用 Q-Learning 和深度 Q-Learning 的在线家庭能源管理系统","authors":"Hasan İzmitligil ,&nbsp;Abdurrahman Karamancıoğlu","doi":"10.1016/j.suscom.2024.101005","DOIUrl":null,"url":null,"abstract":"<div><p>The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.</p></div>","PeriodicalId":48686,"journal":{"name":"Sustainable Computing-Informatics & Systems","volume":"43 ","pages":"Article 101005"},"PeriodicalIF":3.8000,"publicationDate":"2024-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Online Home Energy Management System using Q-Learning and Deep Q-Learning\",\"authors\":\"Hasan İzmitligil ,&nbsp;Abdurrahman Karamancıoğlu\",\"doi\":\"10.1016/j.suscom.2024.101005\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.</p></div>\",\"PeriodicalId\":48686,\"journal\":{\"name\":\"Sustainable Computing-Informatics & Systems\",\"volume\":\"43 \",\"pages\":\"Article 101005\"},\"PeriodicalIF\":3.8000,\"publicationDate\":\"2024-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Sustainable Computing-Informatics & Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2210537924000507\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Sustainable Computing-Informatics & Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2210537924000507","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
引用次数: 0

摘要

由于通信技术和智能计量基础设施的进步,家庭能源管理系统的用户可以安排自己的实时能源消耗。本文提出了一种数据驱动策略,即在线家庭能源管理系统(ON-HEM),它使用强化学习算法(Q-Learning 和 Deep Q-Learning)来控制智能家居系统的最佳能耗。拟议的系统由电力资源(电网、光伏)、通信网络和电器组成,其代理分为四组:可延期、不可延期、功率水平可控和电动汽车。该系统降低了电费成本和高峰需求,同时考虑到了用户对真实数据不满的成本。我们使用 PyCharm 专业版软件对拟议的 ON-HEM 进行了仿真,考虑了不同的定价方法(实时定价和使用时间定价)以及 Q-Learning 和 Deep Q-Learning (DQL) 算法。研究结果表明,DQL 比 Q-Learning 更优越,而且建议的 ON-HEM 在降低高峰需求、电费和客户不满成本方面也很有效。通过使用 IBM SPSS 统计软件将基于模拟的研究结果与实际数据相结合,验证了所建议系统的效率和可靠性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
An Online Home Energy Management System using Q-Learning and Deep Q-Learning

The users of home energy management systems schedule their real-time energy consumption thanks to advancements in communication technology and smart metering infrastructures. In this paper, a data-driven strategy is proposed, which is an Online Home Energy Management System (ON-HEM) that uses reinforcement learning algorithms (Q-Learning and Deep Q-Learning) to control the optimal energy consumption of a smart home system. The proposed system comprises power resources (grid, photovoltaic), communication networks, and appliances with their agents classified into four groups: deferrable, non-deferrable, power level controllable, and electric vehicle. The system reduces electricity costs and high peak demands while considering the cost of user dissatisfaction with real-life data. Simulations are performed on the proposed ON-HEM considering different pricing approaches (Real Time Pricing and Time of Use Pricing) with Q-Learning and Deep Q-Learning (DQL) algorithms using PyCharm Professional Edition software. The findings demonstrate both the superiority of DQL over Q-Learning and the efficiency of the proposed ON-HEM in decreasing high peak demand, electricity costs, and customer dissatisfaction costs. The efficiency and dependability of the proposed system were verified by utilizing simulation-based findings with real-life data using IBM SPSS Statistics software.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Sustainable Computing-Informatics & Systems
Sustainable Computing-Informatics & Systems COMPUTER SCIENCE, HARDWARE & ARCHITECTUREC-COMPUTER SCIENCE, INFORMATION SYSTEMS
CiteScore
10.70
自引率
4.40%
发文量
142
期刊介绍: Sustainable computing is a rapidly expanding research area spanning the fields of computer science and engineering, electrical engineering as well as other engineering disciplines. The aim of Sustainable Computing: Informatics and Systems (SUSCOM) is to publish the myriad research findings related to energy-aware and thermal-aware management of computing resource. Equally important is a spectrum of related research issues such as applications of computing that can have ecological and societal impacts. SUSCOM publishes original and timely research papers and survey articles in current areas of power, energy, temperature, and environment related research areas of current importance to readers. SUSCOM has an editorial board comprising prominent researchers from around the world and selects competitively evaluated peer-reviewed papers.
期刊最新文献
Editorial Board Secured and energy efficient cluster based routing in WSN via hybrid optimization model, TICOA Multiobjective hybrid Al-Biruni Earth Namib Beetle Optimization and deep learning based task scheduling in cloud computing Analysing the radiation reliability, performance and energy consumption of low-power SoC through heterogeneous parallelism Nearest data processing in GPU
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1