{"title":"强化学习建筑控制:使用替代模型引导探索的在线方法","authors":"Sourav Dey, Gregor Henze","doi":"10.1115/1.4064842","DOIUrl":null,"url":null,"abstract":"\n The incorporation of emerging technologies, including solar photovoltaics, electric vehicles, battery energy storage, smart devices, internet-of-things (IoT) devices, and sensors in buildings, desirable control objectives are becoming increasingly complex, calling for advanced control approaches. Reinforcement learning (RL) is a powerful method for this, that can adapt and learn from environmental interaction, but it can take a long time to learn and can be unstable initially due to limited environmental knowledge. In our research, we propose an online RL approach for buildings that uses data-driven surrogate models to guide the RL agent during its early training. This helps the controller learn faster and more stably than the traditional direct plug-and-learn online learning approach. In this research, we propose an online approach in buildings with RL where, with the help of data-driven surrogate models, the RL agent is guided during its early exploratory training stage, aiding the controller to learn a near-optimal policy faster and exhibiting more stable training progress than a traditional direct plug-and-learn online learning RL approach. The agents are assisted in their learning and action with information gained from the surrogate models generating multiple artificial trajectories starting from the current state. The research presented an exploration of various surrogate model-assisted training methods and revealed that models focusing on artificial trajectories around rule-based controls yielded the most stable performance. In contrast, models employing random exploration with a one-step look-ahead approach demonstrated superior overall performance.","PeriodicalId":326594,"journal":{"name":"ASME Journal of Engineering for Sustainable Buildings and Cities","volume":"6 6","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Reinforcement Learning Building Control: An Online Approach with Guided Exploration using Surrogate Models\",\"authors\":\"Sourav Dey, Gregor Henze\",\"doi\":\"10.1115/1.4064842\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n The incorporation of emerging technologies, including solar photovoltaics, electric vehicles, battery energy storage, smart devices, internet-of-things (IoT) devices, and sensors in buildings, desirable control objectives are becoming increasingly complex, calling for advanced control approaches. Reinforcement learning (RL) is a powerful method for this, that can adapt and learn from environmental interaction, but it can take a long time to learn and can be unstable initially due to limited environmental knowledge. In our research, we propose an online RL approach for buildings that uses data-driven surrogate models to guide the RL agent during its early training. This helps the controller learn faster and more stably than the traditional direct plug-and-learn online learning approach. In this research, we propose an online approach in buildings with RL where, with the help of data-driven surrogate models, the RL agent is guided during its early exploratory training stage, aiding the controller to learn a near-optimal policy faster and exhibiting more stable training progress than a traditional direct plug-and-learn online learning RL approach. The agents are assisted in their learning and action with information gained from the surrogate models generating multiple artificial trajectories starting from the current state. The research presented an exploration of various surrogate model-assisted training methods and revealed that models focusing on artificial trajectories around rule-based controls yielded the most stable performance. In contrast, models employing random exploration with a one-step look-ahead approach demonstrated superior overall performance.\",\"PeriodicalId\":326594,\"journal\":{\"name\":\"ASME Journal of Engineering for Sustainable Buildings and Cities\",\"volume\":\"6 6\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ASME Journal of Engineering for Sustainable Buildings and Cities\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1115/1.4064842\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ASME Journal of Engineering for Sustainable Buildings and Cities","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1115/1.4064842","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Reinforcement Learning Building Control: An Online Approach with Guided Exploration using Surrogate Models
The incorporation of emerging technologies, including solar photovoltaics, electric vehicles, battery energy storage, smart devices, internet-of-things (IoT) devices, and sensors in buildings, desirable control objectives are becoming increasingly complex, calling for advanced control approaches. Reinforcement learning (RL) is a powerful method for this, that can adapt and learn from environmental interaction, but it can take a long time to learn and can be unstable initially due to limited environmental knowledge. In our research, we propose an online RL approach for buildings that uses data-driven surrogate models to guide the RL agent during its early training. This helps the controller learn faster and more stably than the traditional direct plug-and-learn online learning approach. In this research, we propose an online approach in buildings with RL where, with the help of data-driven surrogate models, the RL agent is guided during its early exploratory training stage, aiding the controller to learn a near-optimal policy faster and exhibiting more stable training progress than a traditional direct plug-and-learn online learning RL approach. The agents are assisted in their learning and action with information gained from the surrogate models generating multiple artificial trajectories starting from the current state. The research presented an exploration of various surrogate model-assisted training methods and revealed that models focusing on artificial trajectories around rule-based controls yielded the most stable performance. In contrast, models employing random exploration with a one-step look-ahead approach demonstrated superior overall performance.