Caio Fabio Oliveira da Silva, Azita Dabiri, Bart De Schutter
{"title":"集成强化学习和模型预测控制,并将其应用于微网","authors":"Caio Fabio Oliveira da Silva, Azita Dabiri, Bart De Schutter","doi":"arxiv-2409.11267","DOIUrl":null,"url":null,"abstract":"This work proposes an approach that integrates reinforcement learning and\nmodel predictive control (MPC) to efficiently solve finite-horizon optimal\ncontrol problems in mixed-logical dynamical systems. Optimization-based control\nof such systems with discrete and continuous decision variables entails the\nonline solution of mixed-integer quadratic or linear programs, which suffer\nfrom the curse of dimensionality. Our approach aims at mitigating this issue by\neffectively decoupling the decision on the discrete variables and the decision\non the continuous variables. Moreover, to mitigate the combinatorial growth in\nthe number of possible actions due to the prediction horizon, we conceive the\ndefinition of decoupled Q-functions to make the learning problem more\ntractable. The use of reinforcement learning reduces the online optimization\nproblem of the MPC controller from a mixed-integer linear (quadratic) program\nto a linear (quadratic) program, greatly reducing the computational time.\nSimulation experiments for a microgrid, based on real-world data, demonstrate\nthat the proposed method significantly reduces the online computation time of\nthe MPC approach and that it generates policies with small optimality gaps and\nhigh feasibility rates.","PeriodicalId":501175,"journal":{"name":"arXiv - EE - Systems and Control","volume":"9 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Integrating Reinforcement Learning and Model Predictive Control with Applications to Microgrids\",\"authors\":\"Caio Fabio Oliveira da Silva, Azita Dabiri, Bart De Schutter\",\"doi\":\"arxiv-2409.11267\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work proposes an approach that integrates reinforcement learning and\\nmodel predictive control (MPC) to efficiently solve finite-horizon optimal\\ncontrol problems in mixed-logical dynamical systems. Optimization-based control\\nof such systems with discrete and continuous decision variables entails the\\nonline solution of mixed-integer quadratic or linear programs, which suffer\\nfrom the curse of dimensionality. Our approach aims at mitigating this issue by\\neffectively decoupling the decision on the discrete variables and the decision\\non the continuous variables. Moreover, to mitigate the combinatorial growth in\\nthe number of possible actions due to the prediction horizon, we conceive the\\ndefinition of decoupled Q-functions to make the learning problem more\\ntractable. The use of reinforcement learning reduces the online optimization\\nproblem of the MPC controller from a mixed-integer linear (quadratic) program\\nto a linear (quadratic) program, greatly reducing the computational time.\\nSimulation experiments for a microgrid, based on real-world data, demonstrate\\nthat the proposed method significantly reduces the online computation time of\\nthe MPC approach and that it generates policies with small optimality gaps and\\nhigh feasibility rates.\",\"PeriodicalId\":501175,\"journal\":{\"name\":\"arXiv - EE - Systems and Control\",\"volume\":\"9 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - EE - Systems and Control\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.11267\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - EE - Systems and Control","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11267","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Integrating Reinforcement Learning and Model Predictive Control with Applications to Microgrids
This work proposes an approach that integrates reinforcement learning and
model predictive control (MPC) to efficiently solve finite-horizon optimal
control problems in mixed-logical dynamical systems. Optimization-based control
of such systems with discrete and continuous decision variables entails the
online solution of mixed-integer quadratic or linear programs, which suffer
from the curse of dimensionality. Our approach aims at mitigating this issue by
effectively decoupling the decision on the discrete variables and the decision
on the continuous variables. Moreover, to mitigate the combinatorial growth in
the number of possible actions due to the prediction horizon, we conceive the
definition of decoupled Q-functions to make the learning problem more
tractable. The use of reinforcement learning reduces the online optimization
problem of the MPC controller from a mixed-integer linear (quadratic) program
to a linear (quadratic) program, greatly reducing the computational time.
Simulation experiments for a microgrid, based on real-world data, demonstrate
that the proposed method significantly reduces the online computation time of
the MPC approach and that it generates policies with small optimality gaps and
high feasibility rates.