{"title":"基于积分强化学习的非线性多代理系统分布式优化控制","authors":"Ying Xu, Kewen Li, Yongming Li","doi":"10.1002/oca.3174","DOIUrl":null,"url":null,"abstract":"In this article, a distributed optimal control approach is proposed for a class of affine nonlinear multi‐agent systems (MASs) with unknown nonlinear dynamics. The game theory is used to formulate the distributed optimal control problem into a differential graphical game problem with synchronized updates of all nodes. A data‐based integral reinforcement learning (IRL) algorithm is used to learn the solution of the coupled Hamilton–Jacobi (HJ) equation without prior knowledge of the drift dynamics, and the actor‐critic neural networks (A‐C NNs) are used to approximate the control law and the cost function, respectively. To update the parameters synchronously, the gradient descent algorithm is used to design the weight update laws of the A‐C NNs. Combining the IRL and the A‐C NNs, a distributed consensus optimal control method is designed. By using the Lyapunov stability theory, the developed optimal control method can show that all signals in the considered system are uniformly ultimately bounded (UUB), and the systems can achieve Nash equilibrium when all agents update their controllers simultaneously. Finally, simulation results are given to illustrate the effectiveness of the developed optimal control approach.","PeriodicalId":501055,"journal":{"name":"Optimal Control Applications and Methods","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Distributed optimal control of nonlinear multi‐agent systems based on integral reinforcement learning\",\"authors\":\"Ying Xu, Kewen Li, Yongming Li\",\"doi\":\"10.1002/oca.3174\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this article, a distributed optimal control approach is proposed for a class of affine nonlinear multi‐agent systems (MASs) with unknown nonlinear dynamics. The game theory is used to formulate the distributed optimal control problem into a differential graphical game problem with synchronized updates of all nodes. A data‐based integral reinforcement learning (IRL) algorithm is used to learn the solution of the coupled Hamilton–Jacobi (HJ) equation without prior knowledge of the drift dynamics, and the actor‐critic neural networks (A‐C NNs) are used to approximate the control law and the cost function, respectively. To update the parameters synchronously, the gradient descent algorithm is used to design the weight update laws of the A‐C NNs. Combining the IRL and the A‐C NNs, a distributed consensus optimal control method is designed. By using the Lyapunov stability theory, the developed optimal control method can show that all signals in the considered system are uniformly ultimately bounded (UUB), and the systems can achieve Nash equilibrium when all agents update their controllers simultaneously. Finally, simulation results are given to illustrate the effectiveness of the developed optimal control approach.\",\"PeriodicalId\":501055,\"journal\":{\"name\":\"Optimal Control Applications and Methods\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Optimal Control Applications and Methods\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1002/oca.3174\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Optimal Control Applications and Methods","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/oca.3174","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Distributed optimal control of nonlinear multi‐agent systems based on integral reinforcement learning
In this article, a distributed optimal control approach is proposed for a class of affine nonlinear multi‐agent systems (MASs) with unknown nonlinear dynamics. The game theory is used to formulate the distributed optimal control problem into a differential graphical game problem with synchronized updates of all nodes. A data‐based integral reinforcement learning (IRL) algorithm is used to learn the solution of the coupled Hamilton–Jacobi (HJ) equation without prior knowledge of the drift dynamics, and the actor‐critic neural networks (A‐C NNs) are used to approximate the control law and the cost function, respectively. To update the parameters synchronously, the gradient descent algorithm is used to design the weight update laws of the A‐C NNs. Combining the IRL and the A‐C NNs, a distributed consensus optimal control method is designed. By using the Lyapunov stability theory, the developed optimal control method can show that all signals in the considered system are uniformly ultimately bounded (UUB), and the systems can achieve Nash equilibrium when all agents update their controllers simultaneously. Finally, simulation results are given to illustrate the effectiveness of the developed optimal control approach.