{"title":"Control Law Learning Based on LQR Reconstruction With Inverse Optimal Control","authors":"Chendi Qu;Jianping He;Xiaoming Duan","doi":"10.1109/TAC.2024.3469788","DOIUrl":null,"url":null,"abstract":"Designing controllers to generate various trajectories has been studied for years, while recently, recovering an optimal controller from trajectories receives increasing attention. In this article, we reveal that the inherent linear quadratic regulator (LQR) problem of a moving agent can be reconstructed based on its trajectory observations only, which enables one to learn the control law of the target agent autonomously. Specifically, we propose a novel inverse optimal control method to identify the weighting matrices of a discrete-time finite horizon LQR, and we also provide the corresponding identifiability conditions. Then, we obtain the optimal estimate of the control horizon using binary search, and finally, reconstruct the LQR problem with aforementioned estimates. The strength of the learning control law with optimization problem recovery lies in less computation consumption and strong generalization ability. We apply our algorithm to the future control input prediction and the discrepancy loss is further derived. Simulations and hardware experiments on a self-designed robot platform illustrate the effectiveness of our work.","PeriodicalId":13201,"journal":{"name":"IEEE Transactions on Automatic Control","volume":"70 2","pages":"1350-1357"},"PeriodicalIF":7.0000,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automatic Control","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10697272/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Designing controllers to generate various trajectories has been studied for years, while recently, recovering an optimal controller from trajectories receives increasing attention. In this article, we reveal that the inherent linear quadratic regulator (LQR) problem of a moving agent can be reconstructed based on its trajectory observations only, which enables one to learn the control law of the target agent autonomously. Specifically, we propose a novel inverse optimal control method to identify the weighting matrices of a discrete-time finite horizon LQR, and we also provide the corresponding identifiability conditions. Then, we obtain the optimal estimate of the control horizon using binary search, and finally, reconstruct the LQR problem with aforementioned estimates. The strength of the learning control law with optimization problem recovery lies in less computation consumption and strong generalization ability. We apply our algorithm to the future control input prediction and the discrepancy loss is further derived. Simulations and hardware experiments on a self-designed robot platform illustrate the effectiveness of our work.
期刊介绍:
In the IEEE Transactions on Automatic Control, the IEEE Control Systems Society publishes high-quality papers on the theory, design, and applications of control engineering. Two types of contributions are regularly considered:
1) Papers: Presentation of significant research, development, or application of control concepts.
2) Technical Notes and Correspondence: Brief technical notes, comments on published areas or established control topics, corrections to papers and notes published in the Transactions.
In addition, special papers (tutorials, surveys, and perspectives on the theory and applications of control systems topics) are solicited.