{"title":"Model-Free Tracking Control for Linear Stochastic Systems via Integral Reinforcement Learning","authors":"Kun Zhang;Yunjian Peng","doi":"10.1109/TASE.2025.3529895","DOIUrl":null,"url":null,"abstract":"This paper investigates optimal tracking control of linear stochastic systems with multiplicative state-dependent and input-dependent noise via a novel model-free integral reinforcement learning algorithm. We have conquered two major obstacles in this work. Firstly, the model-free tracking control on linear stochastic systems has scarcely been pivoted academically and it demands innovative methods and mathematical strategies to progress. An augmented stochastic differential equation of Itô’s type has been constructed while the control objective has been equated to minimization of the expected quadratic cost function. A model-dependent algorithm is shown to solve the stochastic algebraic Riccati equation. Second, control of linear stochastic systems essentially values for its sophisticated formation and sustained application. Inspired by optimal stationary control for linear stochastic systems, an integral reinforcement learning algorithm utilizing the adaptive dynamic programming method has been developed to scrap the reliance on the complete knowledge of system dynamics. The convergence of the core matrix and the sequence of control policies and the system stability has been rigorously studied following. Finally, numerical simulations are performed to demonstrate the efficacy of the proposed integral reinforcement learning methodology. Note to Practitioners—This paper focuses on the problem of optimal tracking control on linear stochastic systems, which is motivated by optimal stationary control and robust adaptive dynamic programming. Stochastic multiplicative noises are pervasive in modern control systems and engineering fields, such as power systems, aerospace systems, and industrial processes, thus important in modeling the random perturbation in system parameters and coefficients. However, research on tracking control on linear stochastic systems are challenging and languishing. First, we need to eradicate the dependence on complete knowledge system dynamics since machinery parameters in industrial design are constantly varying through time and space which builds a lasting impact on its structure, performance, and growth trajectory. Second, we innovatively apply integral reinforcement learning techniques on linear stochastic systems, which can be seamlessly integrated into networked control systems with noisy communication channels or neuronal networks to constructively solve stochastic multiplayer differential games online. Numerical simulations demonstrate that the proposed algorithm is feasible, which has also facilitated the compounding of robust adaptive dynamic programming and optimal tracking control on the industrial society. This study exhaustively presents the proposed approach to iteratively find optimal policies of the optimal tracking control problem directly from input and state data without explicitly identifying any system dynamics, and thoroughly unveils the explicit mathematical proof of convergence and system stability.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"10835-10844"},"PeriodicalIF":6.4000,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10843284/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
This paper investigates optimal tracking control of linear stochastic systems with multiplicative state-dependent and input-dependent noise via a novel model-free integral reinforcement learning algorithm. We have conquered two major obstacles in this work. Firstly, the model-free tracking control on linear stochastic systems has scarcely been pivoted academically and it demands innovative methods and mathematical strategies to progress. An augmented stochastic differential equation of Itô’s type has been constructed while the control objective has been equated to minimization of the expected quadratic cost function. A model-dependent algorithm is shown to solve the stochastic algebraic Riccati equation. Second, control of linear stochastic systems essentially values for its sophisticated formation and sustained application. Inspired by optimal stationary control for linear stochastic systems, an integral reinforcement learning algorithm utilizing the adaptive dynamic programming method has been developed to scrap the reliance on the complete knowledge of system dynamics. The convergence of the core matrix and the sequence of control policies and the system stability has been rigorously studied following. Finally, numerical simulations are performed to demonstrate the efficacy of the proposed integral reinforcement learning methodology. Note to Practitioners—This paper focuses on the problem of optimal tracking control on linear stochastic systems, which is motivated by optimal stationary control and robust adaptive dynamic programming. Stochastic multiplicative noises are pervasive in modern control systems and engineering fields, such as power systems, aerospace systems, and industrial processes, thus important in modeling the random perturbation in system parameters and coefficients. However, research on tracking control on linear stochastic systems are challenging and languishing. First, we need to eradicate the dependence on complete knowledge system dynamics since machinery parameters in industrial design are constantly varying through time and space which builds a lasting impact on its structure, performance, and growth trajectory. Second, we innovatively apply integral reinforcement learning techniques on linear stochastic systems, which can be seamlessly integrated into networked control systems with noisy communication channels or neuronal networks to constructively solve stochastic multiplayer differential games online. Numerical simulations demonstrate that the proposed algorithm is feasible, which has also facilitated the compounding of robust adaptive dynamic programming and optimal tracking control on the industrial society. This study exhaustively presents the proposed approach to iteratively find optimal policies of the optimal tracking control problem directly from input and state data without explicitly identifying any system dynamics, and thoroughly unveils the explicit mathematical proof of convergence and system stability.
期刊介绍:
The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.