Model-Free Tracking Control for Linear Stochastic Systems via Integral Reinforcement Learning

IF 6.4 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS IEEE Transactions on Automation Science and Engineering Pub Date : 2025-01-15 DOI:10.1109/TASE.2025.3529895
Kun Zhang;Yunjian Peng
{"title":"Model-Free Tracking Control for Linear Stochastic Systems via Integral Reinforcement Learning","authors":"Kun Zhang;Yunjian Peng","doi":"10.1109/TASE.2025.3529895","DOIUrl":null,"url":null,"abstract":"This paper investigates optimal tracking control of linear stochastic systems with multiplicative state-dependent and input-dependent noise via a novel model-free integral reinforcement learning algorithm. We have conquered two major obstacles in this work. Firstly, the model-free tracking control on linear stochastic systems has scarcely been pivoted academically and it demands innovative methods and mathematical strategies to progress. An augmented stochastic differential equation of Itô’s type has been constructed while the control objective has been equated to minimization of the expected quadratic cost function. A model-dependent algorithm is shown to solve the stochastic algebraic Riccati equation. Second, control of linear stochastic systems essentially values for its sophisticated formation and sustained application. Inspired by optimal stationary control for linear stochastic systems, an integral reinforcement learning algorithm utilizing the adaptive dynamic programming method has been developed to scrap the reliance on the complete knowledge of system dynamics. The convergence of the core matrix and the sequence of control policies and the system stability has been rigorously studied following. Finally, numerical simulations are performed to demonstrate the efficacy of the proposed integral reinforcement learning methodology. Note to Practitioners—This paper focuses on the problem of optimal tracking control on linear stochastic systems, which is motivated by optimal stationary control and robust adaptive dynamic programming. Stochastic multiplicative noises are pervasive in modern control systems and engineering fields, such as power systems, aerospace systems, and industrial processes, thus important in modeling the random perturbation in system parameters and coefficients. However, research on tracking control on linear stochastic systems are challenging and languishing. First, we need to eradicate the dependence on complete knowledge system dynamics since machinery parameters in industrial design are constantly varying through time and space which builds a lasting impact on its structure, performance, and growth trajectory. Second, we innovatively apply integral reinforcement learning techniques on linear stochastic systems, which can be seamlessly integrated into networked control systems with noisy communication channels or neuronal networks to constructively solve stochastic multiplayer differential games online. Numerical simulations demonstrate that the proposed algorithm is feasible, which has also facilitated the compounding of robust adaptive dynamic programming and optimal tracking control on the industrial society. This study exhaustively presents the proposed approach to iteratively find optimal policies of the optimal tracking control problem directly from input and state data without explicitly identifying any system dynamics, and thoroughly unveils the explicit mathematical proof of convergence and system stability.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"10835-10844"},"PeriodicalIF":6.4000,"publicationDate":"2025-01-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10843284/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

This paper investigates optimal tracking control of linear stochastic systems with multiplicative state-dependent and input-dependent noise via a novel model-free integral reinforcement learning algorithm. We have conquered two major obstacles in this work. Firstly, the model-free tracking control on linear stochastic systems has scarcely been pivoted academically and it demands innovative methods and mathematical strategies to progress. An augmented stochastic differential equation of Itô’s type has been constructed while the control objective has been equated to minimization of the expected quadratic cost function. A model-dependent algorithm is shown to solve the stochastic algebraic Riccati equation. Second, control of linear stochastic systems essentially values for its sophisticated formation and sustained application. Inspired by optimal stationary control for linear stochastic systems, an integral reinforcement learning algorithm utilizing the adaptive dynamic programming method has been developed to scrap the reliance on the complete knowledge of system dynamics. The convergence of the core matrix and the sequence of control policies and the system stability has been rigorously studied following. Finally, numerical simulations are performed to demonstrate the efficacy of the proposed integral reinforcement learning methodology. Note to Practitioners—This paper focuses on the problem of optimal tracking control on linear stochastic systems, which is motivated by optimal stationary control and robust adaptive dynamic programming. Stochastic multiplicative noises are pervasive in modern control systems and engineering fields, such as power systems, aerospace systems, and industrial processes, thus important in modeling the random perturbation in system parameters and coefficients. However, research on tracking control on linear stochastic systems are challenging and languishing. First, we need to eradicate the dependence on complete knowledge system dynamics since machinery parameters in industrial design are constantly varying through time and space which builds a lasting impact on its structure, performance, and growth trajectory. Second, we innovatively apply integral reinforcement learning techniques on linear stochastic systems, which can be seamlessly integrated into networked control systems with noisy communication channels or neuronal networks to constructively solve stochastic multiplayer differential games online. Numerical simulations demonstrate that the proposed algorithm is feasible, which has also facilitated the compounding of robust adaptive dynamic programming and optimal tracking control on the industrial society. This study exhaustively presents the proposed approach to iteratively find optimal policies of the optimal tracking control problem directly from input and state data without explicitly identifying any system dynamics, and thoroughly unveils the explicit mathematical proof of convergence and system stability.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于积分强化学习的线性随机系统无模型跟踪控制
利用一种新的无模型积分强化学习算法研究了具有状态依赖和输入依赖多重噪声的线性随机系统的最优跟踪控制。我们已经克服了这项工作中的两个主要障碍。首先,线性随机系统的无模型跟踪控制在学术界的研究很少,需要创新的方法和数学策略才能取得进展。构造了一个Itô型的增广随机微分方程,将控制目标等效为期望二次代价函数的最小化。给出了求解随机代数Riccati方程的模型依赖算法。其次,线性随机系统的控制,其本质价值在于其复杂的形成和持续的应用。受线性随机系统最优平稳控制的启发,利用自适应动态规划方法开发了一种积分强化学习算法,以摆脱对系统动力学完整知识的依赖。其次,对核心矩阵的收敛性、控制策略序列的收敛性以及系统的稳定性进行了严格的研究。最后,进行了数值模拟,以证明所提出的积分强化学习方法的有效性。本文主要研究线性随机系统的最优跟踪控制问题,该问题由最优平稳控制和鲁棒自适应动态规划驱动。随机乘性噪声普遍存在于现代控制系统和工程领域,如电力系统、航空航天系统和工业过程,因此对系统参数和系数的随机扰动建模非常重要。然而,线性随机系统的跟踪控制研究是一个充满挑战和困难的课题。首先,我们需要消除对完整知识系统动力学的依赖,因为工业设计中的机械参数是随时间和空间不断变化的,这对其结构、性能和生长轨迹产生了持久的影响。其次,我们创新地将积分强化学习技术应用于线性随机系统,该系统可以无缝集成到具有噪声通信信道或神经网络的网络控制系统中,以在线建设性地解决随机多人微分游戏。数值仿真结果表明,该算法是可行的,为鲁棒自适应动态规划与最优跟踪控制在工业社会中的结合提供了方便。本研究详尽地提出了直接从输入和状态数据中迭代找到最优跟踪控制问题的最优策略的方法,而无需明确识别任何系统动力学,并彻底揭示了收敛性和系统稳定性的显式数学证明。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Transactions on Automation Science and Engineering
IEEE Transactions on Automation Science and Engineering 工程技术-自动化与控制系统
CiteScore
12.50
自引率
14.30%
发文量
404
审稿时长
3.0 months
期刊介绍: The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.
期刊最新文献
RflySimSaT: A Safety Assessment Platform for UAVs Based on Hardware-in-the-loop Simulation Dynamic Memory Event-triggered Predefined-time Control for Stochastic Nonlinear Time-delay Systems with Unknown Input Hysteresis Machine Learning Achieves Accurate and Smooth Navigation Control for Microrobots Performance Improvement of a High-Speed On/Off Valve-Piloted Proportional Valve via Nonlinear Modeling and Load-Adaptive Sliding Mode Control Sea-Trial Validated Adaptive Control for Unmanned Surface Vehicles: Integrating Robust Line-of-Sight With Dual-Mode Model Predictive Control
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1