Lixin Yang;Weijun Lv;Yong Xu;Jie Tao;Daniel E. Quevedo
{"title":"Structure-Aware Reinforcement Learning for Optimal Transmission Scheduling Over Packet Length-Dependent Lossy Networks","authors":"Lixin Yang;Weijun Lv;Yong Xu;Jie Tao;Daniel E. Quevedo","doi":"10.1109/TAC.2024.3488823","DOIUrl":null,"url":null,"abstract":"The optimal transmission scheduling over packet length-dependent lossy networks is investigated. A set of sensors observe some linear time-invariant systems, and obtain their local state estimates using Kalman filters independently. The local state estimates are encoded in a data packet, which is sent to a remote estimator over packet length-dependent lossy networks, i.e., the packet arrival probability is exponentially decreasing with the packet length. The tradeoff arises between the packet arrival probability and the data size. To improve the remote estimation performance, one needs to design a transmission scheduling policy to determine how many as well as which sensors transmit data at each time instant, which is formulated here as a Markov decision process (MDP) model. There exists an optimal stationary policy for the MDP model, which is verified to possess a threshold structure. Based on this, a structure-aware reinforcement learning algorithm is proposed to approximate the MDP's optimal policy. Some simulation examples are given to verify the optimal policy's structure, and illustrate the performance of the proposed structure-aware reinforcement learning algorithm.","PeriodicalId":13201,"journal":{"name":"IEEE Transactions on Automatic Control","volume":"70 4","pages":"2576-2583"},"PeriodicalIF":7.0000,"publicationDate":"2024-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automatic Control","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10740304/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The optimal transmission scheduling over packet length-dependent lossy networks is investigated. A set of sensors observe some linear time-invariant systems, and obtain their local state estimates using Kalman filters independently. The local state estimates are encoded in a data packet, which is sent to a remote estimator over packet length-dependent lossy networks, i.e., the packet arrival probability is exponentially decreasing with the packet length. The tradeoff arises between the packet arrival probability and the data size. To improve the remote estimation performance, one needs to design a transmission scheduling policy to determine how many as well as which sensors transmit data at each time instant, which is formulated here as a Markov decision process (MDP) model. There exists an optimal stationary policy for the MDP model, which is verified to possess a threshold structure. Based on this, a structure-aware reinforcement learning algorithm is proposed to approximate the MDP's optimal policy. Some simulation examples are given to verify the optimal policy's structure, and illustrate the performance of the proposed structure-aware reinforcement learning algorithm.
期刊介绍:
In the IEEE Transactions on Automatic Control, the IEEE Control Systems Society publishes high-quality papers on the theory, design, and applications of control engineering. Two types of contributions are regularly considered:
1) Papers: Presentation of significant research, development, or application of control concepts.
2) Technical Notes and Correspondence: Brief technical notes, comments on published areas or established control topics, corrections to papers and notes published in the Transactions.
In addition, special papers (tutorials, surveys, and perspectives on the theory and applications of control systems topics) are solicited.