Structure-Aware Reinforcement Learning for Optimal Transmission Scheduling Over Packet Length-Dependent Lossy Networks

IF 7 1区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS IEEE Transactions on Automatic Control Pub Date : 2024-10-31 DOI:10.1109/TAC.2024.3488823
Lixin Yang;Weijun Lv;Yong Xu;Jie Tao;Daniel E. Quevedo
{"title":"Structure-Aware Reinforcement Learning for Optimal Transmission Scheduling Over Packet Length-Dependent Lossy Networks","authors":"Lixin Yang;Weijun Lv;Yong Xu;Jie Tao;Daniel E. Quevedo","doi":"10.1109/TAC.2024.3488823","DOIUrl":null,"url":null,"abstract":"The optimal transmission scheduling over packet length-dependent lossy networks is investigated. A set of sensors observe some linear time-invariant systems, and obtain their local state estimates using Kalman filters independently. The local state estimates are encoded in a data packet, which is sent to a remote estimator over packet length-dependent lossy networks, i.e., the packet arrival probability is exponentially decreasing with the packet length. The tradeoff arises between the packet arrival probability and the data size. To improve the remote estimation performance, one needs to design a transmission scheduling policy to determine how many as well as which sensors transmit data at each time instant, which is formulated here as a Markov decision process (MDP) model. There exists an optimal stationary policy for the MDP model, which is verified to possess a threshold structure. Based on this, a structure-aware reinforcement learning algorithm is proposed to approximate the MDP's optimal policy. Some simulation examples are given to verify the optimal policy's structure, and illustrate the performance of the proposed structure-aware reinforcement learning algorithm.","PeriodicalId":13201,"journal":{"name":"IEEE Transactions on Automatic Control","volume":"70 4","pages":"2576-2583"},"PeriodicalIF":7.0000,"publicationDate":"2024-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automatic Control","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10740304/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

The optimal transmission scheduling over packet length-dependent lossy networks is investigated. A set of sensors observe some linear time-invariant systems, and obtain their local state estimates using Kalman filters independently. The local state estimates are encoded in a data packet, which is sent to a remote estimator over packet length-dependent lossy networks, i.e., the packet arrival probability is exponentially decreasing with the packet length. The tradeoff arises between the packet arrival probability and the data size. To improve the remote estimation performance, one needs to design a transmission scheduling policy to determine how many as well as which sensors transmit data at each time instant, which is formulated here as a Markov decision process (MDP) model. There exists an optimal stationary policy for the MDP model, which is verified to possess a threshold structure. Based on this, a structure-aware reinforcement learning algorithm is proposed to approximate the MDP's optimal policy. Some simulation examples are given to verify the optimal policy's structure, and illustrate the performance of the proposed structure-aware reinforcement learning algorithm.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
在依赖数据包长度的有损网络上优化传输调度的结构感知强化学习
研究了包长度相关的有损网络的最优传输调度问题。利用一组传感器对某线性时不变系统进行观测,并利用卡尔曼滤波独立获得其局部状态估计。本地状态估计被编码在数据包中,该数据包通过数据包长度相关的有损网络发送给远程估计器,即数据包到达概率随数据包长度呈指数递减。在数据包到达概率和数据大小之间进行权衡。为了提高远程估计性能,需要设计一个传输调度策略,以确定每个时刻有多少传感器以及哪些传感器传输数据,这里将其表述为马尔可夫决策过程(MDP)模型。MDP模型存在一个最优平稳策略,该策略具有阈值结构。在此基础上,提出了一种结构感知强化学习算法来逼近MDP的最优策略。通过仿真实例验证了最优策略的结构,并说明了结构感知强化学习算法的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Transactions on Automatic Control
IEEE Transactions on Automatic Control 工程技术-工程:电子与电气
CiteScore
11.30
自引率
5.90%
发文量
824
审稿时长
9 months
期刊介绍: In the IEEE Transactions on Automatic Control, the IEEE Control Systems Society publishes high-quality papers on the theory, design, and applications of control engineering. Two types of contributions are regularly considered: 1) Papers: Presentation of significant research, development, or application of control concepts. 2) Technical Notes and Correspondence: Brief technical notes, comments on published areas or established control topics, corrections to papers and notes published in the Transactions. In addition, special papers (tutorials, surveys, and perspectives on the theory and applications of control systems topics) are solicited.
期刊最新文献
Reaching Resilient Leader-Follower Consensus in Time-Varying Networks via Multi-Hop Relays Dynamical System Approach for Optimal Control Problems with Equilibrium Constraints Using Gap-Constraint-Based Reformulation Set-Based State Estimation for Discrete-Time Semi-Markov Jump Linear Systems Using Zonotopes Safe Event-triggered Gaussian Process Learning for Barrier-Constrained Control Energy-Gain Control of Time-Varying Systems: Receding Horizon Approximation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1