Toward Multi-Agent Reinforcement Learning for Distributed Event-Triggered Control

Lukas Kesper, Sebastian Trimpe, Dominik Baumann
{"title":"Toward Multi-Agent Reinforcement Learning for Distributed Event-Triggered Control","authors":"Lukas Kesper, Sebastian Trimpe, Dominik Baumann","doi":"10.48550/arXiv.2305.08723","DOIUrl":null,"url":null,"abstract":"Event-triggered communication and control provide high control performance in networked control systems without overloading the communication network. However, most approaches require precise mathematical models of the system dynamics, which may not always be available. Model-free learning of communication and control policies provides an alternative. Nevertheless, existing methods typically consider single-agent settings. This paper proposes a model-free reinforcement learning algorithm that jointly learns resource-aware communication and control policies for distributed multi-agent systems from data. We evaluate the algorithm in a high-dimensional and nonlinear simulation example and discuss promising avenues for further research.","PeriodicalId":268449,"journal":{"name":"Conference on Learning for Dynamics & Control","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Conference on Learning for Dynamics & Control","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.48550/arXiv.2305.08723","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Event-triggered communication and control provide high control performance in networked control systems without overloading the communication network. However, most approaches require precise mathematical models of the system dynamics, which may not always be available. Model-free learning of communication and control policies provides an alternative. Nevertheless, existing methods typically consider single-agent settings. This paper proposes a model-free reinforcement learning algorithm that jointly learns resource-aware communication and control policies for distributed multi-agent systems from data. We evaluate the algorithm in a high-dimensional and nonlinear simulation example and discuss promising avenues for further research.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
分布式事件触发控制的多智能体强化学习研究
事件触发通信和控制在网络控制系统中提供了高控制性能,而不会使通信网络过载。然而,大多数方法需要精确的系统动力学数学模型,这可能并不总是可用的。通信和控制策略的无模型学习提供了另一种选择。然而,现有的方法通常考虑单代理设置。本文提出了一种无模型强化学习算法,该算法从数据中共同学习分布式多智能体系统的资源感知通信和控制策略。我们在一个高维和非线性的仿真例子中评估了该算法,并讨论了进一步研究的有前途的途径。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Agile Catching with Whole-Body MPC and Blackbox Policy Learning Time Dependent Inverse Optimal Control using Trigonometric Basis Functions Provably Efficient Generalized Lagrangian Policy Optimization for Safe Multi-Agent Reinforcement Learning Black-Box vs. Gray-Box: A Case Study on Learning Table Tennis Ball Trajectory Prediction with Spin and Impacts Model-based Validation as Probabilistic Inference
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1