基于反应防御的对抗多智能体强化学习模型检验

Dennis Gross, C. Schmidl, N. Jansen, G. Pérez
{"title":"基于反应防御的对抗多智能体强化学习模型检验","authors":"Dennis Gross, C. Schmidl, N. Jansen, G. Pérez","doi":"10.1609/icaps.v33i1.27191","DOIUrl":null,"url":null,"abstract":"Cooperative multi-agent reinforcement learning (CMARL) enables agents to achieve a common objective. \nHowever, the safety (a.k.a. robustness) of the CMARL agents operating in critical environments is not guaranteed. \nIn particular, agents are susceptible to adversarial noise in their observations that can mislead their decision-making.\nSo-called denoisers aim to remove adversarial noise from observations, yet, they are often error-prone.\nA key challenge for any rigorous safety verification technique in CMARL settings is the large number of states and transitions, which generally prohibits the construction of a (monolithic) model of the whole system.\nIn this paper, we present a verification method for CMARL agents in settings with or without adversarial attacks or denoisers.\nOur method relies on a tight integration of CMARL and a verification technique referred to as model checking.\nWe showcase the applicability of our method on various benchmarks from different domains.\nOur experiments show that our method is indeed suited to verify CMARL agents and that it scales better than a naive approach to model checking.","PeriodicalId":239898,"journal":{"name":"International Conference on Automated Planning and Scheduling","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Model Checking for Adversarial Multi-Agent Reinforcement Learning with Reactive Defense Methods\",\"authors\":\"Dennis Gross, C. Schmidl, N. Jansen, G. Pérez\",\"doi\":\"10.1609/icaps.v33i1.27191\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Cooperative multi-agent reinforcement learning (CMARL) enables agents to achieve a common objective. \\nHowever, the safety (a.k.a. robustness) of the CMARL agents operating in critical environments is not guaranteed. \\nIn particular, agents are susceptible to adversarial noise in their observations that can mislead their decision-making.\\nSo-called denoisers aim to remove adversarial noise from observations, yet, they are often error-prone.\\nA key challenge for any rigorous safety verification technique in CMARL settings is the large number of states and transitions, which generally prohibits the construction of a (monolithic) model of the whole system.\\nIn this paper, we present a verification method for CMARL agents in settings with or without adversarial attacks or denoisers.\\nOur method relies on a tight integration of CMARL and a verification technique referred to as model checking.\\nWe showcase the applicability of our method on various benchmarks from different domains.\\nOur experiments show that our method is indeed suited to verify CMARL agents and that it scales better than a naive approach to model checking.\",\"PeriodicalId\":239898,\"journal\":{\"name\":\"International Conference on Automated Planning and Scheduling\",\"volume\":\"21 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Conference on Automated Planning and Scheduling\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1609/icaps.v33i1.27191\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Automated Planning and Scheduling","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1609/icaps.v33i1.27191","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

协作式多智能体强化学习(CMARL)使智能体能够实现一个共同的目标。然而,在关键环境中运行的CMARL代理的安全性(即鲁棒性)并不能得到保证。特别是,代理人在他们的观察中容易受到对抗性噪音的影响,这可能会误导他们的决策。所谓的去噪器旨在从观测中去除对抗性噪声,然而,它们往往容易出错。在CMARL设置中,任何严格的安全验证技术的一个关键挑战是大量的状态和转换,这通常禁止构建整个系统的(单片)模型。在本文中,我们提出了一种CMARL代理在有或没有对抗性攻击或去噪设置下的验证方法。我们的方法依赖于CMARL和一种被称为模型检查的验证技术的紧密集成。我们展示了我们的方法在不同领域的各种基准测试中的适用性。我们的实验表明,我们的方法确实适合于验证CMARL代理,并且它比简单的模型检查方法具有更好的扩展性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Model Checking for Adversarial Multi-Agent Reinforcement Learning with Reactive Defense Methods
Cooperative multi-agent reinforcement learning (CMARL) enables agents to achieve a common objective. However, the safety (a.k.a. robustness) of the CMARL agents operating in critical environments is not guaranteed. In particular, agents are susceptible to adversarial noise in their observations that can mislead their decision-making. So-called denoisers aim to remove adversarial noise from observations, yet, they are often error-prone. A key challenge for any rigorous safety verification technique in CMARL settings is the large number of states and transitions, which generally prohibits the construction of a (monolithic) model of the whole system. In this paper, we present a verification method for CMARL agents in settings with or without adversarial attacks or denoisers. Our method relies on a tight integration of CMARL and a verification technique referred to as model checking. We showcase the applicability of our method on various benchmarks from different domains. Our experiments show that our method is indeed suited to verify CMARL agents and that it scales better than a naive approach to model checking.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Fast and Robust Resource-Constrained Scheduling with Graph Neural Networks Solving the Multi-Choice Two Dimensional Shelf Strip Packing Problem with Time Windows Generalizing Action Justification and Causal Links to Policies Exact Anytime Multi-Agent Path Finding Using Branch-and-Cut-and-Price and Large Neighborhood Search A Constraint Programming Solution to the Guillotine Rectangular Cutting Problem
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1