{"title":"Cooperative co-learning: a model-based approach for solving multi-agent reinforcement problems","authors":"B. Scherrer, F. Charpillet","doi":"10.1109/TAI.2002.1180839","DOIUrl":null,"url":null,"abstract":"Solving multiagent reinforcement learning problems is a key issue. Indeed, the complexity of deriving multiagent plans, especially when one uses an explicit model of the problem, is dramatically increasing with the number of agents. This papers introduces a general iterative heuristic: at each step one chooses a sub-group of agents and update their policies to optimize the task given the rest of agents have fixed plans. We analyse this process in a general purpose and show how it can be applied to Markov decision processes, partially observable Markov decision processes and decentralized partially observable Markov decision processes.","PeriodicalId":197064,"journal":{"name":"14th IEEE International Conference on Tools with Artificial Intelligence, 2002. (ICTAI 2002). Proceedings.","volume":"189 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2002-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"13","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"14th IEEE International Conference on Tools with Artificial Intelligence, 2002. (ICTAI 2002). Proceedings.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TAI.2002.1180839","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 13
Abstract
Solving multiagent reinforcement learning problems is a key issue. Indeed, the complexity of deriving multiagent plans, especially when one uses an explicit model of the problem, is dramatically increasing with the number of agents. This papers introduces a general iterative heuristic: at each step one chooses a sub-group of agents and update their policies to optimize the task given the rest of agents have fixed plans. We analyse this process in a general purpose and show how it can be applied to Markov decision processes, partially observable Markov decision processes and decentralized partially observable Markov decision processes.