Keyi Ni, Jing Chen, Jian Wang, Bo-Lan Liu, Ting Lei
{"title":"基于多头关注的多智能体强化学习","authors":"Keyi Ni, Jing Chen, Jian Wang, Bo-Lan Liu, Ting Lei","doi":"10.1109/ITNEC56291.2023.10082248","DOIUrl":null,"url":null,"abstract":"Multi-agent reinforcement learning(MARL) methods have become an important approach to solving the decision making problems of agents. As the environment’s complexity increases, the attention model can effectively solve the problem of information redundancy. However, the introduction of attention models in reinforcement learning may also lead to over-focusing and neglecting other potentially useful information. Moreover, the presence of attention would slow the convergence in the early stages of training. To address the above problem, we propose a divided attention reinforcement learning approach: (i) the involvement of an attention regularization term to make agents more divergent in their focus on different directions; (ii) the use of a layer normalization network structure and the use of a Pre-Layer Normalization(Pre-LN) network structure for the attention optimization in the initialization phase of training. It allows the agents to have a more stable and smooth gradient descent in the early stages of learning. Our approach has been tested in several multi-agent environment tasks. Compared to other related multi-agent methods, our method obtains higher final rewards and training efficiency.","PeriodicalId":218770,"journal":{"name":"2023 IEEE 6th Information Technology,Networking,Electronic and Automation Control Conference (ITNEC)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-agent Reinforcement Learning with Multi-head Attention\",\"authors\":\"Keyi Ni, Jing Chen, Jian Wang, Bo-Lan Liu, Ting Lei\",\"doi\":\"10.1109/ITNEC56291.2023.10082248\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multi-agent reinforcement learning(MARL) methods have become an important approach to solving the decision making problems of agents. As the environment’s complexity increases, the attention model can effectively solve the problem of information redundancy. However, the introduction of attention models in reinforcement learning may also lead to over-focusing and neglecting other potentially useful information. Moreover, the presence of attention would slow the convergence in the early stages of training. To address the above problem, we propose a divided attention reinforcement learning approach: (i) the involvement of an attention regularization term to make agents more divergent in their focus on different directions; (ii) the use of a layer normalization network structure and the use of a Pre-Layer Normalization(Pre-LN) network structure for the attention optimization in the initialization phase of training. It allows the agents to have a more stable and smooth gradient descent in the early stages of learning. Our approach has been tested in several multi-agent environment tasks. Compared to other related multi-agent methods, our method obtains higher final rewards and training efficiency.\",\"PeriodicalId\":218770,\"journal\":{\"name\":\"2023 IEEE 6th Information Technology,Networking,Electronic and Automation Control Conference (ITNEC)\",\"volume\":\"6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-02-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE 6th Information Technology,Networking,Electronic and Automation Control Conference (ITNEC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITNEC56291.2023.10082248\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 6th Information Technology,Networking,Electronic and Automation Control Conference (ITNEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITNEC56291.2023.10082248","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-agent Reinforcement Learning with Multi-head Attention
Multi-agent reinforcement learning(MARL) methods have become an important approach to solving the decision making problems of agents. As the environment’s complexity increases, the attention model can effectively solve the problem of information redundancy. However, the introduction of attention models in reinforcement learning may also lead to over-focusing and neglecting other potentially useful information. Moreover, the presence of attention would slow the convergence in the early stages of training. To address the above problem, we propose a divided attention reinforcement learning approach: (i) the involvement of an attention regularization term to make agents more divergent in their focus on different directions; (ii) the use of a layer normalization network structure and the use of a Pre-Layer Normalization(Pre-LN) network structure for the attention optimization in the initialization phase of training. It allows the agents to have a more stable and smooth gradient descent in the early stages of learning. Our approach has been tested in several multi-agent environment tasks. Compared to other related multi-agent methods, our method obtains higher final rewards and training efficiency.