{"title":"Cumulative Training and Transfer Learning for Multi-Robots Collision-Free Navigation Problems","authors":"Trung-Thanh Nguyen, Amartya Hatua, A. Sung","doi":"10.1109/UEMCON47517.2019.8992945","DOIUrl":null,"url":null,"abstract":"Recently, the characteristics of robot autonomy, decentralized control, collective decision-making ability, high fault tolerance, etc. have significantly increased the applications of swarm robotics in targeted material delivery, precision farming, surveillance, defense and many other areas. In these multi-agent systems, safe collision avoidance is one of the most fundamental and important problems. Difference approaches, especially reinforcement learning, have been applied to solve this problem. This paper introduces a new cumulative learning approach which comprises of application of transfer learning with distributed multi-agent reinforcement learning techniques to solve collision-free navigation for swarm robotics. In our method, throughout the learning processes from the least complexity scenario to the most complex one, multiple agents can improve the shared policy through parameter sharing, reward shaping and multi-round multi-steps learning. We have adapted two policy gradient algorithms (TRPO and PPO) as the core of our distributed multiagent reinforcement learning method. The performance has shown that our new methodology can help reduce the training time and generate a robust navigation plan that can easily be generalized to complex in-door scenarios.","PeriodicalId":187022,"journal":{"name":"2019 IEEE 10th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE 10th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/UEMCON47517.2019.8992945","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Recently, the characteristics of robot autonomy, decentralized control, collective decision-making ability, high fault tolerance, etc. have significantly increased the applications of swarm robotics in targeted material delivery, precision farming, surveillance, defense and many other areas. In these multi-agent systems, safe collision avoidance is one of the most fundamental and important problems. Difference approaches, especially reinforcement learning, have been applied to solve this problem. This paper introduces a new cumulative learning approach which comprises of application of transfer learning with distributed multi-agent reinforcement learning techniques to solve collision-free navigation for swarm robotics. In our method, throughout the learning processes from the least complexity scenario to the most complex one, multiple agents can improve the shared policy through parameter sharing, reward shaping and multi-round multi-steps learning. We have adapted two policy gradient algorithms (TRPO and PPO) as the core of our distributed multiagent reinforcement learning method. The performance has shown that our new methodology can help reduce the training time and generate a robust navigation plan that can easily be generalized to complex in-door scenarios.