{"title":"A DDQN-Based Path Planning Method for Multi-UAVs in a 3D Indoor Environment","authors":"Yuchen Ma, Yancai Xu","doi":"10.1109/ICCR55715.2022.10053884","DOIUrl":null,"url":null,"abstract":"Path planning is one of the most essential for unmanned aerial vehicle (UAV) autonomous navigation. The deep Q-network (DQN) method is widely used for solving the path planning problem, but most researchers simplify the scene into the 2D environment with a single UAV and ignore the fact that there are always multi-UAVs working in 3D environments. Therefore, a double deep Q-network (DDQN) based global path planning algorithm for multi-UAVs in a 3D indoor environment is proposed in this paper. Firstly, the double deep Q-network was designed to approximate the action of multi-UAVs. The 3D space is discretized into grids while each gird is a basic unit of path planning and the whole grid map is the input for the neural network. Then, a continual reward function generated by building an artificial potential field was determined to replace the traditional sparse reward function. Moreover, the action selection strategy is used to determine the current optimal action so that multi-UAVs are able to find the path to reach target points in a simulated indoor environment and avoid crashing into each other and obstacles at the same time. Finally, the experiment verifies the effectiveness of the proposed method. The simulation result demonstrates that the agents can effectively avoid local optimal solution and correctly predict the global optimal action.","PeriodicalId":441511,"journal":{"name":"2022 4th International Conference on Control and Robotics (ICCR)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 4th International Conference on Control and Robotics (ICCR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCR55715.2022.10053884","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Path planning is one of the most essential for unmanned aerial vehicle (UAV) autonomous navigation. The deep Q-network (DQN) method is widely used for solving the path planning problem, but most researchers simplify the scene into the 2D environment with a single UAV and ignore the fact that there are always multi-UAVs working in 3D environments. Therefore, a double deep Q-network (DDQN) based global path planning algorithm for multi-UAVs in a 3D indoor environment is proposed in this paper. Firstly, the double deep Q-network was designed to approximate the action of multi-UAVs. The 3D space is discretized into grids while each gird is a basic unit of path planning and the whole grid map is the input for the neural network. Then, a continual reward function generated by building an artificial potential field was determined to replace the traditional sparse reward function. Moreover, the action selection strategy is used to determine the current optimal action so that multi-UAVs are able to find the path to reach target points in a simulated indoor environment and avoid crashing into each other and obstacles at the same time. Finally, the experiment verifies the effectiveness of the proposed method. The simulation result demonstrates that the agents can effectively avoid local optimal solution and correctly predict the global optimal action.