{"title":"A modified dueling DQN algorithm for robot path planning incorporating priority experience replay and artificial potential fields","authors":"Chang Li, Xiaofeng Yue, Zeyuan Liu, Guoyuan Ma, Hongbo Zhang, Yuan Zhou, Juan Zhu","doi":"10.1007/s10489-024-06149-8","DOIUrl":null,"url":null,"abstract":"<div><p>For the challenges of low learning efficiency, slow convergence speed and slow inference speed in robot path planning. This paper proposes an improved deep reinforcement learning algorithm for robot path planning. Firstly, the Dueling DQN network architecture is employed, combined with a priority experience replay strategy, to more effectively learn from and utilize experience data. Secondly, the mobility space of the robot is expanded, enhancing the diversity and flexibility of the action space. Additionally, in the action selection process, the Artificial Potential Field (APF) algorithm is introduced to intervene in the action selection with a certain probability, thereby accelerating the convergence process of the network. Simultaneously, the <span>\\(\\varepsilon\\)</span> -greedy strategy is employed to balance exploration and exploitation, facilitating better exploration of the environment and utilization of existing knowledge. Furthermore, this paper devises novel composite reward functions that comprehensively integrate multiple reward mechanisms to enhance the convergence performance of the algorithm and the quality of path planning. Finally, the effectiveness and superiority of the proposed method are validated through detailed comparative simulations. Compared to traditional DQN algorithms, Double DQN, and Double DQN with the APF strategy, the method proposed in this paper demonstrates higher learning efficiency and faster convergence speed, enabling more effective planning of shorter paths.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 6","pages":""},"PeriodicalIF":3.4000,"publicationDate":"2025-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Intelligence","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10489-024-06149-8","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
For the challenges of low learning efficiency, slow convergence speed and slow inference speed in robot path planning. This paper proposes an improved deep reinforcement learning algorithm for robot path planning. Firstly, the Dueling DQN network architecture is employed, combined with a priority experience replay strategy, to more effectively learn from and utilize experience data. Secondly, the mobility space of the robot is expanded, enhancing the diversity and flexibility of the action space. Additionally, in the action selection process, the Artificial Potential Field (APF) algorithm is introduced to intervene in the action selection with a certain probability, thereby accelerating the convergence process of the network. Simultaneously, the \(\varepsilon\) -greedy strategy is employed to balance exploration and exploitation, facilitating better exploration of the environment and utilization of existing knowledge. Furthermore, this paper devises novel composite reward functions that comprehensively integrate multiple reward mechanisms to enhance the convergence performance of the algorithm and the quality of path planning. Finally, the effectiveness and superiority of the proposed method are validated through detailed comparative simulations. Compared to traditional DQN algorithms, Double DQN, and Double DQN with the APF strategy, the method proposed in this paper demonstrates higher learning efficiency and faster convergence speed, enabling more effective planning of shorter paths.
期刊介绍:
With a focus on research in artificial intelligence and neural networks, this journal addresses issues involving solutions of real-life manufacturing, defense, management, government and industrial problems which are too complex to be solved through conventional approaches and require the simulation of intelligent thought processes, heuristics, applications of knowledge, and distributed and parallel processing. The integration of these multiple approaches in solving complex problems is of particular importance.
The journal presents new and original research and technological developments, addressing real and complex issues applicable to difficult problems. It provides a medium for exchanging scientific research and technological achievements accomplished by the international community.