DSQN: Robust path planning of mobile robot based on deep spiking Q-network

IF 6.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Neurocomputing Pub Date : 2025-06-14 Epub Date: 2025-03-11 DOI:10.1016/j.neucom.2025.129916
Aakash Kumar , Lei Zhang , Hazrat Bilal , Shifeng Wang , Ali Muhammad Shaikh , Lu Bo , Avinash Rohra , Alisha Khalid
{"title":"DSQN: Robust path planning of mobile robot based on deep spiking Q-network","authors":"Aakash Kumar ,&nbsp;Lei Zhang ,&nbsp;Hazrat Bilal ,&nbsp;Shifeng Wang ,&nbsp;Ali Muhammad Shaikh ,&nbsp;Lu Bo ,&nbsp;Avinash Rohra ,&nbsp;Alisha Khalid","doi":"10.1016/j.neucom.2025.129916","DOIUrl":null,"url":null,"abstract":"<div><div>With the rapid advancement of science and technology, the field of mobile robot applications continues to expand, with path planning emerging as a fundamental yet challenging task. While traditional path planning techniques have developed into a relatively complete theoretical system, their performance in uncertain environments remains a critical area of research. To address this, we propose a novel Deep Spiking Q-Network (DSQN) algorithm that significantly enhances path planning performance by leveraging the unique advantages of spiking neural networks (SNNs). Unlike classic Q-learning and its contemporary variants, the DSQN algorithm integrates global and local information simultaneously, resulting in superior overall performance. As the third generation of neural networks, SNNs offer unparalleled robustness and energy efficiency by mimicking biological neural systems. By introducing spiking neurons into the conventional Deep Q-learning (DQN) framework, the DSQN algorithm overcomes key challenges in deep reinforcement learning (DRL), such as limited robustness and high energy consumption. The DSQN training process incorporates both surrogate gradient learning (SGL) and ANN-to-SNN conversion techniques, with SGL demonstrating remarkable effectiveness in mobile robot path planning tasks. Experimental results validate the practicality and efficiency of DSQN, showcasing improved performance across diverse test scenarios compared to the original DQN algorithm. These findings highlight the potential of DSQN to advance path planning in complex and uncertain environments, establishing it as a robust and energy-efficient solution for mobile robotics.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"634 ","pages":"Article 129916"},"PeriodicalIF":6.5000,"publicationDate":"2025-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225005880","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/3/11 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

With the rapid advancement of science and technology, the field of mobile robot applications continues to expand, with path planning emerging as a fundamental yet challenging task. While traditional path planning techniques have developed into a relatively complete theoretical system, their performance in uncertain environments remains a critical area of research. To address this, we propose a novel Deep Spiking Q-Network (DSQN) algorithm that significantly enhances path planning performance by leveraging the unique advantages of spiking neural networks (SNNs). Unlike classic Q-learning and its contemporary variants, the DSQN algorithm integrates global and local information simultaneously, resulting in superior overall performance. As the third generation of neural networks, SNNs offer unparalleled robustness and energy efficiency by mimicking biological neural systems. By introducing spiking neurons into the conventional Deep Q-learning (DQN) framework, the DSQN algorithm overcomes key challenges in deep reinforcement learning (DRL), such as limited robustness and high energy consumption. The DSQN training process incorporates both surrogate gradient learning (SGL) and ANN-to-SNN conversion techniques, with SGL demonstrating remarkable effectiveness in mobile robot path planning tasks. Experimental results validate the practicality and efficiency of DSQN, showcasing improved performance across diverse test scenarios compared to the original DQN algorithm. These findings highlight the potential of DSQN to advance path planning in complex and uncertain environments, establishing it as a robust and energy-efficient solution for mobile robotics.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
DSQN:基于深度峰值q网络的移动机器人鲁棒路径规划
随着科学技术的飞速发展,移动机器人的应用领域不断扩大,路径规划成为一项基础性但又具有挑战性的任务。虽然传统的路径规划技术已经发展成为一个相对完整的理论体系,但其在不确定环境下的性能仍然是一个重要的研究领域。为了解决这个问题,我们提出了一种新的深度峰值q -网络(DSQN)算法,该算法利用峰值神经网络(snn)的独特优势,显著提高了路径规划性能。与经典的Q-learning及其当代变体不同,DSQN算法同时集成了全局和局部信息,从而获得了卓越的整体性能。作为第三代神经网络,snn通过模仿生物神经系统提供无与伦比的鲁棒性和能量效率。通过将尖峰神经元引入传统的深度q -学习(DQN)框架,DSQN算法克服了深度强化学习(DRL)中的关键挑战,如鲁棒性有限和高能量消耗。DSQN训练过程结合了代理梯度学习(SGL)和ann - snn转换技术,其中SGL在移动机器人路径规划任务中显示出显著的有效性。实验结果验证了DSQN的实用性和效率,与原始DQN算法相比,DSQN算法在不同测试场景下的性能有所提高。这些发现突出了DSQN在复杂和不确定环境中推进路径规划的潜力,使其成为移动机器人的强大且节能的解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
期刊最新文献
Multi-resolution aware pivot-guided dynamic graph neural network for multivariate time series forecasting PC-SNN: Predictive coding-based local Hebbian plasticity learning in spiking neural networks FedCapD: Federated class-incremental learning via capsule distillation and diffusion replay Adaptive critic learning based robust prescribed performance tracking control of nonlinear systems with unmatched uncertainties Towards unbiased action value estimation in reinforcement learning
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1