Multi-attribute dynamic attenuation learning improved spiking actor network

IF 6.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Neurocomputing Pub Date : 2025-01-21 Epub Date: 2024-11-08 DOI:10.1016/j.neucom.2024.128819
Rong Xiao, Zhiyuan Hu, Jie Zhang, Chenwei Tang, Jiancheng Lv
{"title":"Multi-attribute dynamic attenuation learning improved spiking actor network","authors":"Rong Xiao,&nbsp;Zhiyuan Hu,&nbsp;Jie Zhang,&nbsp;Chenwei Tang,&nbsp;Jiancheng Lv","doi":"10.1016/j.neucom.2024.128819","DOIUrl":null,"url":null,"abstract":"<div><div>Deep reinforcement learning (DRL) has shown promising results in solving robotic control and decision tasks, which can learn the high-dimensional state and action information well. Despite their successes, conventional neural-based DRL models are criticized for low energy efficiency, making them laborious to be widely applied in low-power electronics. With more biologically plausible plasticity principles, spiking neural networks (SNNs) are now considered an energy-efficient and robust alternative. The most existing dynamics and learning paradigms for spiking neurons with a common Leaky Integrate-and-Fire (LIF) neuron model often result in relatively low efficiency and poor robustness. To address these limitations, we propose a multi-attribute dynamic attenuation learning improved spiking actor network (MADA-SAN) for reinforcement learning to achieve effective decision-making. The resistance, membrane voltage and membrane current of spiking neurons are updated from a fixed value into dynamic attenuation. By enhancing the temporal relation dependencies in neurons, this model can learn the spatio-temporal relevance of complex continuous information well. Extensive experimental results show MADA-SAN performs better than its counterpart deep actor network on six continuous control tasks from OpenAI gym. Besides, we further validated the proposed MADA-LIF can achieve comparable performance with other state-of-the-art algorithms on MNIST and DVS-gesture recognition tasks.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128819"},"PeriodicalIF":6.5000,"publicationDate":"2025-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S092523122401590X","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/11/8 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Deep reinforcement learning (DRL) has shown promising results in solving robotic control and decision tasks, which can learn the high-dimensional state and action information well. Despite their successes, conventional neural-based DRL models are criticized for low energy efficiency, making them laborious to be widely applied in low-power electronics. With more biologically plausible plasticity principles, spiking neural networks (SNNs) are now considered an energy-efficient and robust alternative. The most existing dynamics and learning paradigms for spiking neurons with a common Leaky Integrate-and-Fire (LIF) neuron model often result in relatively low efficiency and poor robustness. To address these limitations, we propose a multi-attribute dynamic attenuation learning improved spiking actor network (MADA-SAN) for reinforcement learning to achieve effective decision-making. The resistance, membrane voltage and membrane current of spiking neurons are updated from a fixed value into dynamic attenuation. By enhancing the temporal relation dependencies in neurons, this model can learn the spatio-temporal relevance of complex continuous information well. Extensive experimental results show MADA-SAN performs better than its counterpart deep actor network on six continuous control tasks from OpenAI gym. Besides, we further validated the proposed MADA-LIF can achieve comparable performance with other state-of-the-art algorithms on MNIST and DVS-gesture recognition tasks.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
多属性动态衰减学习改进型尖峰行动者网络
深度强化学习(DRL)能够很好地学习高维状态和动作信息,在解决机器人控制和决策任务方面取得了可喜的成果。尽管取得了成功,但传统的基于神经的 DRL 模型因能效低而饱受诟病,使其难以在低功耗电子设备中广泛应用。尖峰神经网络(SNN)的可塑性原理更符合生物学原理,因此现在被认为是一种高能效、稳健的替代方案。现有的大多数尖峰神经元动力学和学习范式都采用常见的 "漏电积分-放电(LIF)"神经元模型,这往往导致效率相对较低,鲁棒性较差。针对这些局限性,我们提出了一种用于强化学习的多属性动态衰减学习改进型尖峰行为网络(MADA-SAN),以实现有效决策。尖峰神经元的电阻、膜电压和膜电流从固定值更新为动态衰减。通过增强神经元的时间关系依赖性,该模型可以很好地学习复杂连续信息的时空相关性。广泛的实验结果表明,MADA-SAN 在 OpenAI gym 的六个连续控制任务上的表现优于其对应的深度演员网络。此外,我们还进一步验证了所提出的 MADA-LIF 可以在 MNIST 和 DVS 手势识别任务上实现与其他最先进算法相当的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
期刊最新文献
Unseen TCR-epitope interaction prediction with self-supervised contrastive learning Incomplete multi-view clustering based on kernel graph fusion tensor under self-expression framework TriSSR: Tri-expert fusion in state space models for sequential recommendation Optimized Bayesian asymmetric nutcracker quantized neural networks for area-efficient 10T2C capacitive SRAM design with reconfigurable SAR ADCs TOC-UCO: a comprehensive repository of tabular ordinal classification datasets
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1