Dynamic Spike Bundling for Energy-Efficient Spiking Neural Networks

Sarada Krithivasan, Sanchari Sen, Swagath Venkataramani, A. Raghunathan
{"title":"Dynamic Spike Bundling for Energy-Efficient Spiking Neural Networks","authors":"Sarada Krithivasan, Sanchari Sen, Swagath Venkataramani, A. Raghunathan","doi":"10.1109/ISLPED.2019.8824897","DOIUrl":null,"url":null,"abstract":"Spiking Neural Networks (SNNs), which represent information as sequences of spikes, are gaining interest due to the emergence of low-power hardware platforms such as IBM TrueNorth and Intel Loihi, and their intrinsic ability to process temporal streams of data (e.g., outputs from event-based cameras). A spike produced by a neuron in an SNN is an event that triggers updates to the membrane potentials of each of the fanout neurons based on the weight associated with the synaptic connection, possibly resulting in other spikes being generated. The time and energy consumption in SNN implementations are dominated by accesses to the synaptic weights from memory and communication of spikes through the on-chip network. To improve the energy-efficiency of SNNs, we therefore propose Dynamic Spike Bundling (DSB), wherein an event to fanout neurons is not generated for every spike; instead, spikes produced by a neuron that occur close in time are dynamically bundled, with a single event being generated for the entire spike bundle. This reduces memory accesses as the synaptic weight can be fetched just once and reused across all spikes in the bundle. The communication traffic is also reduced as fewer messages are communicated between neurons.To evaluate DSB, we develop B-SNNAP, an event-driven SNN accelerator with hardware support for dynamically bundling spikes with minimal overheads. Across 7 image recognition benchmarks including CIFAR100 and ImageNet datasets, DSB achieves 1.15×-3.8× reduction in energy for <0.1% loss in accuracy, and upto 5.1× savings when <1% accuracy loss is tolerable.","PeriodicalId":101794,"journal":{"name":"2019 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED)","volume":"356 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"16","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE/ACM International Symposium on Low Power Electronics and Design (ISLPED)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISLPED.2019.8824897","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 16

Abstract

Spiking Neural Networks (SNNs), which represent information as sequences of spikes, are gaining interest due to the emergence of low-power hardware platforms such as IBM TrueNorth and Intel Loihi, and their intrinsic ability to process temporal streams of data (e.g., outputs from event-based cameras). A spike produced by a neuron in an SNN is an event that triggers updates to the membrane potentials of each of the fanout neurons based on the weight associated with the synaptic connection, possibly resulting in other spikes being generated. The time and energy consumption in SNN implementations are dominated by accesses to the synaptic weights from memory and communication of spikes through the on-chip network. To improve the energy-efficiency of SNNs, we therefore propose Dynamic Spike Bundling (DSB), wherein an event to fanout neurons is not generated for every spike; instead, spikes produced by a neuron that occur close in time are dynamically bundled, with a single event being generated for the entire spike bundle. This reduces memory accesses as the synaptic weight can be fetched just once and reused across all spikes in the bundle. The communication traffic is also reduced as fewer messages are communicated between neurons.To evaluate DSB, we develop B-SNNAP, an event-driven SNN accelerator with hardware support for dynamically bundling spikes with minimal overheads. Across 7 image recognition benchmarks including CIFAR100 and ImageNet datasets, DSB achieves 1.15×-3.8× reduction in energy for <0.1% loss in accuracy, and upto 5.1× savings when <1% accuracy loss is tolerable.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
高能效尖峰神经网络的动态尖峰捆绑
由于低功耗硬件平台(如IBM TrueNorth和Intel Loihi)的出现,以及它们处理时间数据流(例如,基于事件的相机的输出)的内在能力,以峰值序列的形式表示信息的峰值神经网络(snn)正受到越来越多的关注。SNN中神经元产生的尖峰是一个触发每个扇出神经元膜电位更新的事件,该事件基于与突触连接相关的权重,可能导致其他尖峰的产生。SNN实现的时间和能量消耗主要来自于从存储器访问突触权值和通过片上网络进行尖峰通信。因此,为了提高snn的能量效率,我们提出了动态Spike Bundling (DSB),其中不为每个Spike生成扇出神经元的事件;相反,由一个神经元产生的在时间上接近的尖峰被动态地捆绑在一起,整个尖峰束产生一个单一的事件。这减少了内存访问,因为突触权重可以只获取一次,并在bundle中的所有峰值中重用。由于神经元之间传递的信息更少,通信流量也减少了。为了评估DSB,我们开发了B-SNNAP,这是一个事件驱动的SNN加速器,硬件支持以最小的开销动态捆绑峰值。在包括CIFAR100和ImageNet数据集在内的7个图像识别基准测试中,DSB在精度损失<0.1%的情况下实现了1.15×-3.8×能耗降低,在可容忍的精度损失<1%的情况下节能高达5.1倍。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Non-Volatile Memory utilizing Reconfigurable Ferroelectric Transistors to enable Differential Read and Energy-Efficient In-Memory Computation Exploring the Relation between Monolithic 3D L1 GPU Cache Capacity and Warp Scheduling Efficiency Improving Energy Efficiency by Memoizing Data Access Information A Low-Energy Inductive Transceiver using Spike-Latency Encoding for Wireless 3D Integration Dynamic Spike Bundling for Energy-Efficient Spiking Neural Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1