{"title":"Reconsidering the energy efficiency of spiking neural networks","authors":"Zhanglu Yan, Zhenyu Bai, Weng-Fai Wong","doi":"arxiv-2409.08290","DOIUrl":null,"url":null,"abstract":"Spiking neural networks (SNNs) are generally regarded as more\nenergy-efficient because they do not use multiplications. However, most SNN\nworks only consider the counting of additions to evaluate energy consumption,\nneglecting other overheads such as memory accesses and data movement\noperations. This oversight can lead to a misleading perception of efficiency,\nespecially when state-of-the-art SNN accelerators operate with very small time\nwindow sizes. In this paper, we present a detailed comparison of the energy\nconsumption of artificial neural networks (ANNs) and SNNs from a hardware\nperspective. We provide accurate formulas for energy consumption based on\nclassical multi-level memory hierarchy architectures, commonly used\nneuromorphic dataflow architectures, and our proposed improved spatial-dataflow\narchitecture. Our research demonstrates that to achieve comparable accuracy and\ngreater energy efficiency than ANNs, SNNs require strict limitations on both\ntime window size T and sparsity s. For instance, with the VGG16 model and a\nfixed T of 6, the neuron sparsity rate must exceed 93% to ensure energy\nefficiency across most architectures. Inspired by our findings, we explore\nstrategies to enhance energy efficiency by increasing sparsity. We introduce\ntwo regularization terms during training that constrain weights and\nactivations, effectively boosting the sparsity rate. Our experiments on the\nCIFAR-10 dataset, using T of 6, show that our SNNs consume 69% of the energy\nused by optimized ANNs on spatial-dataflow architectures, while maintaining an\nSNN accuracy of 94.18%. This framework, developed using PyTorch, is publicly\navailable for use and further research.","PeriodicalId":501347,"journal":{"name":"arXiv - CS - Neural and Evolutionary Computing","volume":"15 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Neural and Evolutionary Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.08290","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Spiking neural networks (SNNs) are generally regarded as more
energy-efficient because they do not use multiplications. However, most SNN
works only consider the counting of additions to evaluate energy consumption,
neglecting other overheads such as memory accesses and data movement
operations. This oversight can lead to a misleading perception of efficiency,
especially when state-of-the-art SNN accelerators operate with very small time
window sizes. In this paper, we present a detailed comparison of the energy
consumption of artificial neural networks (ANNs) and SNNs from a hardware
perspective. We provide accurate formulas for energy consumption based on
classical multi-level memory hierarchy architectures, commonly used
neuromorphic dataflow architectures, and our proposed improved spatial-dataflow
architecture. Our research demonstrates that to achieve comparable accuracy and
greater energy efficiency than ANNs, SNNs require strict limitations on both
time window size T and sparsity s. For instance, with the VGG16 model and a
fixed T of 6, the neuron sparsity rate must exceed 93% to ensure energy
efficiency across most architectures. Inspired by our findings, we explore
strategies to enhance energy efficiency by increasing sparsity. We introduce
two regularization terms during training that constrain weights and
activations, effectively boosting the sparsity rate. Our experiments on the
CIFAR-10 dataset, using T of 6, show that our SNNs consume 69% of the energy
used by optimized ANNs on spatial-dataflow architectures, while maintaining an
SNN accuracy of 94.18%. This framework, developed using PyTorch, is publicly
available for use and further research.