首页 > 最新文献

IEEE Journal on Emerging and Selected Topics in Circuits and Systems最新文献

英文 中文
Dynamic-HDC: A Two-Stage Dynamic Inference Framework for Brain-Inspired Hyperdimensional Computing Dynamic-HDC:脑启发超维计算的两阶段动态推理框架
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328857
Yu-Chuan Chuang;Cheng-Yang Chang;An-Yeu Wu
Brain-inspired hyperdimensional computing (HDC) has attracted attention due to its energy efficiency and noise resilience in various IoT applications. However, striking the right balance between accuracy and efficiency in HDC remains a challenge. Specifically, HDC represents data as high-dimensional vectors known as hypervectors (HVs), where each component of HVs can be a high-precision integer or a low-cost bipolar number (+1/−1). However, this choice presents HDC with a significant trade-off between accuracy and efficiency. To address this challenge, we propose a two-stage dynamic inference framework called Dynamic-HDC that offers IoT applications a more flexible solution rather than limiting them to choose between the two extreme options. Dynamic-HDC leverages the strategies of early exit and model parameter adaptation. Unlike prior works that use a single HDC model to classify all data, Dynamic-HDC employs a cascade of models for two-stage inference. The first stage involves a low-cost, low-precision bipolar model, while the second stage utilizes a high-cost, high-precision integer model. By doing so, Dynamic-HDC can save computational resources for easy samples by performing an early exit when the low-cost bipolar model exhibits high confidence in its classification. For difficult samples, the high-precision integer model is conditionally activated to achieve more accurate predictions. To further enhance the efficiency of Dynamic-HDC, we introduce dynamic dimension selection (DDS) and dynamic class selection (DCS). These techniques enable the framework to dynamically adapt the dimensions and the number of classes in the HDC model, further optimizing performance. We evaluate the effectiveness of Dynamic-HDC on three commonly used benchmarks in HDC research, namely MNIST, ISOLET, and UCIHAR. Our simulation results demonstrate that Dynamic-HDC with different configurations can reduce energy consumption by 19.8-51.1% and execution time by 22.5-49.9% with negligible 0.02-0.36 % accuracy degradation compared to a single integer model. Compared to a single bipolar model, Dynamic-HDC improves 3.1% accuracy with a slight 10% energy and 14% execution time overhead.
大脑启发的超维计算(HDC)因其在各种物联网应用中的能效和抗噪能力而备受关注。然而,如何在 HDC 的准确性和效率之间取得适当的平衡仍然是一个挑战。具体来说,HDC 将数据表示为称为超向量(HVs)的高维向量,其中 HVs 的每个分量可以是高精度整数或低成本双极性数字(+1/-1)。然而,这种选择使得 HDC 需要在精度和效率之间做出重大权衡。为了应对这一挑战,我们提出了一个名为 "动态-HDC "的两阶段动态推理框架,为物联网应用提供更灵活的解决方案,而不是局限于在两个极端选项中做出选择。动态-HDC 利用了早期退出和模型参数适应策略。与之前使用单一 HDC 模型对所有数据进行分类的工作不同,Dynamic-HDC 采用级联模型进行两阶段推理。第一阶段使用低成本、低精度的双极模型,第二阶段使用高成本、高精度的整数模型。这样,当低成本双极模型在分类中表现出较高的置信度时,Dynamic-HDC 就会提前退出,从而为简单样本节省计算资源。对于困难样本,则有条件地激活高精度整数模型,以实现更准确的预测。为了进一步提高 Dynamic-HDC 的效率,我们引入了动态维度选择 (DDS) 和动态类别选择 (DCS)。这些技术使框架能够动态调整 HDC 模型中的维数和类数,从而进一步优化性能。我们在 HDC 研究中常用的三个基准(即 MNIST、ISOLET 和 UCIHAR)上评估了动态 HDC 的有效性。仿真结果表明,与单一整数模型相比,不同配置的 Dynamic-HDC 可减少 19.8-51.1% 的能耗和 22.5-49.9% 的执行时间,而 0.02-0.36% 的精度下降可以忽略不计。与单一双极性模型相比,Dynamic-HDC 提高了 3.1% 的精度,但能耗和执行时间的开销分别为 10%和 14%。
{"title":"Dynamic-HDC: A Two-Stage Dynamic Inference Framework for Brain-Inspired Hyperdimensional Computing","authors":"Yu-Chuan Chuang;Cheng-Yang Chang;An-Yeu Wu","doi":"10.1109/JETCAS.2023.3328857","DOIUrl":"10.1109/JETCAS.2023.3328857","url":null,"abstract":"Brain-inspired hyperdimensional computing (HDC) has attracted attention due to its energy efficiency and noise resilience in various IoT applications. However, striking the right balance between accuracy and efficiency in HDC remains a challenge. Specifically, HDC represents data as high-dimensional vectors known as hypervectors (HVs), where each component of HVs can be a high-precision integer or a low-cost bipolar number (+1/−1). However, this choice presents HDC with a significant trade-off between accuracy and efficiency. To address this challenge, we propose a two-stage dynamic inference framework called Dynamic-HDC that offers IoT applications a more flexible solution rather than limiting them to choose between the two extreme options. Dynamic-HDC leverages the strategies of early exit and model parameter adaptation. Unlike prior works that use a single HDC model to classify all data, Dynamic-HDC employs a cascade of models for two-stage inference. The first stage involves a low-cost, low-precision bipolar model, while the second stage utilizes a high-cost, high-precision integer model. By doing so, Dynamic-HDC can save computational resources for easy samples by performing an early exit when the low-cost bipolar model exhibits high confidence in its classification. For difficult samples, the high-precision integer model is conditionally activated to achieve more accurate predictions. To further enhance the efficiency of Dynamic-HDC, we introduce dynamic dimension selection (DDS) and dynamic class selection (DCS). These techniques enable the framework to dynamically adapt the dimensions and the number of classes in the HDC model, further optimizing performance. We evaluate the effectiveness of Dynamic-HDC on three commonly used benchmarks in HDC research, namely MNIST, ISOLET, and UCIHAR. Our simulation results demonstrate that Dynamic-HDC with different configurations can reduce energy consumption by 19.8-51.1% and execution time by 22.5-49.9% with negligible 0.02-0.36 % accuracy degradation compared to a single integer model. Compared to a single bipolar model, Dynamic-HDC improves 3.1% accuracy with a slight 10% energy and 14% execution time overhead.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GEBA: Gradient-Error-Based Approximation of Activation Functions GEBA:基于梯度误差的激活函数逼近
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328890
Changmin Ye;Doo Seok Jeong
Computing-in-memory (CIM) macros aiming at accelerating deep learning operations at low power need activation function (AF) units on the same die to reduce their host-dependency. Versatile CIM macros need to include reconfigurable AF units at high precision and high efficiency in hardware usage. To this end, we propose the gradient-error-based approximation (GEBA) of AFs, which approximates various types of AFs in discrete input domains at high precision. GEBA reduces the approximation error by ca. 49.7%, 67.3%, 81.4%, 60.1% (for sigmoid, tanh, GELU, swish in FP32), compared with the uniform input-based approximation using the same memory as GEBA.
旨在以低功耗加速深度学习操作的内存计算(CIM)宏需要同一芯片上的激活函数(AF)单元,以减少对主机的依赖。多功能 CIM 宏需要包含可重新配置的高精度 AF 单元,并提高硬件使用效率。为此,我们提出了基于梯度误差的 AF 近似 (GEBA),可以高精度逼近离散输入域中的各类 AF。与使用与 GEBA 相同内存的基于均匀输入的近似方法相比,GEBA 将近似误差分别降低了约 49.7%、67.3%、81.4% 和 60.1%(对于 FP32 中的 sigmoid、tanh、GELU 和 swish)。
{"title":"GEBA: Gradient-Error-Based Approximation of Activation Functions","authors":"Changmin Ye;Doo Seok Jeong","doi":"10.1109/JETCAS.2023.3328890","DOIUrl":"10.1109/JETCAS.2023.3328890","url":null,"abstract":"Computing-in-memory (CIM) macros aiming at accelerating deep learning operations at low power need activation function (AF) units on the same die to reduce their host-dependency. Versatile CIM macros need to include reconfigurable AF units at high precision and high efficiency in hardware usage. To this end, we propose the gradient-error-based approximation (GEBA) of AFs, which approximates various types of AFs in discrete input domains at high precision. GEBA reduces the approximation error by ca. 49.7%, 67.3%, 81.4%, 60.1% (for sigmoid, tanh, GELU, swish in FP32), compared with the uniform input-based approximation using the same memory as GEBA.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Operating Coupled VO₂-Based Oscillators for Solving Ising Models 基于 VO₂的操作耦合振荡器用于求解等效模型
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328887
Maria J. Avedillo;Manuel Jiménez Través;Corentin Delacour;Aida Todri-Sanial;Bernabé Linares-Barranco;Juan Núñez
Coupled nano-oscillators are attracting increasing interest because of their potential to perform computation efficiently, enabling new applications in computing and information processing. The potential of phase transition devices for such dynamical systems has recently been recognized. This paper investigates the implementation of coupled VO2-based oscillator networks to solve combinatorial optimization problems. The target problem is mapped to an Ising model, which is solved by the synchronization dynamics of the system. Different factors that impact the probability of the system reaching the ground state of the Ising Hamiltonian and, therefore, the optimum solution to the corresponding optimization problem, are analyzed. The simulation-based analysis has led to the proposal of a novel Second-Harmonic Injection Locking (SHIL) schedule. Its main feature is that SHIL signal amplitude is repeatedly smoothly increased and decreased. Reducing SHIL strength is the mechanism that enables escaping from local minimum energy states. Our experiments show better results in terms of success probability than previously reported approaches. An experimental Oscillatory Ising Machine (OIM) has been built to validate our proposal.
由于耦合纳米振荡器具有高效计算的潜力,可在计算和信息处理领域实现新的应用,因此正吸引着越来越多的关注。相变器件在此类动态系统中的潜力最近已得到认可。本文研究了如何利用基于 VO2 的耦合振荡器网络来解决组合优化问题。目标问题被映射到一个伊辛模型,通过系统的同步动力学来解决。分析了影响系统达到伊辛哈密顿的基态概率的不同因素,从而分析了相应优化问题的最优解。通过模拟分析,提出了一种新颖的二次谐波注入锁定(SHIL)计划。它的主要特点是 SHIL 信号幅度反复平滑地增加和减少。降低 SHIL 强度是摆脱局部最小能量状态的机制。我们的实验结果表明,与之前报道的方法相比,我们的成功概率更高。为了验证我们的建议,我们建立了一个实验性的振荡伊辛机(OIM)。
{"title":"Operating Coupled VO₂-Based Oscillators for Solving Ising Models","authors":"Maria J. Avedillo;Manuel Jiménez Través;Corentin Delacour;Aida Todri-Sanial;Bernabé Linares-Barranco;Juan Núñez","doi":"10.1109/JETCAS.2023.3328887","DOIUrl":"10.1109/JETCAS.2023.3328887","url":null,"abstract":"Coupled nano-oscillators are attracting increasing interest because of their potential to perform computation efficiently, enabling new applications in computing and information processing. The potential of phase transition devices for such dynamical systems has recently been recognized. This paper investigates the implementation of coupled VO2-based oscillator networks to solve combinatorial optimization problems. The target problem is mapped to an Ising model, which is solved by the synchronization dynamics of the system. Different factors that impact the probability of the system reaching the ground state of the Ising Hamiltonian and, therefore, the optimum solution to the corresponding optimization problem, are analyzed. The simulation-based analysis has led to the proposal of a novel Second-Harmonic Injection Locking (SHIL) schedule. Its main feature is that SHIL signal amplitude is repeatedly smoothly increased and decreased. Reducing SHIL strength is the mechanism that enables escaping from local minimum energy states. Our experiments show better results in terms of success probability than previously reported approaches. An experimental Oscillatory Ising Machine (OIM) has been built to validate our proposal.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ESSM: Extended Synaptic Sampling Machine With Stochastic Echo State Neuro-Memristive Circuits ESSM:带有随机回波状态神经迷走电路的扩展突触采样机
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328875
Vineeta V. Nair;Chithra Reghuvaran;Deepu John;Bhaskar Choubey;Alex James
Synaptic stochasticity is an important feature of biological neural networks that is not widely explored in analog memristor networks. Synaptic Sampling Machine (SSM) is one of the recent models of the neural network that explores the importance of the synaptic stochasticity. In this paper, we present a memristive Echo State Network (ESN) with Extended-SSM (ESSM). The circuit-level design of the single synaptic sampling cell that can introduce stochasticity to the neural network is presented. The architecture of synaptic sampling cells is proposed that have the ability to adaptively reprogram the arrays and respond to stimuli of various strengths. The effect of stochasticity is achieved by randomly blocking the input with the probability that follows Bernoulli distribution, and can lead to the reduction of the memory capacity requirements. The blocking signals are randomly generated using Circular Shift Registers (CSRs). The network processing is handled in analog domain and the training is performed offline. The performance of the neural network is analyzed with a view to benchmark for hardware performance without compromising the system performance. The neural system was tested on ECG, MNIST, Fashion MNIST and CIFAR10 dataset for classification problem. The advantage of memristive CSR in comparison with conventional CMOS based CSR is presented. The ESSM-ESN performance is evaluated with the effect of device variations like resistance variations, noise and quantization. The advantage of ESSM-ESN is demonstrated in terms of performance and power requirements in comparison with other neural architectures.
突触随机性是生物神经网络的一个重要特征,但在模拟忆阻器网络中并未得到广泛探讨。突触采样机(SSM)是近年来探索突触随机性重要性的神经网络模型之一。在本文中,我们提出了一种带有扩展 SSM(ESSM)的忆阻器回声状态网络(ESN)。本文介绍了可为神经网络引入随机性的单个突触采样单元的电路级设计。提出的突触采样单元结构能够自适应地对阵列进行重新编程,并对不同强度的刺激做出响应。随机性的效果是通过按照伯努利分布的概率随机阻断输入来实现的,这可以降低对内存容量的要求。阻塞信号通过循环移位寄存器(CSR)随机产生。网络处理在模拟域中进行,训练在离线状态下进行。对神经网络的性能进行了分析,目的是在不影响系统性能的情况下确定硬件性能基准。神经系统在 ECG、MNIST、Fashion MNIST 和 CIFAR10 数据集上进行了分类测试。与传统的基于 CMOS 的 CSR 相比,忆阻式 CSR 的优势显而易见。在评估 ESSM-ESN 性能时,考虑了电阻变化、噪声和量化等器件变化的影响。与其他神经架构相比,ESSM-ESN 在性能和功耗要求方面的优势得到了证明。
{"title":"ESSM: Extended Synaptic Sampling Machine With Stochastic Echo State Neuro-Memristive Circuits","authors":"Vineeta V. Nair;Chithra Reghuvaran;Deepu John;Bhaskar Choubey;Alex James","doi":"10.1109/JETCAS.2023.3328875","DOIUrl":"10.1109/JETCAS.2023.3328875","url":null,"abstract":"Synaptic stochasticity is an important feature of biological neural networks that is not widely explored in analog memristor networks. Synaptic Sampling Machine (SSM) is one of the recent models of the neural network that explores the importance of the synaptic stochasticity. In this paper, we present a memristive Echo State Network (ESN) with Extended-SSM (ESSM). The circuit-level design of the single synaptic sampling cell that can introduce stochasticity to the neural network is presented. The architecture of synaptic sampling cells is proposed that have the ability to adaptively reprogram the arrays and respond to stimuli of various strengths. The effect of stochasticity is achieved by randomly blocking the input with the probability that follows Bernoulli distribution, and can lead to the reduction of the memory capacity requirements. The blocking signals are randomly generated using Circular Shift Registers (CSRs). The network processing is handled in analog domain and the training is performed offline. The performance of the neural network is analyzed with a view to benchmark for hardware performance without compromising the system performance. The neural system was tested on ECG, MNIST, Fashion MNIST and CIFAR10 dataset for classification problem. The advantage of memristive CSR in comparison with conventional CMOS based CSR is presented. The ESSM-ESN performance is evaluated with the effect of device variations like resistance variations, noise and quantization. The advantage of ESSM-ESN is demonstrated in terms of performance and power requirements in comparison with other neural architectures.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10302278","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spike Timing Dependent Gradient for Direct Training of Fast and Efficient Binarized Spiking Neural Networks 用于直接训练快速高效二值化尖峰神经网络的尖峰时序相关梯度
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328926
Zhengyu Cai;Hamid Rahimian Kalatehbali;Ben Walters;Mostafa Rahimi Azghadi;Amirali Amirsoleimani;Roman Genov
Spiking neural networks (SNNs) are well-suited for neuromorphic hardware due to their biological plausibility and energy efficiency. These networks utilize sparse, asynchronous spikes for communication and can be binarized. However, the training of such networks presents several challenges due to their non-differentiable activation function and binarized inter-layer data movement. The well-established backpropagation through time (BPTT) algorithm used to train SNNs encounters notable difficulties because of its substantial memory consumption and extensive computational demands. These limitations restrict its practical utility in real-world scenarios. Therefore, effective techniques are required to train such networks efficiently while preserving accuracy. In this paper, we propose Binarized Spike Timing Dependent Gradient (BSTDG), a novel method that utilizes presynaptic and postsynaptic timings to bypass the non-differentiable gradient and the need of BPTT. Additionally, we employ binarized weights with a threshold training strategy to enhance energy savings and performance. Moreover, we exploit latency/temporal-based coding and the Integrate-and-Fire (IF) model to achieve significant computational advantages. We evaluate the proposed method on Caltech101 Face/Motorcycle, MNIST, Fashion-MNIST, and Spiking Heidelberg Digits. The results demonstrate that the accuracy attained surpasses that of existing BSNNs and single-spike networks under the same structure. Furthermore, the proposed model achieves up to 30 $times times times $ speedup in inference and effectively reduces the number of spikes emitted in the hidden layer by 50% compared to previous works.
尖峰神经网络(SNN)具有生物学上的合理性和能效,非常适合神经形态硬件。这些网络利用稀疏的异步尖峰进行通信,并可进行二值化。然而,由于其激活函数的不可分性和二值化的层间数据移动,这类网络的训练面临着一些挑战。用于训练 SNN 的成熟的时间反向传播(BPTT)算法遇到了显著的困难,因为它需要消耗大量内存和大量计算需求。这些局限性限制了它在现实世界中的实际应用。因此,需要有效的技术来高效地训练此类网络,同时保持准确性。在本文中,我们提出了二值化尖峰时序相关梯度(BSTDG),这是一种利用突触前和突触后时序的新方法,可以绕过无差异梯度和 BPTT 的需要。此外,我们还采用了二值化权重和阈值训练策略,以提高节能效果和性能。此外,我们还利用基于时延/时序的编码和 "集成-发射"(IF)模型来实现显著的计算优势。我们在 Caltech101 Face/Motorcycle、MNIST、Fashion-MNIST 和 Spiking Heidelberg Digits 上对所提出的方法进行了评估。结果表明,在相同结构下,所达到的准确度超过了现有的 BSNN 和单尖峰网络。此外,与之前的研究相比,该模型的推理速度提高了30倍,并有效地减少了50%的隐层尖峰数量。
{"title":"Spike Timing Dependent Gradient for Direct Training of Fast and Efficient Binarized Spiking Neural Networks","authors":"Zhengyu Cai;Hamid Rahimian Kalatehbali;Ben Walters;Mostafa Rahimi Azghadi;Amirali Amirsoleimani;Roman Genov","doi":"10.1109/JETCAS.2023.3328926","DOIUrl":"10.1109/JETCAS.2023.3328926","url":null,"abstract":"Spiking neural networks (SNNs) are well-suited for neuromorphic hardware due to their biological plausibility and energy efficiency. These networks utilize sparse, asynchronous spikes for communication and can be binarized. However, the training of such networks presents several challenges due to their non-differentiable activation function and binarized inter-layer data movement. The well-established backpropagation through time (BPTT) algorithm used to train SNNs encounters notable difficulties because of its substantial memory consumption and extensive computational demands. These limitations restrict its practical utility in real-world scenarios. Therefore, effective techniques are required to train such networks efficiently while preserving accuracy. In this paper, we propose Binarized Spike Timing Dependent Gradient (BSTDG), a novel method that utilizes presynaptic and postsynaptic timings to bypass the non-differentiable gradient and the need of BPTT. Additionally, we employ binarized weights with a threshold training strategy to enhance energy savings and performance. Moreover, we exploit latency/temporal-based coding and the Integrate-and-Fire (IF) model to achieve significant computational advantages. We evaluate the proposed method on Caltech101 Face/Motorcycle, MNIST, Fashion-MNIST, and Spiking Heidelberg Digits. The results demonstrate that the accuracy attained surpasses that of existing BSNNs and single-spike networks under the same structure. Furthermore, the proposed model achieves up to 30\u0000<inline-formula> <tex-math>$times times times $ </tex-math></inline-formula>\u0000 speedup in inference and effectively reduces the number of spikes emitted in the hidden layer by 50% compared to previous works.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CBP-QSNN: Spiking Neural Networks Quantized Using Constrained Backpropagation CBP-QSNN:使用受限反向传播量化的尖峰神经网络
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328911
Donghyung Yoo;Doo Seok Jeong
Spiking Neural Networks (SNNs) support sparse event-based data processing at high power efficiency when implemented in event-based neuromorphic processors. However, the limited on- chip memory capacity of neuromorphic processors strictly delimits the depth and width of SNNs implemented. A direct solution is the use of quantized SNNs (QSNNs) in place of SNNs with FP32 weights. To this end, we propose a method to quantize the weights using constrained backpropagation (CBP) with the Lagrangian function (conventional loss function plus well-defined weight-constraint functions) as an objective function. This work utilizes CBP as a post-training algorithm for deep SNNs pre-trained using various state-of-the-art methods including direct training (TSSL-BP, STBP, and surrogate gradient) and DNN-to-SNN conversion (SNN-Calibration), validating CBP as a general framework for QSNNs. CBP-QSNNs highlight their high accuracy insomuch as the degradation of accuracy on CIFAR-10, DVS128 Gesture, and CIFAR10-DVS in the worst case is less than 1%. Particularly, CBP-QSNNs for SNN-Calibration-pretrained SNNs on CIFAR-100 highlight an unexpected large increase in accuracy by 3.72% while using small weight-memory (3.5% of the FP32 case).
尖峰神经网络(SNN)在基于事件的神经形态处理器中实现时,能以高能效支持基于稀疏事件的数据处理。然而,神经形态处理器有限的芯片内存容量严格限制了 SNN 的深度和宽度。一个直接的解决方案是使用量化 SNN(QSNN)来替代具有 FP32 权重的 SNN。为此,我们提出了一种使用约束反向传播(CBP)量化权重的方法,并将拉格朗日函数(传统损失函数加上定义明确的权重约束函数)作为目标函数。这项研究将 CBP 作为后训练算法,用于使用各种先进方法(包括直接训练(TSSL-BP、STBP 和代理梯度)和 DNN 到 SNN 转换(SNN-Calibration))预训练的深度 SNN,从而验证了 CBP 作为 QSNN 通用框架的有效性。CBP-QSNNs 突出了其高准确度,因为在最坏情况下,CIFAR-10、DVS128 Gesture 和 CIFAR10-DVS 的准确度下降不到 1%。特别是,CBP-QSNN 在 CIFAR-100 上用于 SNN-Calibration 训练的 SNN,在使用少量权重内存(FP32 案例的 3.5%)的情况下,准确率意外大幅提高了 3.72%。
{"title":"CBP-QSNN: Spiking Neural Networks Quantized Using Constrained Backpropagation","authors":"Donghyung Yoo;Doo Seok Jeong","doi":"10.1109/JETCAS.2023.3328911","DOIUrl":"10.1109/JETCAS.2023.3328911","url":null,"abstract":"Spiking Neural Networks (SNNs) support sparse event-based data processing at high power efficiency when implemented in event-based neuromorphic processors. However, the limited on- chip memory capacity of neuromorphic processors strictly delimits the depth and width of SNNs implemented. A direct solution is the use of quantized SNNs (QSNNs) in place of SNNs with FP32 weights. To this end, we propose a method to quantize the weights using constrained backpropagation (CBP) with the Lagrangian function (conventional loss function plus well-defined weight-constraint functions) as an objective function. This work utilizes CBP as a post-training algorithm for deep SNNs pre-trained using various state-of-the-art methods including direct training (TSSL-BP, STBP, and surrogate gradient) and DNN-to-SNN conversion (SNN-Calibration), validating CBP as a general framework for QSNNs. CBP-QSNNs highlight their high accuracy insomuch as the degradation of accuracy on CIFAR-10, DVS128 Gesture, and CIFAR10-DVS in the worst case is less than 1%. Particularly, CBP-QSNNs for SNN-Calibration-pretrained SNNs on CIFAR-100 highlight an unexpected large increase in accuracy by 3.72% while using small weight-memory (3.5% of the FP32 case).","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MiCE: An ANN-to-SNN Conversion Technique to Enable High Accuracy and Low Latency MiCE:实现高精度和低延迟的 ANN 到 SNN 转换技术
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328863
Nguyen-Dong Ho;Ik-Joon Chang
Spiking Neural Networks (SNNs) mimic the behavior of biological neurons. Unlike traditional Artificial Neural Networks (ANNs) that operate in a continuous time domain and use activation functions to process information, SNNs operate discrete event-driven, where data is encoded and communicated through spikes or discrete events. This unique approach offers several advantages, such as efficient computation and lower power consumption, making SNNs particularly attractive for energy-constrained and neuromorphic applications. However, training SNNs poses significant challenges due to the discrete nature of spikes and the non-differentiable behavior they exhibit. As a result, converting pre-trained ANNs into SNNs has gained attention as a convenient approach. While this approach simplifies the training process, it introduces certain drawbacks, including high latency. The conversion of ANNs to SNNs typically leads to a loss of accuracy, which can be attributed to various factors, including quantization, clipping, and timing errors. Previous studies have proposed techniques to mitigate quantization and clipping errors during the conversion process. However, they do not consider timing errors, degrading SNN accuracies at low latency conditions. This work introduces the MiCE conversion method, which offers a comprehensive joint optimization strategy to simultaneously alleviate quantization, clipping, and timing errors. At a moderate latency of 8 time-steps, our converted ResNet-20 achieves classification accuracies of 79.02% and 95.74% on the CIFAR-100 and CIFAR-10 datasets, respectively.
尖峰神经网络(SNN)模仿生物神经元的行为。传统的人工神经网络(ANN)在连续时域中运行并使用激活函数来处理信息,与之不同的是,SNN 以离散事件为驱动,通过尖峰或离散事件对数据进行编码和通信。这种独特的方法具有多种优势,例如计算效率高、功耗低,因此 SNN 对能源受限和神经形态应用特别有吸引力。然而,由于尖峰的离散性及其表现出的无差异行为,训练 SNNs 面临着巨大的挑战。因此,将预先训练好的 ANNs 转换为 SNNs 作为一种便捷的方法受到了关注。虽然这种方法简化了训练过程,但也带来了一些缺点,包括高延迟。将 ANNs 转换为 SNNs 通常会导致精度下降,这可归因于量化、削波和定时误差等各种因素。以往的研究提出了一些技术,以减少转换过程中的量化和削波误差。但是,它们没有考虑时序误差,从而降低了低延迟条件下的 SNN 精度。这项工作引入了 MiCE 转换方法,它提供了一种全面的联合优化策略,可同时减轻量化、削波和时序误差。在 8 个时间步的中等延迟条件下,我们转换后的 ResNet-20 在 CIFAR-100 和 CIFAR-10 数据集上的分类准确率分别达到了 79.02% 和 95.74%。
{"title":"MiCE: An ANN-to-SNN Conversion Technique to Enable High Accuracy and Low Latency","authors":"Nguyen-Dong Ho;Ik-Joon Chang","doi":"10.1109/JETCAS.2023.3328863","DOIUrl":"10.1109/JETCAS.2023.3328863","url":null,"abstract":"Spiking Neural Networks (SNNs) mimic the behavior of biological neurons. Unlike traditional Artificial Neural Networks (ANNs) that operate in a continuous time domain and use activation functions to process information, SNNs operate discrete event-driven, where data is encoded and communicated through spikes or discrete events. This unique approach offers several advantages, such as efficient computation and lower power consumption, making SNNs particularly attractive for energy-constrained and neuromorphic applications. However, training SNNs poses significant challenges due to the discrete nature of spikes and the non-differentiable behavior they exhibit. As a result, converting pre-trained ANNs into SNNs has gained attention as a convenient approach. While this approach simplifies the training process, it introduces certain drawbacks, including high latency. The conversion of ANNs to SNNs typically leads to a loss of accuracy, which can be attributed to various factors, including quantization, clipping, and timing errors. Previous studies have proposed techniques to mitigate quantization and clipping errors during the conversion process. However, they do not consider timing errors, degrading SNN accuracies at low latency conditions. This work introduces the MiCE conversion method, which offers a comprehensive joint optimization strategy to simultaneously alleviate quantization, clipping, and timing errors. At a moderate latency of 8 time-steps, our converted ResNet-20 achieves classification accuracies of 79.02% and 95.74% on the CIFAR-100 and CIFAR-10 datasets, respectively.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Neuromorphic Computing With Address-Event-Representation Using Time-to-Event Margin Propagation 利用时间到事件边际传播进行地址到事件表示的神经形态计算
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328916
R. Madhuvanthi Srivatsav;Shantanu Chakrabartty;Chetan Singh Thakur
Address-Event-Representation (AER) is a spike-routing protocol that allows the scaling of neuromorphic and spiking neural network (SNN) architectures. However, in conventional neuromorphic architectures, the AER protocol and in general, any virtual interconnect plays only a passive role in computation, i.e., only for routing spikes and events. In this paper, we show how causal temporal primitives like delay, triggering, and sorting inherent in the AER protocol itself can be exploited for scalable neuromorphic computing using our proposed technique called Time-to-Event Margin Propagation (TEMP). The proposed TEMP-based AER architecture is fully asynchronous and relies on interconnect delays for memory and computing as opposed to conventional and local multiply-and-accumulate (MAC) operations. We show that the time-based encoding in the TEMP neural network produces a spatio-temporal representation that can encode a large number of discriminatory patterns. As a proof-of-concept, we show that a trained TEMP-based convolutional neural network (CNN) can demonstrate an accuracy greater than 99% on the MNIST dataset and 91.2% on the Fashion MNIST Dataset. Overall, our work is a biologically inspired computing paradigm that brings forth a new dimension of research to the field of neuromorphic computing.
地址-事件-表示(AER)是一种尖峰路由协议,可以扩展神经形态和尖峰神经网络(SNN)架构。然而,在传统的神经形态架构中,AER 协议和一般的虚拟互连在计算中仅扮演被动角色,即仅用于路由尖峰和事件。在本文中,我们展示了如何利用 AER 协议本身固有的因果时间基元(如延迟、触发和排序),通过我们提出的时间到事件边际传播(TEMP)技术,实现可扩展的神经形态计算。所提出的基于 TEMP 的 AER 架构是完全异步的,它依赖于内存和计算的互连延迟,而不是传统的本地乘法累加 (MAC) 操作。我们的研究表明,TEMP 神经网络中基于时间的编码产生了一种时空表示,可以编码大量的判别模式。作为概念验证,我们展示了经过训练的基于 TEMP 的卷积神经网络 (CNN) 在 MNIST 数据集上的准确率超过 99%,在时尚 MNIST 数据集上的准确率超过 91.2%。总之,我们的工作是一种受生物启发的计算范例,为神经形态计算领域带来了新的研究维度。
{"title":"Neuromorphic Computing With Address-Event-Representation Using Time-to-Event Margin Propagation","authors":"R. Madhuvanthi Srivatsav;Shantanu Chakrabartty;Chetan Singh Thakur","doi":"10.1109/JETCAS.2023.3328916","DOIUrl":"10.1109/JETCAS.2023.3328916","url":null,"abstract":"Address-Event-Representation (AER) is a spike-routing protocol that allows the scaling of neuromorphic and spiking neural network (SNN) architectures. However, in conventional neuromorphic architectures, the AER protocol and in general, any virtual interconnect plays only a passive role in computation, i.e., only for routing spikes and events. In this paper, we show how causal temporal primitives like delay, triggering, and sorting inherent in the AER protocol itself can be exploited for scalable neuromorphic computing using our proposed technique called Time-to-Event Margin Propagation (TEMP). The proposed TEMP-based AER architecture is fully asynchronous and relies on interconnect delays for memory and computing as opposed to conventional and local multiply-and-accumulate (MAC) operations. We show that the time-based encoding in the TEMP neural network produces a spatio-temporal representation that can encode a large number of discriminatory patterns. As a proof-of-concept, we show that a trained TEMP-based convolutional neural network (CNN) can demonstrate an accuracy greater than 99% on the MNIST dataset and 91.2% on the Fashion MNIST Dataset. Overall, our work is a biologically inspired computing paradigm that brings forth a new dimension of research to the field of neuromorphic computing.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking DNN Mapping Methods for the in-Memory Computing Accelerators 内存计算加速器的 DNN 映射方法基准测试
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-31 DOI: 10.1109/JETCAS.2023.3328864
Yimin Wang;Xuanyao Fong
This paper presents a study of methods for mapping the convolutional workloads in deep neural networks (DNNs) onto the computing hardware in the in-memory computing (IMC) architecture. Specifically, we focus on categorizing and benchmarking the processing element (PE)-level mapping methods, which have not been investigated in detail for IMC-based architectures. First, we categorize the PE-level mapping methods from the loop unrolling perspective and discuss the corresponding implications on input data reuse and output data reduction. Then, a mapping-oriented architecture is proposed by considering the input and output datapaths under various mapping methods. The architecture is evaluated on the 45 nm technology showing good area-efficiency and scalability, providing a hardware substrate for further performance improvements via PE-level mappings. Furthermore, we present an evaluation framework that captures the architecture behaviors and enables extensive benchmarking of mapping methods under various neural network workloads, main memory bandwidth, and digital computing throughput. The benchmarking results demonstrate significant tradeoffs in the design space and unlock new design possibilities. We present case studies to showcase preferred mapping methods for best energy consumption and/or execution time and demonstrate that a hybrid-mapping scheme enhances minimum execution time by up to 30% for the publicly-available DNN benchmarks.
本文研究了将深度神经网络(DNN)中的卷积工作量映射到内存计算(IMC)架构中的计算硬件上的方法。具体而言,我们将重点放在处理元件(PE)级映射方法的分类和基准测试上,这些方法尚未针对基于 IMC 的架构进行详细研究。首先,我们从循环展开的角度对 PE 级映射方法进行分类,并讨论其对输入数据重用和输出数据缩减的相应影响。然后,通过考虑各种映射方法下的输入和输出数据通路,提出了一种面向映射的架构。该架构在 45 纳米技术上进行了评估,显示出良好的面积效率和可扩展性,为通过 PE 级映射进一步提高性能提供了硬件基础。此外,我们还提出了一个评估框架,可捕捉架构行为,并在各种神经网络工作负载、主存储器带宽和数字计算吞吐量下对映射方法进行广泛的基准测试。基准测试结果表明了设计空间中的重大权衡,并揭示了新的设计可能性。我们通过案例研究展示了最佳能耗和/或执行时间的首选映射方法,并证明混合映射方案可将公开的 DNN 基准的最短执行时间最多延长 30%。
{"title":"Benchmarking DNN Mapping Methods for the in-Memory Computing Accelerators","authors":"Yimin Wang;Xuanyao Fong","doi":"10.1109/JETCAS.2023.3328864","DOIUrl":"10.1109/JETCAS.2023.3328864","url":null,"abstract":"This paper presents a study of methods for mapping the convolutional workloads in deep neural networks (DNNs) onto the computing hardware in the in-memory computing (IMC) architecture. Specifically, we focus on categorizing and benchmarking the processing element (PE)-level mapping methods, which have not been investigated in detail for IMC-based architectures. First, we categorize the PE-level mapping methods from the loop unrolling perspective and discuss the corresponding implications on input data reuse and output data reduction. Then, a mapping-oriented architecture is proposed by considering the input and output datapaths under various mapping methods. The architecture is evaluated on the 45 nm technology showing good area-efficiency and scalability, providing a hardware substrate for further performance improvements via PE-level mappings. Furthermore, we present an evaluation framework that captures the architecture behaviors and enables extensive benchmarking of mapping methods under various neural network workloads, main memory bandwidth, and digital computing throughput. The benchmarking results demonstrate significant tradeoffs in the design space and unlock new design possibilities. We present case studies to showcase preferred mapping methods for best energy consumption and/or execution time and demonstrate that a hybrid-mapping scheme enhances minimum execution time by up to 30% for the publicly-available DNN benchmarks.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135263023","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design Space Exploration of Sparsity-Aware Application-Specific Spiking Neural Network Accelerators 稀疏感知特定应用尖峰神经网络加速器的设计空间探索
IF 4.6 2区 工程技术 Q1 Engineering Pub Date : 2023-10-30 DOI: 10.1109/JETCAS.2023.3327746
Ilkin Aliyev;Kama Svoboda;Tosiron Adegbija
Spiking Neural Networks (SNNs) offer a promising alternative to Artificial Neural Networks (ANNs) for deep learning applications, particularly in resource-constrained systems. This is largely due to their inherent sparsity, influenced by factors such as the input dataset, the length of the spike train, and the network topology. While a few prior works have demonstrated the advantages of incorporating sparsity into the hardware design, especially in terms of reducing energy consumption, the impact on hardware resources has not yet been explored. This is where design space exploration (DSE) becomes crucial, as it allows for the optimization of hardware performance by tailoring both the hardware and model parameters to suit specific application needs. However, DSE can be extremely challenging given the potentially large design space and the interplay of hardware architecture design choices and application-specific model parameters. In this paper, we propose a flexible hardware design that leverages the sparsity of SNNs to identify highly efficient, application-specific accelerator designs. We develop a high-level, cycle-accurate simulation framework for this hardware and demonstrate the framework’s benefits in enabling detailed and fine-grained exploration of SNN design choices, such as the layer-wise logical-to-hardware ratio (LHR). Our experimental results show that our design can (i) achieve up to 76% reduction in hardware resources and (ii) deliver a speed increase of up to $31.25times $ , while requiring 27% fewer hardware resources compared to sparsity-oblivious designs. We further showcase the robustness of our framework by varying spike train lengths with different neuron population sizes to find the optimal trade-off points between accuracy and hardware latency.
尖峰神经网络(SNN)为深度学习应用,尤其是资源受限系统的深度学习应用,提供了一种替代人工神经网络(ANN)的前景广阔的选择。这主要是由于其固有的稀疏性,受输入数据集、尖峰训练长度和网络拓扑结构等因素的影响。虽然之前的一些工作已经证明了将稀疏性纳入硬件设计的优势,特别是在降低能耗方面,但对硬件资源的影响还没有进行过探索。这正是设计空间探索(DSE)的关键所在,因为它可以通过调整硬件和模型参数来优化硬件性能,以满足特定的应用需求。然而,由于设计空间可能很大,而且硬件架构设计选择与特定应用模型参数之间存在相互作用,因此 DSE 极具挑战性。在本文中,我们提出了一种灵活的硬件设计,利用 SNN 的稀疏性来识别高效的特定应用加速器设计。我们为该硬件开发了一个高层次、周期精确的仿真框架,并展示了该框架在详细、精细地探索 SNN 设计选择(如层逻辑硬件比 (LHR))方面的优势。我们的实验结果表明,我们的设计可以:(i) 减少高达 76% 的硬件资源;(ii) 实现高达 31.25/times $ 的速度提升,同时与不依赖稀疏性的设计相比,所需的硬件资源减少了 27%。我们进一步展示了我们框架的鲁棒性,即通过不同的神经元群体大小来改变尖峰训练长度,从而在准确性和硬件延迟之间找到最佳权衡点。
{"title":"Design Space Exploration of Sparsity-Aware Application-Specific Spiking Neural Network Accelerators","authors":"Ilkin Aliyev;Kama Svoboda;Tosiron Adegbija","doi":"10.1109/JETCAS.2023.3327746","DOIUrl":"10.1109/JETCAS.2023.3327746","url":null,"abstract":"Spiking Neural Networks (SNNs) offer a promising alternative to Artificial Neural Networks (ANNs) for deep learning applications, particularly in resource-constrained systems. This is largely due to their inherent sparsity, influenced by factors such as the input dataset, the length of the spike train, and the network topology. While a few prior works have demonstrated the advantages of incorporating sparsity into the hardware design, especially in terms of reducing energy consumption, the impact on hardware resources has not yet been explored. This is where design space exploration (DSE) becomes crucial, as it allows for the optimization of hardware performance by tailoring both the hardware and model parameters to suit specific application needs. However, DSE can be extremely challenging given the potentially large design space and the interplay of hardware architecture design choices and application-specific model parameters. In this paper, we propose a flexible hardware design that leverages the sparsity of SNNs to identify highly efficient, application-specific accelerator designs. We develop a high-level, cycle-accurate simulation framework for this hardware and demonstrate the framework’s benefits in enabling detailed and fine-grained exploration of SNN design choices, such as the layer-wise logical-to-hardware ratio (LHR). Our experimental results show that our design can (i) achieve up to 76% reduction in hardware resources and (ii) deliver a speed increase of up to \u0000<inline-formula> <tex-math>$31.25times $ </tex-math></inline-formula>\u0000, while requiring 27% fewer hardware resources compared to sparsity-oblivious designs. We further showcase the robustness of our framework by varying spike train lengths with different neuron population sizes to find the optimal trade-off points between accuracy and hardware latency.","PeriodicalId":48827,"journal":{"name":"IEEE Journal on Emerging and Selected Topics in Circuits and Systems","volume":null,"pages":null},"PeriodicalIF":4.6,"publicationDate":"2023-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"135260934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Journal on Emerging and Selected Topics in Circuits and Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1