{"title":"A 28-nm 16-kb Aggregation and Combination Computing-in-Memory Macro With Dual-Level Sparsity Modulation and Sparse-Tracking ADCs for GCNs","authors":"Zhaoyang Zhang;Yanqi Zhang;Feiran Liu;Zhichao Liu;Yinhai Gao;Yuchen Ma;Yutong Zhang;An Guo;Tianzhu Xiong;Jinwu Chen;Xi Chen;Bo Wang;Yuchen Tang;Jun Yang;Xin Si","doi":"10.1109/JSSC.2024.3472115","DOIUrl":null,"url":null,"abstract":"Computing-in-memory (CIM) architectures have demonstrated remarkable potential in addressing the memory wall. However, previous CIMs were often designed for multiply-accumulate (MAC) operations, which presents a myriad of challenges when deploying graph convolutional networks (GCNs) on CIM macros. This work presents a compact 6T SRAM-based aggregation and combination CIM macro (ACCIM) using: 1) a cell array with compact 6T bitcells, local jump and computing cells (LJCCs), and sparse-tracking analog-to-digital converters (STADCs) to improve energy efficiency in both input sparsity and weight bit sparsity; 2) an LJCC and STADC to improve the signal margin; 3) a CIM macro architecture with an aggregation input unit (AGINU) to support both GCN aggregation and GCN combination/convolutional neural network (CNN) MAC; 4) a graph pruning algorithm that divides the graph data into memory-friendly subgraphs; and 5) an error modeling-based pre-training method to improve the inference accuracy. A fabricated 28-nm 16-kb charge-domain SRAM-CIM macro achieved an energy efficiency of 86.87 TOPS/W and an area efficiency of 2344 GOPS/mm2 for GCNs with 4-bit degree input, 8-bit features, and 15-bit output.","PeriodicalId":13129,"journal":{"name":"IEEE Journal of Solid-state Circuits","volume":"60 3","pages":"949-962"},"PeriodicalIF":5.6000,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Solid-state Circuits","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10720191/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Computing-in-memory (CIM) architectures have demonstrated remarkable potential in addressing the memory wall. However, previous CIMs were often designed for multiply-accumulate (MAC) operations, which presents a myriad of challenges when deploying graph convolutional networks (GCNs) on CIM macros. This work presents a compact 6T SRAM-based aggregation and combination CIM macro (ACCIM) using: 1) a cell array with compact 6T bitcells, local jump and computing cells (LJCCs), and sparse-tracking analog-to-digital converters (STADCs) to improve energy efficiency in both input sparsity and weight bit sparsity; 2) an LJCC and STADC to improve the signal margin; 3) a CIM macro architecture with an aggregation input unit (AGINU) to support both GCN aggregation and GCN combination/convolutional neural network (CNN) MAC; 4) a graph pruning algorithm that divides the graph data into memory-friendly subgraphs; and 5) an error modeling-based pre-training method to improve the inference accuracy. A fabricated 28-nm 16-kb charge-domain SRAM-CIM macro achieved an energy efficiency of 86.87 TOPS/W and an area efficiency of 2344 GOPS/mm2 for GCNs with 4-bit degree input, 8-bit features, and 15-bit output.
期刊介绍:
The IEEE Journal of Solid-State Circuits publishes papers each month in the broad area of solid-state circuits with particular emphasis on transistor-level design of integrated circuits. It also provides coverage of topics such as circuits modeling, technology, systems design, layout, and testing that relate directly to IC design. Integrated circuits and VLSI are of principal interest; material related to discrete circuit design is seldom published. Experimental verification is strongly encouraged.