L. Chang, Chenglong Li, Xin Zhao, Shuisheng Lin, Jun Zhou
{"title":"IPOCIM: Artificial Intelligent Processor with Adaptive Ping-pong Computing-in-Memory Architecture","authors":"L. Chang, Chenglong Li, Xin Zhao, Shuisheng Lin, Jun Zhou","doi":"10.1109/ICTA56932.2022.9963134","DOIUrl":null,"url":null,"abstract":"Computing-in-memory (CIM) architecture is a promising solution toward energy-efficient artificial intelligent (AI) processor. Practically, the AI processor with CIM engine induces a series of issues including data updating and flexibility. For instance, in AI-oriented applications, the weight stored in the CIM must be reloaded due to the huge gap between limited capacity of CIM and growing weight parameter, which greatly reduces the computation efficiency of the AI processor. Moreover, the natural parallelism of CIM leads to the mismatch of various convolution kernel sizes in different networks and layers, which reduces hardware utilization efficiency. In this work, we explore a CIM engine with a ping-pong strategy as an alternative to traditional CIM macro and weight buffer, hiding the data update latency to enhance data reuse. In addition, we proposed a flexible CIM architecture adapting to different neural networks, namely IPOCIM, with a fine-grained data-flow mapping strategy. Based on the evaluation, IPOCIM achieves 1.4-7.1× performance improvement, and 2.2-6.1× energy efficiency, compared to baseline.","PeriodicalId":325602,"journal":{"name":"2022 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA)","volume":"68 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Integrated Circuits, Technologies and Applications (ICTA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTA56932.2022.9963134","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Computing-in-memory (CIM) architecture is a promising solution toward energy-efficient artificial intelligent (AI) processor. Practically, the AI processor with CIM engine induces a series of issues including data updating and flexibility. For instance, in AI-oriented applications, the weight stored in the CIM must be reloaded due to the huge gap between limited capacity of CIM and growing weight parameter, which greatly reduces the computation efficiency of the AI processor. Moreover, the natural parallelism of CIM leads to the mismatch of various convolution kernel sizes in different networks and layers, which reduces hardware utilization efficiency. In this work, we explore a CIM engine with a ping-pong strategy as an alternative to traditional CIM macro and weight buffer, hiding the data update latency to enhance data reuse. In addition, we proposed a flexible CIM architecture adapting to different neural networks, namely IPOCIM, with a fine-grained data-flow mapping strategy. Based on the evaluation, IPOCIM achieves 1.4-7.1× performance improvement, and 2.2-6.1× energy efficiency, compared to baseline.