NEST-C: A deep learning compiler framework for heterogeneous computing systems with artificial intelligence accelerators
https://doi.org/10.4218/etrij.2024-0139
ETRI Journal, Volume 46, Issue 5, October 2024, pp. 851–864.
In the article entitled “NEST-C: A deep learning compiler framework for heterogeneous computing systems with artificial intelligence accelerators,” the authors would like to correct the funding information of their article. It should be written as follows:
Funding information This study is supported by a grant from the Institute of Information & Communications Technology Planning & Evaluation (IITP), funded by the Korean government (MSIT) (No. RS-2023-00277060, Development of OpenEdge AI SoC hardware and software platform and No. 2018-0-00769, Neuromorphic Computing Software Platform for Artificial Intelligence Systems).
The authors would like to apologize for the inconvenience caused.
Sungryeul Rhyu | Junsik Kim | Gwang Hoon Park | Kyuheon Kim
Low-complexity patch projection method for efficient and lightweight point-cloud compression
https://doi.org/10.4218/etrij.2023-0242
ETRI Journal, Volume 46, Issue 4, August 2024, pp. 683–696.
In the article entitled “Low-complexity patch projection method for efficient and lightweight point-cloud compression”, the authors would like to correct the funding information of their article. It should be written as follows:
Funding information
This study was supported by the Information Technology Research Center of the Ministry of Science and ICT, Korea (grant number: IITP-2024-2021-0-02046) and the Institute of Information & Communications Technology Planning & Evaluation, Korea (grant number: RS-2023-00227431, Development of 3D space digital media standard technology).
The authors would like to apologize for the inconvenience caused.
SoCs with analog-circuit-based unsigned weight-accumulating spiking neural networks (UWA-SNNs) are a highly promising solution for achieving low-power AI-SoCs. This paper addresses the challenges that must be overcome to realize the potential of UWA-SNNs in low-power AI-SoCs: (i) the absence of UWA-SNN learning methods and the lack of an environment for developing applications based on trained SNN models and (ii) the inherent issue of testing and validating applications on the system being nearly impractical until the final chip is fabricated owing to the mixed-signal circuit implementation of UWA-SNN-based SoCs. This paper argues that, by integrating the proposed solutions, the development of an EDA tool that enables the easy and rapid development of UWA-SNN-based SoCs is feasible, and demonstrates this through the development of the SNN eXpress (SNX) tool. The developed SNX automates the generation of RTL code, FPGA prototypes, and a software development kit tailored for UWA-SNN-based application development. Comprehensive details of SNX development and the performance evaluation and verification results of two AI-SoCs developed using SNX are also presented.
Owing to the widespread advancement of transformer-based artificial neural networks, artificial intelligence (AI) processors are now required to perform matrix–vector multiplication in addition to the conventional matrix–matrix multiplication. However, current AI processor architectures are optimized for general matrix–matrix multiplications (GEMMs), which causes significant throughput degradation when processing general matrix–vector multiplications (GEMVs). In this study, we proposed a port-folding GEMV (PF-GEMV) scheme employing multiformat and low-precision techniques while reusing an outer product-based processor optimized for conventional GEMM operations. This approach achieves 93.7% utilization in GEMV operations with an 8-bit format on an 8