Neural architecture search for in-memory computing-based deep learning accelerators

Olga Krestinskaya, Mohammed E. Fouda, Hadjer Benmeziane, Kaoutar El Maghraoui, Abu Sebastian, Wei D. Lu, Mario Lanza, Hai Li, Fadi Kurdahi, Suhaib A. Fahmy, Ahmed Eltawil, Khaled N. Salama
{"title":"Neural architecture search for in-memory computing-based deep learning accelerators","authors":"Olga Krestinskaya, Mohammed E. Fouda, Hadjer Benmeziane, Kaoutar El Maghraoui, Abu Sebastian, Wei D. Lu, Mario Lanza, Hai Li, Fadi Kurdahi, Suhaib A. Fahmy, Ahmed Eltawil, Khaled N. Salama","doi":"10.1038/s44287-024-00052-7","DOIUrl":null,"url":null,"abstract":"The rapid growth of artificial intelligence and the increasing complexity of neural network models are driving demand for efficient hardware architectures that can address power-constrained and resource-constrained deployments. In this context, the emergence of in-memory computing (IMC) stands out as a promising technology. For this purpose, several IMC devices, circuits and architectures have been developed. However, the intricate nature of designing, implementing and deploying such architectures necessitates a well-orchestrated toolchain for hardware–software co-design. This toolchain must allow IMC-aware optimizations across the entire stack, encompassing devices, circuits, chips, compilers, software and neural network design. The complexity and sheer size of the design space involved renders manual optimizations impractical. To mitigate these challenges, hardware-aware neural architecture search (HW-NAS) has emerged as a promising approach to accelerate the design of streamlined neural networks tailored for efficient deployment on IMC hardware. This Review illustrates the application of HW-NAS to the specific features of IMC hardware and compares existing optimization frameworks. Ongoing research and unresolved issues are discussed. A roadmap for the evolution of HW-NAS for IMC architectures is proposed. Hardware-aware neural architecture search (HW-NAS) can be used to design efficient in-memory computing (IMC) hardware for deep learning accelerators. This Review discusses methodologies, frameworks, ongoing research, open issues and recommendations, and provides a roadmap for HW-NAS for IMC.","PeriodicalId":501701,"journal":{"name":"Nature Reviews Electrical Engineering","volume":"1 6","pages":"374-390"},"PeriodicalIF":0.0000,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.nature.com/articles/s44287-024-00052-7.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nature Reviews Electrical Engineering","FirstCategoryId":"1085","ListUrlMain":"https://www.nature.com/articles/s44287-024-00052-7","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The rapid growth of artificial intelligence and the increasing complexity of neural network models are driving demand for efficient hardware architectures that can address power-constrained and resource-constrained deployments. In this context, the emergence of in-memory computing (IMC) stands out as a promising technology. For this purpose, several IMC devices, circuits and architectures have been developed. However, the intricate nature of designing, implementing and deploying such architectures necessitates a well-orchestrated toolchain for hardware–software co-design. This toolchain must allow IMC-aware optimizations across the entire stack, encompassing devices, circuits, chips, compilers, software and neural network design. The complexity and sheer size of the design space involved renders manual optimizations impractical. To mitigate these challenges, hardware-aware neural architecture search (HW-NAS) has emerged as a promising approach to accelerate the design of streamlined neural networks tailored for efficient deployment on IMC hardware. This Review illustrates the application of HW-NAS to the specific features of IMC hardware and compares existing optimization frameworks. Ongoing research and unresolved issues are discussed. A roadmap for the evolution of HW-NAS for IMC architectures is proposed. Hardware-aware neural architecture search (HW-NAS) can be used to design efficient in-memory computing (IMC) hardware for deep learning accelerators. This Review discusses methodologies, frameworks, ongoing research, open issues and recommendations, and provides a roadmap for HW-NAS for IMC.

Abstract Image

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于内存计算的深度学习加速器的神经架构搜索
人工智能的快速发展和神经网络模型复杂性的不断提高,推动了对高效硬件架构的需求,这种架构可以解决功耗和资源受限的部署问题。在此背景下,内存计算(IMC)作为一种前景广阔的技术脱颖而出。为此,人们开发了多种 IMC 设备、电路和架构。然而,由于设计、实施和部署此类架构的复杂性,有必要为硬件-软件协同设计提供一个精心设计的工具链。该工具链必须允许在整个堆栈中进行 IMC 感知优化,包括器件、电路、芯片、编译器、软件和神经网络设计。设计空间的复杂性和规模使得手动优化变得不切实际。为缓解这些挑战,硬件感知神经架构搜索(HW-NAS)已成为一种很有前途的方法,可加速设计精简的神经网络,以便在 IMC 硬件上高效部署。本综述说明了 HW-NAS 在 IMC 硬件特定功能上的应用,并对现有优化框架进行了比较。文中还讨论了正在进行的研究和尚未解决的问题。本文提出了针对 IMC 架构的 HW-NAS 发展路线图。硬件感知神经架构搜索(HW-NAS)可用于为深度学习加速器设计高效的内存计算(IMC)硬件。本综述讨论了方法论、框架、正在进行的研究、开放性问题和建议,并提供了用于 IMC 的 HW-NAS 路线图。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
The future of 2D spintronics Spintronics for ultra-low-power circuits and systems Spin-transfer torque magnetoresistive random access memory technology status and future directions Perpendicularly magnetized materials for energy-efficient orbitronics Spintronic neural systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1