K. Kunal, Jitesh Poojary, S. Ramprasath, Ramesh Harjani, S. Sapatnekar
{"title":"Automated synthesis of mixed-signal ML inference hardware under accuracy constraints","authors":"K. Kunal, Jitesh Poojary, S. Ramprasath, Ramesh Harjani, S. Sapatnekar","doi":"10.1109/ASP-DAC58780.2024.10473942","DOIUrl":null,"url":null,"abstract":"Due to the inherent error-tolerance of machine learning (ML) algorithms, many parts of the inference computation can be performed with adequate accuracy and low power under relatively low precision. Early approaches have used digital approximate computing methods to explore this space. Recent approaches using analog-based operations achieve power-efficient computation at moderate precision. This work proposes a mixed-signal optimization (MiSO) approach that optimally blends analog and digital computation for ML inference. Based on accuracy and power models, an integer linear programming formulation is used to optimize design metrics of analog/digital implementations. The efficacy of the method is demonstrated on multiple ML architectures.","PeriodicalId":518586,"journal":{"name":"2024 29th Asia and South Pacific Design Automation Conference (ASP-DAC)","volume":"186 1","pages":"478-483"},"PeriodicalIF":0.0000,"publicationDate":"2024-01-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 29th Asia and South Pacific Design Automation Conference (ASP-DAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ASP-DAC58780.2024.10473942","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Due to the inherent error-tolerance of machine learning (ML) algorithms, many parts of the inference computation can be performed with adequate accuracy and low power under relatively low precision. Early approaches have used digital approximate computing methods to explore this space. Recent approaches using analog-based operations achieve power-efficient computation at moderate precision. This work proposes a mixed-signal optimization (MiSO) approach that optimally blends analog and digital computation for ML inference. Based on accuracy and power models, an integer linear programming formulation is used to optimize design metrics of analog/digital implementations. The efficacy of the method is demonstrated on multiple ML architectures.
由于机器学习(ML)算法固有的容错性,推理计算的许多部分都可以在相对较低的精度下以足够的精度和较低的功耗执行。早期的方法使用数字近似计算方法来探索这一空间。最近的方法使用基于模拟的运算,在中等精度下实现了高能效计算。本研究提出了一种混合信号优化(MiSO)方法,可将模拟计算和数字计算最佳地融合到 ML 推断中。在精度和功耗模型的基础上,使用整数线性规划公式来优化模拟/数字实现的设计指标。该方法在多种 ML 架构上的功效得到了验证。