{"title":"位深度神经网络近似推理的高效硬件加速","authors":"Sebastian Vogel, A. Guntoro, G. Ascheid","doi":"10.1109/DASIP.2017.8122127","DOIUrl":null,"url":null,"abstract":"In recent years, Deep Neural Networks (DNNs) have been of special interest in the area of image processing and scene perception. Albeit being effective and accurate, DNNs demand challenging computational resources. Fortunately, dedicated low bitwidth accelerators enable efficient, real-time inference of DNNs. We present an approximate evaluation method and a specialized multiplierless accelerator for the recently proposed bitwise DNNs. Our approximate evaluation method is based on the speculative recomputation of selective parts of a bitwise neural network. The selection is based on the intermediate results of a previous input evaluation. In context with limited energy budgets, our method and accelerator enable a fast, power efficient, first decision. If necessary, a reliable and accurate output is available after reevaluating the input data multiple times in an approximate manner. Our experiments on the GTSRB and CIFAR-10 dataset show that this approach results in no loss of classification performance in comparison with floating-point evaluation. Our work contributes to efficient inference of neural networks on power-constrained embedded devices.","PeriodicalId":6637,"journal":{"name":"2017 Conference on Design and Architectures for Signal and Image Processing (DASIP)","volume":"104 1","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2017-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Efficient hardware acceleration for approximate inference of bitwise deep neural networks\",\"authors\":\"Sebastian Vogel, A. Guntoro, G. Ascheid\",\"doi\":\"10.1109/DASIP.2017.8122127\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, Deep Neural Networks (DNNs) have been of special interest in the area of image processing and scene perception. Albeit being effective and accurate, DNNs demand challenging computational resources. Fortunately, dedicated low bitwidth accelerators enable efficient, real-time inference of DNNs. We present an approximate evaluation method and a specialized multiplierless accelerator for the recently proposed bitwise DNNs. Our approximate evaluation method is based on the speculative recomputation of selective parts of a bitwise neural network. The selection is based on the intermediate results of a previous input evaluation. In context with limited energy budgets, our method and accelerator enable a fast, power efficient, first decision. If necessary, a reliable and accurate output is available after reevaluating the input data multiple times in an approximate manner. Our experiments on the GTSRB and CIFAR-10 dataset show that this approach results in no loss of classification performance in comparison with floating-point evaluation. Our work contributes to efficient inference of neural networks on power-constrained embedded devices.\",\"PeriodicalId\":6637,\"journal\":{\"name\":\"2017 Conference on Design and Architectures for Signal and Image Processing (DASIP)\",\"volume\":\"104 1\",\"pages\":\"1-6\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2017-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2017 Conference on Design and Architectures for Signal and Image Processing (DASIP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DASIP.2017.8122127\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 Conference on Design and Architectures for Signal and Image Processing (DASIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DASIP.2017.8122127","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Efficient hardware acceleration for approximate inference of bitwise deep neural networks
In recent years, Deep Neural Networks (DNNs) have been of special interest in the area of image processing and scene perception. Albeit being effective and accurate, DNNs demand challenging computational resources. Fortunately, dedicated low bitwidth accelerators enable efficient, real-time inference of DNNs. We present an approximate evaluation method and a specialized multiplierless accelerator for the recently proposed bitwise DNNs. Our approximate evaluation method is based on the speculative recomputation of selective parts of a bitwise neural network. The selection is based on the intermediate results of a previous input evaluation. In context with limited energy budgets, our method and accelerator enable a fast, power efficient, first decision. If necessary, a reliable and accurate output is available after reevaluating the input data multiple times in an approximate manner. Our experiments on the GTSRB and CIFAR-10 dataset show that this approach results in no loss of classification performance in comparison with floating-point evaluation. Our work contributes to efficient inference of neural networks on power-constrained embedded devices.