{"title":"Scaling Bit-Flexible Neural Networks","authors":"Yun-Nan Chang, Yu-Tang Tin","doi":"10.1109/ISOCC47750.2019.9078506","DOIUrl":null,"url":null,"abstract":"This paper proposes a neural network training scheme in order to obtain the network weights represented in the fixed-point number format such that under the different truncated lengths of the weights, our neural new network can all achieve near-optimized inference accuracy at the corresponding word-length. The similar idea has been explored before; however, the salient feature of our proposed scaling bit-progressive method is we have further taken into account the use and training of weight scaling factor, which can significant improve the inference accuracy. Our experimental results show that our trained Resnet- 18 neural network can improve the top-1 and top-5 accuracies of Tiny-ImageNet dataset by the average of 11.02% and 9.21% compared with the previous work without using the scaling factor. The top-1 and top-5 accuracy losses compared with float-point weights are only about 0.5% and 0.31% under the truncated size of 5-bit. The proposed method can be applied for neural network accelerators especially for those which support bit-serial processing.","PeriodicalId":113802,"journal":{"name":"2019 International SoC Design Conference (ISOCC)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International SoC Design Conference (ISOCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISOCC47750.2019.9078506","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
This paper proposes a neural network training scheme in order to obtain the network weights represented in the fixed-point number format such that under the different truncated lengths of the weights, our neural new network can all achieve near-optimized inference accuracy at the corresponding word-length. The similar idea has been explored before; however, the salient feature of our proposed scaling bit-progressive method is we have further taken into account the use and training of weight scaling factor, which can significant improve the inference accuracy. Our experimental results show that our trained Resnet- 18 neural network can improve the top-1 and top-5 accuracies of Tiny-ImageNet dataset by the average of 11.02% and 9.21% compared with the previous work without using the scaling factor. The top-1 and top-5 accuracy losses compared with float-point weights are only about 0.5% and 0.31% under the truncated size of 5-bit. The proposed method can be applied for neural network accelerators especially for those which support bit-serial processing.