{"title":"Efficient Hardware Architecture of Softmax Layer in Deep Neural Network","authors":"Ruofei Hu, Binren Tian, S. Yin, Shaojun Wei","doi":"10.1109/ICDSP.2018.8631588","DOIUrl":null,"url":null,"abstract":"Deep neural network (DNN), as a very important machine learning technique in classification and detection tasks for images, video, speech as wellas audio, has recently received tremendous attention. Integral Stochastic Computation (Integral SC), on the other hand, has proved its extraordinary ability in hardware implementation of DNNs. Thesoftmax layer is generally used in multi-classification tasks as a very basic and important network layer in DNNs. However, the hardware implementation of softmax layer is expensive cause the exponentiation and division computation. In this paper, we designed an efficient way to simulate softmax layer in DNNs based on Integral stochastic computing, filling the vacancy of previous academic works. Compared to conventional softmax hardware implementation, our method achieves reduction in power and area by 68% and 41%, respectively.","PeriodicalId":218806,"journal":{"name":"2018 IEEE 23rd International Conference on Digital Signal Processing (DSP)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE 23rd International Conference on Digital Signal Processing (DSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSP.2018.8631588","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
Deep neural network (DNN), as a very important machine learning technique in classification and detection tasks for images, video, speech as wellas audio, has recently received tremendous attention. Integral Stochastic Computation (Integral SC), on the other hand, has proved its extraordinary ability in hardware implementation of DNNs. Thesoftmax layer is generally used in multi-classification tasks as a very basic and important network layer in DNNs. However, the hardware implementation of softmax layer is expensive cause the exponentiation and division computation. In this paper, we designed an efficient way to simulate softmax layer in DNNs based on Integral stochastic computing, filling the vacancy of previous academic works. Compared to conventional softmax hardware implementation, our method achieves reduction in power and area by 68% and 41%, respectively.