Log-quantization on GRU networks

Sangki Park, Sang-Soo Park, Ki-Seok Chung
{"title":"Log-quantization on GRU networks","authors":"Sangki Park, Sang-Soo Park, Ki-Seok Chung","doi":"10.1145/3290420.3290443","DOIUrl":null,"url":null,"abstract":"Today, recurrent neural network (RNN) is used in various applications like image captioning, speech recognition and machine translation. However, because of data dependencies, recurrent neural network is hard to parallelize. Furthermore, to increase network's accuracy, recurrent neural network uses complicated cell units such as long short-term memory (LSTM) and gated recurrent unit (GRU). To run such models on an embedded system, the size of the network model and the amount of computation need to be reduced to achieve low power consumption and low required memory bandwidth. In this paper, implementation of RNN based on GRU with a logarithmic quantization method is proposed. The proposed implementation is synthesized using high-level synthesis (HLS) targeting Xilinx ZCU102 FPGA running at 100MHz. The proposed implementation with an 8-bit log-quantization achieves 90.57% accuracy without re-training or fine-tuning. And the memory usage is 31% lower than that for an implementation with 32-bit floating point data representation.","PeriodicalId":259201,"journal":{"name":"International Conference on Critical Infrastructure Protection","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Critical Infrastructure Protection","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3290420.3290443","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Today, recurrent neural network (RNN) is used in various applications like image captioning, speech recognition and machine translation. However, because of data dependencies, recurrent neural network is hard to parallelize. Furthermore, to increase network's accuracy, recurrent neural network uses complicated cell units such as long short-term memory (LSTM) and gated recurrent unit (GRU). To run such models on an embedded system, the size of the network model and the amount of computation need to be reduced to achieve low power consumption and low required memory bandwidth. In this paper, implementation of RNN based on GRU with a logarithmic quantization method is proposed. The proposed implementation is synthesized using high-level synthesis (HLS) targeting Xilinx ZCU102 FPGA running at 100MHz. The proposed implementation with an 8-bit log-quantization achieves 90.57% accuracy without re-training or fine-tuning. And the memory usage is 31% lower than that for an implementation with 32-bit floating point data representation.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
GRU网络的日志量化
如今,递归神经网络(RNN)被用于各种应用,如图像字幕、语音识别和机器翻译。然而,由于数据的依赖性,递归神经网络很难并行化。此外,为了提高网络的准确性,递归神经网络使用了长短期记忆(LSTM)和门控递归单元(GRU)等复杂的细胞单元。要在嵌入式系统上运行这样的模型,需要减少网络模型的大小和计算量,以实现低功耗和低所需的内存带宽。本文提出了一种基于GRU的对数量化方法来实现RNN。提出的实现采用针对运行在100MHz的Xilinx ZCU102 FPGA的高级合成(HLS)进行合成。采用8位日志量化的实现,无需重新训练或微调,准确率达到90.57%。与使用32位浮点数据表示的实现相比,内存使用量降低了31%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Deep learning chips: challenges and opportunities for ubiquitous power internet of things Secrecy relaying strategy over correlated fading channels using CSI estimation 3D anisotropie convolutional neural network with step transfer learning for liver segmentation Image feature fusion and its application based on trace transform and improved GLBP An improved AOR-based precoding for massive MIMO systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1