Efficient deep convolutional model compression with an active stepwise pruning approach

Sheng-sheng Wang, Chunshang Xing, Dong Liu
{"title":"Efficient deep convolutional model compression with an active stepwise pruning approach","authors":"Sheng-sheng Wang, Chunshang Xing, Dong Liu","doi":"10.1504/ijcse.2020.10031600","DOIUrl":null,"url":null,"abstract":"Deep models are structurally tremendous and complex, thus making it hard to deploy on the embedded hardware with restricted memory and computing power. Although, the existing compression methods have pruned the deep models effectively, some issues exist in those methods, such as multiple iterations needed in fine-tuning phase, difficulty in pruning granularity control and numerous hyperparameters needed to set. In this paper, we propose an active stepwise pruning method of a logarithmic function which only needs to set three hyperparameters and a few epochs. We also propose a recovery strategy to repair the incorrect pruning thus ensuring the prediction accuracy of model. Pruning and repairing alternately constitute cyclic process along with updating the weights in layers. Our method can prune the parameters of MobileNet, AlexNet, VGG-16 and ZFNet by a factor of 5.6×, 11.7×, 16.6× and 15× respectively without any accuracy loss, which precedes the existing methods.","PeriodicalId":340410,"journal":{"name":"Int. J. Comput. Sci. Eng.","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Int. J. Comput. Sci. Eng.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1504/ijcse.2020.10031600","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Deep models are structurally tremendous and complex, thus making it hard to deploy on the embedded hardware with restricted memory and computing power. Although, the existing compression methods have pruned the deep models effectively, some issues exist in those methods, such as multiple iterations needed in fine-tuning phase, difficulty in pruning granularity control and numerous hyperparameters needed to set. In this paper, we propose an active stepwise pruning method of a logarithmic function which only needs to set three hyperparameters and a few epochs. We also propose a recovery strategy to repair the incorrect pruning thus ensuring the prediction accuracy of model. Pruning and repairing alternately constitute cyclic process along with updating the weights in layers. Our method can prune the parameters of MobileNet, AlexNet, VGG-16 and ZFNet by a factor of 5.6×, 11.7×, 16.6× and 15× respectively without any accuracy loss, which precedes the existing methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
有效的深度卷积模型压缩与主动逐步修剪方法
深度模型结构庞大且复杂,难以在内存和计算能力有限的嵌入式硬件上部署。现有的压缩方法虽然对深度模型进行了有效的剪枝,但存在微调阶段需要多次迭代、剪枝粒度控制困难以及需要设置大量超参数等问题。本文提出了一种对数函数的主动逐步剪枝方法,该方法只需要设置三个超参数和几个epoch。我们还提出了一种修复错误剪枝的恢复策略,从而保证了模型的预测精度。随着层间权值的更新,剪枝和修复交替构成循环过程。我们的方法可以对MobileNet、AlexNet、VGG-16和ZFNet的参数分别进行5.6倍、11.7倍、16.6倍和15倍的裁剪,而没有任何精度损失,优于现有方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
ECC-based lightweight mutual authentication protocol for fog enabled IoT system using three-way authentication procedure Gene selection and classification combining information gain ratio with fruit fly optimisation algorithm for single-cell RNA-seq data Attitude control of an unmanned patrol helicopter based on an optimised spiking neural membrane system for use in coal mines CEMP-IR: a novel location aware cache invalidation and replacement policy Prediction of consumer preference for the bottom of the pyramid using EEG-based deep model
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1