A study on accelerating convolutional neural networks

Hsien-I Lin, Chung-Sheng Cheng
{"title":"A study on accelerating convolutional neural networks","authors":"Hsien-I Lin, Chung-Sheng Cheng","doi":"10.1063/1.5138068","DOIUrl":null,"url":null,"abstract":"Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, respectively, in THUR15K, Caltech-101, Caltech-256, and GHIM10k datasets. The results show that the parameter amount greatly decreased (76%) but the recognition rate dropped slightly (1.34%).Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, resp...","PeriodicalId":20565,"journal":{"name":"PROCEEDINGS OF THE INTERNATIONAL CONFERENCE OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING 2019 (ICCMSE-2019)","volume":"7 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2019-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"PROCEEDINGS OF THE INTERNATIONAL CONFERENCE OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING 2019 (ICCMSE-2019)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1063/1.5138068","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, respectively, in THUR15K, Caltech-101, Caltech-256, and GHIM10k datasets. The results show that the parameter amount greatly decreased (76%) but the recognition rate dropped slightly (1.34%).Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, resp...
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
加速卷积神经网络的研究
与浅学习方法相比,深度学习方法由于具有较深和复杂的近似函数结构而受到越来越多的关注。深度神经网络的显著特点是使用多层,其中许多层用于提取数据特征,很少用于分类或回归。深度神经网络最严重的问题是使用太多的参数,这会导致过多的内存使用和计算资源用于训练和推理。因此,深度学习方法不适合内存和CPU等计算资源有限的实时工业应用。例如,著名的卷积神经网络(CNN) AlexNet使用多达6000万个参数来训练ImageNet数据集,许多成像项目将AlexNet作为迁移学习应用到自己的应用程序中。因此,本工作提出了一种可行的解决方案,以修剪CNN,加快其速度,并保持准确率相似。两种主要类型的cnn和AlexNet分别在THUR15K、Caltech-101、Caltech-256和GHIM10k数据集上进行了验证。结果表明,参数数量大大减少(76%),但识别率略有下降(1.34%)。与浅学习方法相比,深度学习方法由于具有较深和复杂的近似函数结构而受到越来越多的关注。深度神经网络的显著特点是使用多层,其中许多层用于提取数据特征,很少用于分类或回归。深度神经网络最严重的问题是使用太多的参数,这会导致过多的内存使用和计算资源用于训练和推理。因此,深度学习方法不适合内存和CPU等计算资源有限的实时工业应用。例如,著名的卷积神经网络(CNN) AlexNet使用多达6000万个参数来训练ImageNet数据集,许多成像项目将AlexNet作为迁移学习应用到自己的应用程序中。因此,本工作提出了一种可行的解决方案,以修剪CNN,加快其速度,并保持准确率相似。cnn和AlexNet的两种主要类型得到了验证,分别是…
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Selected time space characteristics in female pole vault Wave polarisation in a dynamic elastic lattice Symbolic-numeric research of leaky modes in planar dielectric electromagnetic waveguide as inhomogeneous waves Derivation of the concepts in data modelling Preface of the “GIS, Remote Sensing and Dendrochronology in Geohazards”
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1