Grouped Pointwise Convolutions Reduce Parameters in Convolutional Neural Networks

Mendel Pub Date : 2022-06-30 DOI:10.13164/mendel.2022.1.023
Joao Paulo Schwarz Schuler, S. Romaní, M. Abdel-Nasser, Hatem A. Rashwan, D. Puig
{"title":"Grouped Pointwise Convolutions Reduce Parameters in Convolutional Neural Networks","authors":"Joao Paulo Schwarz Schuler, S. Romaní, M. Abdel-Nasser, Hatem A. Rashwan, D. Puig","doi":"10.13164/mendel.2022.1.023","DOIUrl":null,"url":null,"abstract":"In DCNNs, the number of parameters in pointwise convolutions rapidly grows due to the multiplication of the number of filters by the number of input channels that come from the previous layer. Our proposal makes pointwise convolutions parameter efficient via grouping filters into parallel branches or groups, where each branch processes a fraction of the input channels. However, by doing so, the learning capability of the DCNN is degraded. To avoid this effect, we suggest interleaving the output of filters from different branches at intermediate layers of consecutive pointwise convolutions. We applied our improvement to the EfficientNet, DenseNet-BC L100, MobileNet and MobileNet V3 Large architectures. We trained these architectures with the CIFAR-10, CIFAR-100, Cropped-PlantDoc and The Oxford-IIIT Pet datasets. When training from scratch, we obtained similar test accuracies to the original EfficientNet and MobileNet V3 Large architectures while saving up to 90% of the parameters and 63% of the flops.","PeriodicalId":38293,"journal":{"name":"Mendel","volume":"1 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Mendel","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.13164/mendel.2022.1.023","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 12

Abstract

In DCNNs, the number of parameters in pointwise convolutions rapidly grows due to the multiplication of the number of filters by the number of input channels that come from the previous layer. Our proposal makes pointwise convolutions parameter efficient via grouping filters into parallel branches or groups, where each branch processes a fraction of the input channels. However, by doing so, the learning capability of the DCNN is degraded. To avoid this effect, we suggest interleaving the output of filters from different branches at intermediate layers of consecutive pointwise convolutions. We applied our improvement to the EfficientNet, DenseNet-BC L100, MobileNet and MobileNet V3 Large architectures. We trained these architectures with the CIFAR-10, CIFAR-100, Cropped-PlantDoc and The Oxford-IIIT Pet datasets. When training from scratch, we obtained similar test accuracies to the original EfficientNet and MobileNet V3 Large architectures while saving up to 90% of the parameters and 63% of the flops.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
分组点卷积减少卷积神经网络中的参数
在DCNNs中,由于滤波器数量乘以来自前一层的输入通道数量,点向卷积中的参数数量迅速增长。我们的建议通过将滤波器分组到并行分支或组来提高点卷积参数的效率,其中每个分支处理一小部分输入通道。然而,这样做会降低DCNN的学习能力。为了避免这种影响,我们建议在连续点向卷积的中间层中,将来自不同分支的滤波器的输出相互交错。我们将我们的改进应用到EfficientNet、DenseNet-BC L100、MobileNet和MobileNet V3 Large架构上。我们使用CIFAR-10、CIFAR-100、croped - plantdoc和the Oxford-IIIT Pet数据集训练这些架构。当从头开始训练时,我们获得了与原始的EfficientNet和MobileNet V3 Large架构相似的测试精度,同时节省了高达90%的参数和63%的失败。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Mendel
Mendel Decision Sciences-Decision Sciences (miscellaneous)
CiteScore
2.20
自引率
0.00%
发文量
7
期刊最新文献
Detecting Outliers Using Modified Recursive PCA Algorithm For Dynamic Streaming Data Stock and Structured Warrant Portfolio Optimization Using Black-Litterman Model and Binomial Method Optimized Fixed-Time Synergetic Controller via a modified Salp Swarm Algorithm for Acute and Chronic HBV Transmission System Initial Coin Offering Prediction Comparison Using Ridge Regression, Artificial Neural Network, Random Forest Regression, and Hybrid ANN-Ridge Predicting Football Match Outcomes with Machine Learning Approaches
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1