A low functional redundancy-based network slimming method for accelerating deep neural networks

IF 6.8 2区 工程技术 Q1 ENGINEERING, MULTIDISCIPLINARY alexandria engineering journal Pub Date : 2025-04-01 Epub Date: 2025-02-07 DOI:10.1016/j.aej.2024.12.118
Zheng Fang , Bo Yin
{"title":"A low functional redundancy-based network slimming method for accelerating deep neural networks","authors":"Zheng Fang ,&nbsp;Bo Yin","doi":"10.1016/j.aej.2024.12.118","DOIUrl":null,"url":null,"abstract":"<div><div>Deep neural networks (DNNs) have been widely criticized for their large parameters and computation demands, hindering deployment to edge and embedded devices. In order to reduce the floating point operations (FLOPs) running DNNs and accelerate the inference speed, we start from the model pruning, and realize this goal by removing useless network parameters. In this research, we propose a low functional redundancy-based network slimming method (LFRNS) that can find and remove functional redundant filters by feature clustering algorithm. However, the redundancy of some key features is beneficial to the model, and removing these features will limit the potential of the model to some extent. Build on this view, we propose feature contribution ranking unit (FCR unit) which can automatically learn the feature maps' contribution to the key information with training iterations. FCR unit can assist LFRNS restore some important elements in the pruning set to break the performance bottleneck of the slimming model. Our method mainly removes feature maps with similar functions instead of only pruning the unimportant parts, thus effectively ensuring the integrity of features’ functions and avoiding network degradation. We conduct experiments on image classification task based on CIFAR-10 and CIFAR-100 datasets. Our framework achieves over 2.0 × parameters and FLOPs reductions, while maintaining &lt; 1 % loss in accuracy, and even improve accuracy of large-volume models. We also introduce our method to the vision transformer model (ViT) and achieve performance comparable to state-of-the-art methods with nearly 1.5 × less computation.</div></div>","PeriodicalId":7484,"journal":{"name":"alexandria engineering journal","volume":"119 ","pages":"Pages 437-450"},"PeriodicalIF":6.8000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"alexandria engineering journal","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1110016824017162","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/7 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

Deep neural networks (DNNs) have been widely criticized for their large parameters and computation demands, hindering deployment to edge and embedded devices. In order to reduce the floating point operations (FLOPs) running DNNs and accelerate the inference speed, we start from the model pruning, and realize this goal by removing useless network parameters. In this research, we propose a low functional redundancy-based network slimming method (LFRNS) that can find and remove functional redundant filters by feature clustering algorithm. However, the redundancy of some key features is beneficial to the model, and removing these features will limit the potential of the model to some extent. Build on this view, we propose feature contribution ranking unit (FCR unit) which can automatically learn the feature maps' contribution to the key information with training iterations. FCR unit can assist LFRNS restore some important elements in the pruning set to break the performance bottleneck of the slimming model. Our method mainly removes feature maps with similar functions instead of only pruning the unimportant parts, thus effectively ensuring the integrity of features’ functions and avoiding network degradation. We conduct experiments on image classification task based on CIFAR-10 and CIFAR-100 datasets. Our framework achieves over 2.0 × parameters and FLOPs reductions, while maintaining < 1 % loss in accuracy, and even improve accuracy of large-volume models. We also introduce our method to the vision transformer model (ViT) and achieve performance comparable to state-of-the-art methods with nearly 1.5 × less computation.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
一种基于低功能冗余的加速深度神经网络瘦身方法
深度神经网络(dnn)因其大参数和计算需求而受到广泛批评,阻碍了边缘和嵌入式设备的部署。为了减少dnn运行的浮点运算,加快推理速度,我们从模型修剪开始,通过去除无用的网络参数来实现这一目标。在本研究中,我们提出了一种基于低功能冗余的网络瘦身方法(LFRNS),该方法可以通过特征聚类算法发现和去除功能冗余过滤器。然而,一些关键特征的冗余对模型是有益的,删除这些特征会在一定程度上限制模型的潜力。在此基础上,我们提出了特征贡献排序单元(FCR unit),该单元可以通过训练迭代自动学习特征映射对关键信息的贡献。FCR单元可以帮助LFRNS恢复剪枝集中的一些重要元素,打破剪枝模型的性能瓶颈。我们的方法主要是去除功能相似的特征映射,而不是只修剪不重要的部分,从而有效地保证了特征功能的完整性,避免了网络退化。我们基于CIFAR-10和CIFAR-100数据集对图像分类任务进行了实验。我们的框架实现了2.0 × 以上的参数和FLOPs降低,同时保持了<; 1 %的精度损失,甚至提高了大容量模型的精度。我们还将我们的方法引入视觉变压器模型(ViT),并实现了与最先进的方法相当的性能,减少了近1.5 × 的计算量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
alexandria engineering journal
alexandria engineering journal Engineering-General Engineering
CiteScore
11.20
自引率
4.40%
发文量
1015
审稿时长
43 days
期刊介绍: Alexandria Engineering Journal is an international journal devoted to publishing high quality papers in the field of engineering and applied science. Alexandria Engineering Journal is cited in the Engineering Information Services (EIS) and the Chemical Abstracts (CA). The papers published in Alexandria Engineering Journal are grouped into five sections, according to the following classification: • Mechanical, Production, Marine and Textile Engineering • Electrical Engineering, Computer Science and Nuclear Engineering • Civil and Architecture Engineering • Chemical Engineering and Applied Sciences • Environmental Engineering
期刊最新文献
Assessing public transport equity: The case of Alexandria, Egypt Design and performance assessment of a high efficiency facade-integrated ventilation unit with membrane-based enthalpy exchanger New insights for enhancing the intelligence of coal mine: A two-stage method for unsupervised low-light image enhancement and lightweight detection A siamese vision transformer-based model for automatic music emotion annotation and classification SCH-Net: A ViT-ResNet hybrid network with STERN module for automatic classification of thoracic diseases on clinical chest X-rays
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1