{"title":"A Theoretical CNN Compression Framework for Resource-Restricted Environments","authors":"Zahra Waheed Awan, S. Khalid, Sajid Gul","doi":"10.1109/ICoDT255437.2022.9787431","DOIUrl":null,"url":null,"abstract":"Convolutional Neural Network (CNN) is considered as one of the most significant algorithms of deep learning that has made impressive achievements in many areas of computer vision and natural language processing. In the current times of big data, input data dimensions keep on increasing which leads to the development of complex CNN models for processing such big data. This has made CNN computationally intensive and limits its practical application to some extent. To address the aforementioned issue, this paper presents a detailed review of various network compression methods existing in literature. Two most commonly deployed network compression methods have been discussed including pruning and quantization which can be coupled with CNN to increase its applicability. The main goal of presenting this comprehensive review of the state-of-the-art pruning and quantization-based network compression schemes is to significantly improve trade-off between CNN architectural size and computational cost versus its performance in resource restricted environments. Additionally, this paper also exploits the challenges posed by these techniques when implemented for large-scale CNNs. In this context, paper also presents a novel framework to perform network compression of CNN to meet the requirements of resource-restricted devices.","PeriodicalId":291030,"journal":{"name":"2022 2nd International Conference on Digital Futures and Transformative Technologies (ICoDT2)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 2nd International Conference on Digital Futures and Transformative Technologies (ICoDT2)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICoDT255437.2022.9787431","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Convolutional Neural Network (CNN) is considered as one of the most significant algorithms of deep learning that has made impressive achievements in many areas of computer vision and natural language processing. In the current times of big data, input data dimensions keep on increasing which leads to the development of complex CNN models for processing such big data. This has made CNN computationally intensive and limits its practical application to some extent. To address the aforementioned issue, this paper presents a detailed review of various network compression methods existing in literature. Two most commonly deployed network compression methods have been discussed including pruning and quantization which can be coupled with CNN to increase its applicability. The main goal of presenting this comprehensive review of the state-of-the-art pruning and quantization-based network compression schemes is to significantly improve trade-off between CNN architectural size and computational cost versus its performance in resource restricted environments. Additionally, this paper also exploits the challenges posed by these techniques when implemented for large-scale CNNs. In this context, paper also presents a novel framework to perform network compression of CNN to meet the requirements of resource-restricted devices.