{"title":"一种基于强化学习的神经网络结构冗余自动识别框架","authors":"Tingting Wu, Chunhe Song, Peng Zeng","doi":"10.1117/12.2668217","DOIUrl":null,"url":null,"abstract":"The increasing structure of neural networks makes it difficult to deploy on edge devices with limited computing resources. Network pruning has become one of the most successful model compression methods in recent years. Existing works usually compress models by removing unimportant filters based on the importance. However, the importance-based algorithms tend to ignore the parameters that extract edge features with small criterion values. And recent studies have shown that the existing criteria rely on norm and lead to similar model compression structures. Aiming at the problems of ignoring edge features and manually specifying the pruning rate in current importance-based model pruning algorithms, this paper proposes an automatic recognition framework for neural network structure redundancy based on reinforcement learning. First, we perform cluster analysis on the filters of each layer, and map the filters into a multi-dimensional space to generate similar sets with different functions. We then propose a criterion for identifying redundant filters within similar sets. Finally, we use reinforcement learning to automatically optimize the cluster dimension, and then determine the optimal pruning rate for each layer to reduce the performance loss caused by pruning. Extensive experiments on various benchmark network architectures and datasets demonstrate the effectiveness of our proposed framework.","PeriodicalId":345723,"journal":{"name":"Fifth International Conference on Computer Information Science and Artificial Intelligence","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A framework for automatic identification of neural network structural redundancy based on reinforcement learning\",\"authors\":\"Tingting Wu, Chunhe Song, Peng Zeng\",\"doi\":\"10.1117/12.2668217\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The increasing structure of neural networks makes it difficult to deploy on edge devices with limited computing resources. Network pruning has become one of the most successful model compression methods in recent years. Existing works usually compress models by removing unimportant filters based on the importance. However, the importance-based algorithms tend to ignore the parameters that extract edge features with small criterion values. And recent studies have shown that the existing criteria rely on norm and lead to similar model compression structures. Aiming at the problems of ignoring edge features and manually specifying the pruning rate in current importance-based model pruning algorithms, this paper proposes an automatic recognition framework for neural network structure redundancy based on reinforcement learning. First, we perform cluster analysis on the filters of each layer, and map the filters into a multi-dimensional space to generate similar sets with different functions. We then propose a criterion for identifying redundant filters within similar sets. Finally, we use reinforcement learning to automatically optimize the cluster dimension, and then determine the optimal pruning rate for each layer to reduce the performance loss caused by pruning. Extensive experiments on various benchmark network architectures and datasets demonstrate the effectiveness of our proposed framework.\",\"PeriodicalId\":345723,\"journal\":{\"name\":\"Fifth International Conference on Computer Information Science and Artificial Intelligence\",\"volume\":\"44 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Fifth International Conference on Computer Information Science and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.2668217\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Fifth International Conference on Computer Information Science and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2668217","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A framework for automatic identification of neural network structural redundancy based on reinforcement learning
The increasing structure of neural networks makes it difficult to deploy on edge devices with limited computing resources. Network pruning has become one of the most successful model compression methods in recent years. Existing works usually compress models by removing unimportant filters based on the importance. However, the importance-based algorithms tend to ignore the parameters that extract edge features with small criterion values. And recent studies have shown that the existing criteria rely on norm and lead to similar model compression structures. Aiming at the problems of ignoring edge features and manually specifying the pruning rate in current importance-based model pruning algorithms, this paper proposes an automatic recognition framework for neural network structure redundancy based on reinforcement learning. First, we perform cluster analysis on the filters of each layer, and map the filters into a multi-dimensional space to generate similar sets with different functions. We then propose a criterion for identifying redundant filters within similar sets. Finally, we use reinforcement learning to automatically optimize the cluster dimension, and then determine the optimal pruning rate for each layer to reduce the performance loss caused by pruning. Extensive experiments on various benchmark network architectures and datasets demonstrate the effectiveness of our proposed framework.