{"title":"An Efficient Model Compression Method of Pruning for Object Detection","authors":"Junjie Yin, Li Wei, Ding Pu, Q. Miao","doi":"10.1145/3404555.3404612","DOIUrl":null,"url":null,"abstract":"In this paper, we propose an efficient model compression method for object detection network. The key to this method is that we combine pruning and training into a single process. This design benefits in two aspects. First, we have a full control on pruning of convolution kernel, which ensures the model's accuracy to maximum extent. Second, compared with previous works, we overlap pruning with the training process instead of waiting for the model to be trained before pruning. In such a way, we can directly get a compressed model that is ready to use once training finished. We took experiments based on SSD(Single Shot MultiBox Detector) for verification. Firstly, when compressing the ssd300 model with dataset of Pascal VOC, we got model compression of 7.7X while the model accuracy only drops by 1.8%. Then on the COCO dataset, under the premise that the accuracy of the model remains unchanged, we got the model compressed by 2.8X.","PeriodicalId":220526,"journal":{"name":"Proceedings of the 2020 6th International Conference on Computing and Artificial Intelligence","volume":"28 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2020 6th International Conference on Computing and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3404555.3404612","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we propose an efficient model compression method for object detection network. The key to this method is that we combine pruning and training into a single process. This design benefits in two aspects. First, we have a full control on pruning of convolution kernel, which ensures the model's accuracy to maximum extent. Second, compared with previous works, we overlap pruning with the training process instead of waiting for the model to be trained before pruning. In such a way, we can directly get a compressed model that is ready to use once training finished. We took experiments based on SSD(Single Shot MultiBox Detector) for verification. Firstly, when compressing the ssd300 model with dataset of Pascal VOC, we got model compression of 7.7X while the model accuracy only drops by 1.8%. Then on the COCO dataset, under the premise that the accuracy of the model remains unchanged, we got the model compressed by 2.8X.