{"title":"加速卷积神经网络的研究","authors":"Hsien-I Lin, Chung-Sheng Cheng","doi":"10.1063/1.5138068","DOIUrl":null,"url":null,"abstract":"Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, respectively, in THUR15K, Caltech-101, Caltech-256, and GHIM10k datasets. The results show that the parameter amount greatly decreased (76%) but the recognition rate dropped slightly (1.34%).Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, resp...","PeriodicalId":20565,"journal":{"name":"PROCEEDINGS OF THE INTERNATIONAL CONFERENCE OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING 2019 (ICCMSE-2019)","volume":"7 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2019-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"A study on accelerating convolutional neural networks\",\"authors\":\"Hsien-I Lin, Chung-Sheng Cheng\",\"doi\":\"10.1063/1.5138068\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, respectively, in THUR15K, Caltech-101, Caltech-256, and GHIM10k datasets. The results show that the parameter amount greatly decreased (76%) but the recognition rate dropped slightly (1.34%).Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, resp...\",\"PeriodicalId\":20565,\"journal\":{\"name\":\"PROCEEDINGS OF THE INTERNATIONAL CONFERENCE OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING 2019 (ICCMSE-2019)\",\"volume\":\"7 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-12-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"PROCEEDINGS OF THE INTERNATIONAL CONFERENCE OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING 2019 (ICCMSE-2019)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1063/1.5138068\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"PROCEEDINGS OF THE INTERNATIONAL CONFERENCE OF COMPUTATIONAL METHODS IN SCIENCES AND ENGINEERING 2019 (ICCMSE-2019)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1063/1.5138068","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A study on accelerating convolutional neural networks
Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, respectively, in THUR15K, Caltech-101, Caltech-256, and GHIM10k datasets. The results show that the parameter amount greatly decreased (76%) but the recognition rate dropped slightly (1.34%).Recent deep-learning methods have been paid more attention than shallow-learning ones because they have deep and complex structures to approximate functions. The salient feature of deep neural networks is to use many layers where many of them are used to extract data features and few are for classification or regression. The most severe problem of a deep neural network is using too many parameters that cause too much memory usage and computing resources for both training and inference. Thus, deep learning approaches are not suitable for real-time industrial applications that have limited computing resources such as memory and CPU. For example, a famous convolutional neural network (CNN), AlexNet, uses up to 60 million parameters to train ImageNet dataset and many imaging projects apply AlexNet to their own applications as transfer learning. Thus, this work proposes a feasible solution to trim the CNN, speed it up, and keep the accuracy rate similar. Two main types of CNNs and AlexNet, were validated, resp...