{"title":"Neural Network Optimization Objective Vector Representation based on Genetic Algorithm and Its Multi-objective Optimization Method","authors":"Yunke Xiong, Qun Hou, Xin Liu","doi":"10.1145/3583788.3583796","DOIUrl":null,"url":null,"abstract":"Deep learning algorithms mostly have network parameters that can affect their training results, and the combination of neural network architectures also has a significant impact on the algorithm performance. The performance of deep learning algorithms is usually proportional to the overall number of network parameters, leading to excessive resource consumption for exploring neural network architectures with a large number of hyper-parameters. To solve this problem, a vector representation is proposed which for neural network architectures, and a multi-objective optimization model is established based on genetic algorithms in this paper, and it is short for “NNOO Vector Representation based on GA and Its Optimization Method”. The multi-objective optimization model can automatically optimize the neural network architecture and hyper-parameters in the network, improve the network accuracy, and reduce the overall number of network parameters. It is shown in the test results with the MNIST data set, and the accuracy is 95.61% for the traditional empirical setting network, and the average accuracy is 86.2% for the network optimized by TensorFlow’s optimization algorithm. While the network accuracy is improved to 96.86% with the proposed optimization method in this paper and the network parameters are reduced by 32.6% compared with the traditional empirical network, and the network parameters are reduced by13.2% compared with the network by TensorFlow’s optimization algorithm. Therefore, the method is presented which has obvious practical application value in neural network optimization problems and provides a new way of thinking for large and deep network optimization problems.","PeriodicalId":292167,"journal":{"name":"Proceedings of the 2023 7th International Conference on Machine Learning and Soft Computing","volume":"503 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2023 7th International Conference on Machine Learning and Soft Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3583788.3583796","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Deep learning algorithms mostly have network parameters that can affect their training results, and the combination of neural network architectures also has a significant impact on the algorithm performance. The performance of deep learning algorithms is usually proportional to the overall number of network parameters, leading to excessive resource consumption for exploring neural network architectures with a large number of hyper-parameters. To solve this problem, a vector representation is proposed which for neural network architectures, and a multi-objective optimization model is established based on genetic algorithms in this paper, and it is short for “NNOO Vector Representation based on GA and Its Optimization Method”. The multi-objective optimization model can automatically optimize the neural network architecture and hyper-parameters in the network, improve the network accuracy, and reduce the overall number of network parameters. It is shown in the test results with the MNIST data set, and the accuracy is 95.61% for the traditional empirical setting network, and the average accuracy is 86.2% for the network optimized by TensorFlow’s optimization algorithm. While the network accuracy is improved to 96.86% with the proposed optimization method in this paper and the network parameters are reduced by 32.6% compared with the traditional empirical network, and the network parameters are reduced by13.2% compared with the network by TensorFlow’s optimization algorithm. Therefore, the method is presented which has obvious practical application value in neural network optimization problems and provides a new way of thinking for large and deep network optimization problems.