{"title":"最大化梯度函数的有效神经网络训练","authors":"S. U. Ahmed, M. Shahjahan, K. Murase","doi":"10.1109/ICCITECHN.2010.5723895","DOIUrl":null,"url":null,"abstract":"In this paper, a faster supervised algorithm (BPfast) for the neural network training is proposed that maximizes the derivative of sigmoid activation function during back-propagation (BP) training. BP adjusts the weights of neural network with minimizing an error function. Due to the presence of derivative information in the weight update rule, BP goes to ‘premature saturation’ that slows down the training convergence. In the saturation region, the derivative information tends to zero. To overcome the problem, BPfast maximizes the derivative of activation function together with minimizing the error function. BPfast is tested on five real world benchmark problems such as breast cancer, diabetes, heart disease, Australian credit card, and horse. BPfast exhibits faster convergence and good generalization ability over standard BP algorithm.","PeriodicalId":149135,"journal":{"name":"2010 13th International Conference on Computer and Information Technology (ICCIT)","volume":"103 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Maximization of the gradient function for efficient neural network training\",\"authors\":\"S. U. Ahmed, M. Shahjahan, K. Murase\",\"doi\":\"10.1109/ICCITECHN.2010.5723895\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, a faster supervised algorithm (BPfast) for the neural network training is proposed that maximizes the derivative of sigmoid activation function during back-propagation (BP) training. BP adjusts the weights of neural network with minimizing an error function. Due to the presence of derivative information in the weight update rule, BP goes to ‘premature saturation’ that slows down the training convergence. In the saturation region, the derivative information tends to zero. To overcome the problem, BPfast maximizes the derivative of activation function together with minimizing the error function. BPfast is tested on five real world benchmark problems such as breast cancer, diabetes, heart disease, Australian credit card, and horse. BPfast exhibits faster convergence and good generalization ability over standard BP algorithm.\",\"PeriodicalId\":149135,\"journal\":{\"name\":\"2010 13th International Conference on Computer and Information Technology (ICCIT)\",\"volume\":\"103 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 13th International Conference on Computer and Information Technology (ICCIT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCITECHN.2010.5723895\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 13th International Conference on Computer and Information Technology (ICCIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCITECHN.2010.5723895","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Maximization of the gradient function for efficient neural network training
In this paper, a faster supervised algorithm (BPfast) for the neural network training is proposed that maximizes the derivative of sigmoid activation function during back-propagation (BP) training. BP adjusts the weights of neural network with minimizing an error function. Due to the presence of derivative information in the weight update rule, BP goes to ‘premature saturation’ that slows down the training convergence. In the saturation region, the derivative information tends to zero. To overcome the problem, BPfast maximizes the derivative of activation function together with minimizing the error function. BPfast is tested on five real world benchmark problems such as breast cancer, diabetes, heart disease, Australian credit card, and horse. BPfast exhibits faster convergence and good generalization ability over standard BP algorithm.