N. Kamiyama, N. Iijima, A. Taguchi, H. Mitsui, Y. Yoshida, M. Sone
{"title":"Tuning of learning rate and momentum on back-propagation","authors":"N. Kamiyama, N. Iijima, A. Taguchi, H. Mitsui, Y. Yoshida, M. Sone","doi":"10.1109/ICCS.1992.254895","DOIUrl":null,"url":null,"abstract":"It is known well that backpropagation is used in recognition and learning on neural networks. The backpropagation, modification of the weight is calculated by learning rate ( eta =0.2) and momentum ( alpha =0.9). The number of training cycles depends on eta and alpha , so that it is necessary to choose the most suitable values for eta and alpha . Then, changing eta and alpha , the authors tried to search for the most suitable values for the learning. As a result, the combination of eta and alpha given the minimum value of the number of training cycles behave under the constant rule. Thus eta =K(1- alpha ). Moreover, the constant K is decided by the ratio between the number of output units and hidden units or the initialized weight.<<ETX>>","PeriodicalId":223769,"journal":{"name":"[Proceedings] Singapore ICCS/ISITA `92","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1992-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"[Proceedings] Singapore ICCS/ISITA `92","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCS.1992.254895","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
It is known well that backpropagation is used in recognition and learning on neural networks. The backpropagation, modification of the weight is calculated by learning rate ( eta =0.2) and momentum ( alpha =0.9). The number of training cycles depends on eta and alpha , so that it is necessary to choose the most suitable values for eta and alpha . Then, changing eta and alpha , the authors tried to search for the most suitable values for the learning. As a result, the combination of eta and alpha given the minimum value of the number of training cycles behave under the constant rule. Thus eta =K(1- alpha ). Moreover, the constant K is decided by the ratio between the number of output units and hidden units or the initialized weight.<>