{"title":"Accelerating the training of feedforward neural networks using generalized Hebbian rules for initializing the internal representations","authors":"N. Karayiannis","doi":"10.1109/ICNN.1994.374134","DOIUrl":null,"url":null,"abstract":"It is argued in this paper that most of the problems associated with the application of existing learning algorithms in complex training tasks can be overcome by using only the input data to determine the role of the hidden units, which form a data compression or a data expansion layer. The initial set of internal representations can be formed through an unsupervised learning process applied before the supervised training algorithm. The synaptic weights that connect the input of the network with the hidden units can be determined through various linear or nonlinear variations of a generalized Hebbian learning rule, known as the Oja's rule. Several experiments indicated that the use of the proposed initialization of the internal representations improves significantly the convergence of various gradient-descent-based algorithms used to perform nontrivial training tasks.<<ETX>>","PeriodicalId":209128,"journal":{"name":"Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1996-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"25","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of 1994 IEEE International Conference on Neural Networks (ICNN'94)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNN.1994.374134","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 25
Abstract
It is argued in this paper that most of the problems associated with the application of existing learning algorithms in complex training tasks can be overcome by using only the input data to determine the role of the hidden units, which form a data compression or a data expansion layer. The initial set of internal representations can be formed through an unsupervised learning process applied before the supervised training algorithm. The synaptic weights that connect the input of the network with the hidden units can be determined through various linear or nonlinear variations of a generalized Hebbian learning rule, known as the Oja's rule. Several experiments indicated that the use of the proposed initialization of the internal representations improves significantly the convergence of various gradient-descent-based algorithms used to perform nontrivial training tasks.<>