{"title":"Self-structuring hidden control neural models","authors":"H. Sørensen, U. Hartmann","doi":"10.1109/NNSP.1992.253698","DOIUrl":null,"url":null,"abstract":"The authors propose a self-structuring hidden control (SHC) neural model for pattern recognition which establishes a near-optimal architecture during training. A significant network architecture reduction in terms of the number of hidden processing elements (PEs) is typically achieved. The SHC model combines self-structuring architecture generation with nonlinear prediction and hidden Markov modelling. The authors present a theorem for self-structuring neural models stating that these models are universal approximators and thus relevant to real-world pattern recognition. Using SHC models containing as few as five hidden PEs each for an isolated word recognition task resulted in a recognition rate of 98.4%. SHC models can also be applied to continuous speech recognition.<<ETX>>","PeriodicalId":438250,"journal":{"name":"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop","volume":"48 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1992-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NNSP.1992.253698","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The authors propose a self-structuring hidden control (SHC) neural model for pattern recognition which establishes a near-optimal architecture during training. A significant network architecture reduction in terms of the number of hidden processing elements (PEs) is typically achieved. The SHC model combines self-structuring architecture generation with nonlinear prediction and hidden Markov modelling. The authors present a theorem for self-structuring neural models stating that these models are universal approximators and thus relevant to real-world pattern recognition. Using SHC models containing as few as five hidden PEs each for an isolated word recognition task resulted in a recognition rate of 98.4%. SHC models can also be applied to continuous speech recognition.<>