{"title":"A Study on the Control of Nonlinear System Using Growing RBFN and Reinforcement Learning","authors":"Hyun-Seob Cho","doi":"10.1109/ICNC.2007.151","DOIUrl":null,"url":null,"abstract":"The proposed approach is neural-network based and combines the self-tuning principle with reinforcement learning. The proposed control scheme consists of a controller, a utility estimator, an exploration module, a learning module and a rewarding module. The controller and the utility estimator are implemented together in a single radial basis function network (RBFN). The learning method involves structural adaptation (growing RBFN) and parameter adaptation. No prior knowledge of the plant is assumed, and the controller has to begin with exploration of the state space. The exploration versus exploitation dilemma of reinforcement learning is solved through smooth transitions between the two modes. The controller is capable of asymptotically approaching the desired reference trajectory, which is showed in simulation result.","PeriodicalId":250881,"journal":{"name":"Third International Conference on Natural Computation (ICNC 2007)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Third International Conference on Natural Computation (ICNC 2007)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICNC.2007.151","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
The proposed approach is neural-network based and combines the self-tuning principle with reinforcement learning. The proposed control scheme consists of a controller, a utility estimator, an exploration module, a learning module and a rewarding module. The controller and the utility estimator are implemented together in a single radial basis function network (RBFN). The learning method involves structural adaptation (growing RBFN) and parameter adaptation. No prior knowledge of the plant is assumed, and the controller has to begin with exploration of the state space. The exploration versus exploitation dilemma of reinforcement learning is solved through smooth transitions between the two modes. The controller is capable of asymptotically approaching the desired reference trajectory, which is showed in simulation result.