{"title":"一种具有不确定性的离散线性系统的Actor-Critic强化学习控制方法","authors":"Hsin-Chang Chen, Yu‐Chen Lin, Yu-Heng Chang","doi":"10.1109/CACS.2018.8606740","DOIUrl":null,"url":null,"abstract":"This paper is concerned with an adaptive optimal controller based an actor-critic architecture for solving discrete-time linear system with uncertainty. The actor-critic reinforcement learning progress is similar to the produce of dopamine in human brain and the mechanism which acts on the motoneuron, which dopamine enhances specific actions by reinforce the synaptic contact of the frontal lobe. As same as artificial intelligence (AI), it means the reward signal of dopamine in the neural network can be used to adjust weights in artificial neural which makes the system find the right way to solve the work. The actor-critic scheme is applied to solve the dynamic programming equation problem, using actor and critic neural networks (NNs) for solving optimal controller and optimal value function, respectively. The weights of actor and critic NNs are updated using policy gradient and recursive least squares temporal-difference learning (RLS-TD) scheme at each sampling instant. Finally, time and frequency domain simulations performed using a typical quarter-car suspension systems that an active suspension systems with the proposed control strategy is able to improve ride comfort significantly, compared with the conventional passive suspension systems.","PeriodicalId":282633,"journal":{"name":"2018 International Automatic Control Conference (CACS)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"An Actor-Critic Reinforcement Learning Control Approach for Discrete-Time Linear System with Uncertainty\",\"authors\":\"Hsin-Chang Chen, Yu‐Chen Lin, Yu-Heng Chang\",\"doi\":\"10.1109/CACS.2018.8606740\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper is concerned with an adaptive optimal controller based an actor-critic architecture for solving discrete-time linear system with uncertainty. The actor-critic reinforcement learning progress is similar to the produce of dopamine in human brain and the mechanism which acts on the motoneuron, which dopamine enhances specific actions by reinforce the synaptic contact of the frontal lobe. As same as artificial intelligence (AI), it means the reward signal of dopamine in the neural network can be used to adjust weights in artificial neural which makes the system find the right way to solve the work. The actor-critic scheme is applied to solve the dynamic programming equation problem, using actor and critic neural networks (NNs) for solving optimal controller and optimal value function, respectively. The weights of actor and critic NNs are updated using policy gradient and recursive least squares temporal-difference learning (RLS-TD) scheme at each sampling instant. Finally, time and frequency domain simulations performed using a typical quarter-car suspension systems that an active suspension systems with the proposed control strategy is able to improve ride comfort significantly, compared with the conventional passive suspension systems.\",\"PeriodicalId\":282633,\"journal\":{\"name\":\"2018 International Automatic Control Conference (CACS)\",\"volume\":\"54 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 International Automatic Control Conference (CACS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CACS.2018.8606740\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 International Automatic Control Conference (CACS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CACS.2018.8606740","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Actor-Critic Reinforcement Learning Control Approach for Discrete-Time Linear System with Uncertainty
This paper is concerned with an adaptive optimal controller based an actor-critic architecture for solving discrete-time linear system with uncertainty. The actor-critic reinforcement learning progress is similar to the produce of dopamine in human brain and the mechanism which acts on the motoneuron, which dopamine enhances specific actions by reinforce the synaptic contact of the frontal lobe. As same as artificial intelligence (AI), it means the reward signal of dopamine in the neural network can be used to adjust weights in artificial neural which makes the system find the right way to solve the work. The actor-critic scheme is applied to solve the dynamic programming equation problem, using actor and critic neural networks (NNs) for solving optimal controller and optimal value function, respectively. The weights of actor and critic NNs are updated using policy gradient and recursive least squares temporal-difference learning (RLS-TD) scheme at each sampling instant. Finally, time and frequency domain simulations performed using a typical quarter-car suspension systems that an active suspension systems with the proposed control strategy is able to improve ride comfort significantly, compared with the conventional passive suspension systems.