{"title":"利用模糊规则将专家知识融入q学习","authors":"M. Pourhassan, N. Mozayani","doi":"10.1109/CIMSA.2009.5069952","DOIUrl":null,"url":null,"abstract":"Incorporating expert knowledge in reinforcement learning is an important issue, especially when a large state space is concerned. In this paper, we present a novel method for accelerating the setting of Q-values in the well-known Q-learning algorithm. Fuzzy rules indicating the state values will be used, and the knowledge will be transformed to the Q-table or Q-function in some first training experiences. There have already been methods to initialize the Q-values using fuzzy rules, but the rules were the kind of state-action rules and needed the expert to know about environment transitions on actions. In the method introduced in this paper, the expert should only apply some rules to estimate the state value while no appreciations about state transitions are required. The introduced method has been examined in a multiagent system which has the shepherding scenario. The obtaining results show that Q-learning requires much less iterations for getting good results if using the fuzzy rules estimating the state value.","PeriodicalId":178669,"journal":{"name":"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Incorporating expert knowledge in Q-learning by means of fuzzy rules\",\"authors\":\"M. Pourhassan, N. Mozayani\",\"doi\":\"10.1109/CIMSA.2009.5069952\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Incorporating expert knowledge in reinforcement learning is an important issue, especially when a large state space is concerned. In this paper, we present a novel method for accelerating the setting of Q-values in the well-known Q-learning algorithm. Fuzzy rules indicating the state values will be used, and the knowledge will be transformed to the Q-table or Q-function in some first training experiences. There have already been methods to initialize the Q-values using fuzzy rules, but the rules were the kind of state-action rules and needed the expert to know about environment transitions on actions. In the method introduced in this paper, the expert should only apply some rules to estimate the state value while no appreciations about state transitions are required. The introduced method has been examined in a multiagent system which has the shepherding scenario. The obtaining results show that Q-learning requires much less iterations for getting good results if using the fuzzy rules estimating the state value.\",\"PeriodicalId\":178669,\"journal\":{\"name\":\"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications\",\"volume\":\"27 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2009-05-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CIMSA.2009.5069952\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIMSA.2009.5069952","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Incorporating expert knowledge in Q-learning by means of fuzzy rules
Incorporating expert knowledge in reinforcement learning is an important issue, especially when a large state space is concerned. In this paper, we present a novel method for accelerating the setting of Q-values in the well-known Q-learning algorithm. Fuzzy rules indicating the state values will be used, and the knowledge will be transformed to the Q-table or Q-function in some first training experiences. There have already been methods to initialize the Q-values using fuzzy rules, but the rules were the kind of state-action rules and needed the expert to know about environment transitions on actions. In the method introduced in this paper, the expert should only apply some rules to estimate the state value while no appreciations about state transitions are required. The introduced method has been examined in a multiagent system which has the shepherding scenario. The obtaining results show that Q-learning requires much less iterations for getting good results if using the fuzzy rules estimating the state value.