利用模糊规则将专家知识融入q学习

M. Pourhassan, N. Mozayani
{"title":"利用模糊规则将专家知识融入q学习","authors":"M. Pourhassan, N. Mozayani","doi":"10.1109/CIMSA.2009.5069952","DOIUrl":null,"url":null,"abstract":"Incorporating expert knowledge in reinforcement learning is an important issue, especially when a large state space is concerned. In this paper, we present a novel method for accelerating the setting of Q-values in the well-known Q-learning algorithm. Fuzzy rules indicating the state values will be used, and the knowledge will be transformed to the Q-table or Q-function in some first training experiences. There have already been methods to initialize the Q-values using fuzzy rules, but the rules were the kind of state-action rules and needed the expert to know about environment transitions on actions. In the method introduced in this paper, the expert should only apply some rules to estimate the state value while no appreciations about state transitions are required. The introduced method has been examined in a multiagent system which has the shepherding scenario. The obtaining results show that Q-learning requires much less iterations for getting good results if using the fuzzy rules estimating the state value.","PeriodicalId":178669,"journal":{"name":"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2009-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Incorporating expert knowledge in Q-learning by means of fuzzy rules\",\"authors\":\"M. Pourhassan, N. Mozayani\",\"doi\":\"10.1109/CIMSA.2009.5069952\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Incorporating expert knowledge in reinforcement learning is an important issue, especially when a large state space is concerned. In this paper, we present a novel method for accelerating the setting of Q-values in the well-known Q-learning algorithm. Fuzzy rules indicating the state values will be used, and the knowledge will be transformed to the Q-table or Q-function in some first training experiences. There have already been methods to initialize the Q-values using fuzzy rules, but the rules were the kind of state-action rules and needed the expert to know about environment transitions on actions. In the method introduced in this paper, the expert should only apply some rules to estimate the state value while no appreciations about state transitions are required. The introduced method has been examined in a multiagent system which has the shepherding scenario. The obtaining results show that Q-learning requires much less iterations for getting good results if using the fuzzy rules estimating the state value.\",\"PeriodicalId\":178669,\"journal\":{\"name\":\"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications\",\"volume\":\"27 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2009-05-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CIMSA.2009.5069952\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2009 IEEE International Conference on Computational Intelligence for Measurement Systems and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIMSA.2009.5069952","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

摘要

在强化学习中引入专家知识是一个重要的问题,特别是当涉及到一个大的状态空间时。在本文中,我们提出了一种新的方法来加速众所周知的q -学习算法中q值的设置。将使用指示状态值的模糊规则,并将知识转换为一些首次训练经验中的q表或q函数。已经有使用模糊规则初始化q值的方法,但这些规则是一种状态-动作规则,需要专家了解动作的环境转换。在本文所介绍的方法中,专家只需要应用一些规则来估计状态值,而不需要对状态转移进行评估。所介绍的方法已经在一个具有放牧场景的多智能体系统中进行了检验。得到的结果表明,如果使用模糊规则估计状态值,q -学习所需的迭代次数要少得多,并且得到了很好的结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Incorporating expert knowledge in Q-learning by means of fuzzy rules
Incorporating expert knowledge in reinforcement learning is an important issue, especially when a large state space is concerned. In this paper, we present a novel method for accelerating the setting of Q-values in the well-known Q-learning algorithm. Fuzzy rules indicating the state values will be used, and the knowledge will be transformed to the Q-table or Q-function in some first training experiences. There have already been methods to initialize the Q-values using fuzzy rules, but the rules were the kind of state-action rules and needed the expert to know about environment transitions on actions. In the method introduced in this paper, the expert should only apply some rules to estimate the state value while no appreciations about state transitions are required. The introduced method has been examined in a multiagent system which has the shepherding scenario. The obtaining results show that Q-learning requires much less iterations for getting good results if using the fuzzy rules estimating the state value.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
An intelligent multi-agent distributed battlefield via Multi-Token Message Passing Research on Supervised Manifold Learning for SAR target classification Fuzzy control system of constant current for spot welding inverter Deviation recognition of high speed rotational arc sensor based on support vector machine Research of improved immune clonal algorithms and its applications
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1