智能体协同进化中的模糊编码函数逼近

L. Tokarchuk, J. Bigham, L. Cuthbert
{"title":"智能体协同进化中的模糊编码函数逼近","authors":"L. Tokarchuk, J. Bigham, L. Cuthbert","doi":"10.5555/1166890.1166950","DOIUrl":null,"url":null,"abstract":"Reinforcement learning (RL) is a machine learning technique for sequential decision making. This approach is well proven in many small-scale domains. The true potential of this technique cannot be fully realised until it can adequately deal with the large domain sizes that typically describe real world problems. RL with function approximation is one method of dealing with the domain size problem. This paper investigates two different function approximation approaches to RL: Fuzzy Sarsa and gradient descent Sarsa(λ) with tile coding. It presents detailed experiments in two different simulation environments on the effectiveness of the two approaches. Initial experiments indicated that the tile coding approach had greater modelling capabilities in both testbed domains. However, experimentation in a coevolutionary scenario has indicated that Fuzzy Sarsa has greater flexibility.","PeriodicalId":91205,"journal":{"name":"Artificial intelligence and applications (Commerce, Calif.)","volume":"66 1","pages":"353-358"},"PeriodicalIF":0.0000,"publicationDate":"2006-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Fuzzy and Tile Coding Function Approximation in Agent Coevolution\",\"authors\":\"L. Tokarchuk, J. Bigham, L. Cuthbert\",\"doi\":\"10.5555/1166890.1166950\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Reinforcement learning (RL) is a machine learning technique for sequential decision making. This approach is well proven in many small-scale domains. The true potential of this technique cannot be fully realised until it can adequately deal with the large domain sizes that typically describe real world problems. RL with function approximation is one method of dealing with the domain size problem. This paper investigates two different function approximation approaches to RL: Fuzzy Sarsa and gradient descent Sarsa(λ) with tile coding. It presents detailed experiments in two different simulation environments on the effectiveness of the two approaches. Initial experiments indicated that the tile coding approach had greater modelling capabilities in both testbed domains. However, experimentation in a coevolutionary scenario has indicated that Fuzzy Sarsa has greater flexibility.\",\"PeriodicalId\":91205,\"journal\":{\"name\":\"Artificial intelligence and applications (Commerce, Calif.)\",\"volume\":\"66 1\",\"pages\":\"353-358\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2006-02-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Artificial intelligence and applications (Commerce, Calif.)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.5555/1166890.1166950\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial intelligence and applications (Commerce, Calif.)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5555/1166890.1166950","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

强化学习(RL)是一种用于顺序决策的机器学习技术。这种方法在许多小规模领域得到了很好的证明。这种技术的真正潜力不能完全实现,直到它能够充分处理通常描述现实世界问题的大域尺寸。函数逼近强化学习是处理域大小问题的一种方法。本文研究了两种不同的RL函数逼近方法:模糊Sarsa和梯度下降Sarsa(λ)。在两种不同的仿真环境中对两种方法的有效性进行了详细的实验。最初的实验表明,tile编码方法在两个测试平台域中都具有更大的建模能力。然而,在共同进化情景下的实验表明,Fuzzy Sarsa具有更大的灵活性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Fuzzy and Tile Coding Function Approximation in Agent Coevolution
Reinforcement learning (RL) is a machine learning technique for sequential decision making. This approach is well proven in many small-scale domains. The true potential of this technique cannot be fully realised until it can adequately deal with the large domain sizes that typically describe real world problems. RL with function approximation is one method of dealing with the domain size problem. This paper investigates two different function approximation approaches to RL: Fuzzy Sarsa and gradient descent Sarsa(λ) with tile coding. It presents detailed experiments in two different simulation environments on the effectiveness of the two approaches. Initial experiments indicated that the tile coding approach had greater modelling capabilities in both testbed domains. However, experimentation in a coevolutionary scenario has indicated that Fuzzy Sarsa has greater flexibility.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Methodology of Measurement Intellectualization based on Regularized Bayesian Approach in Uncertain Conditions Stochastic Dual Coordinate Ascent for Learning Sign Constrained Linear Predictors Data Smoothing Filling Method based on ScRNA-Seq Data Zero-Value Identification Batch-Stochastic Sub-Gradient Method for Solving Non-Smooth Convex Loss Function Problems Teaching Reading Skills More Effectively
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1