A novel reinforcement learning architecture for continuous state and action spaces

Victor Uc-Cetina
{"title":"A novel reinforcement learning architecture for continuous state and action spaces","authors":"Victor Uc-Cetina","doi":"10.1155/2013/492852","DOIUrl":null,"url":null,"abstract":"We introduce a reinforcement learning architecture designed for problems with an infinite number of states, where each state can be seen as a vector of real numbers and with a finite number of actions, where each action requires a vector of real numbers as parameters. The main objective of this architecture is to distribute in two actors the work required to learn the final policy. One actor decideswhat actionmust be performed;meanwhile, a second actor determines the right parameters for the selected action. We tested our architecture and one algorithmbased on it solving the robot dribbling problem, a challenging robot control problem taken from the RoboCup competitions. Our experimental work with three different function approximators provides enough evidence to prove that the proposed architecture can be used to implement fast, robust, and reliable reinforcement learning algorithms.","PeriodicalId":91830,"journal":{"name":"Advances in artificial intelligence. Canadian Society for Computational Studies of Intelligence. Conference","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2013-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in artificial intelligence. Canadian Society for Computational Studies of Intelligence. Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1155/2013/492852","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

We introduce a reinforcement learning architecture designed for problems with an infinite number of states, where each state can be seen as a vector of real numbers and with a finite number of actions, where each action requires a vector of real numbers as parameters. The main objective of this architecture is to distribute in two actors the work required to learn the final policy. One actor decideswhat actionmust be performed;meanwhile, a second actor determines the right parameters for the selected action. We tested our architecture and one algorithmbased on it solving the robot dribbling problem, a challenging robot control problem taken from the RoboCup competitions. Our experimental work with three different function approximators provides enough evidence to prove that the proposed architecture can be used to implement fast, robust, and reliable reinforcement learning algorithms.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
一种新的连续状态和动作空间强化学习架构
我们引入了一个为具有无限数量状态的问题设计的强化学习架构,其中每个状态可以被视为实数向量,并且具有有限数量的动作,其中每个动作需要实数向量作为参数。该体系结构的主要目标是将学习最终策略所需的工作分配给两个参与者。一个参与者决定必须执行什么操作;同时,第二个参与者确定所选操作的正确参数。我们测试了我们的架构和一个基于它的算法来解决机器人运球问题,这是一个来自机器人世界杯比赛的具有挑战性的机器人控制问题。我们用三种不同的函数逼近器进行的实验工作提供了足够的证据来证明所提出的架构可以用于实现快速、鲁棒和可靠的强化学习算法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Solving the Permutation Heijunka Flow Shop Scheduling Problem with Non-unit Demands for Jobs Speech Intention Classification with Multimodal Deep Learning. Unsupervised Extraction of Diagnosis Codes from EMRs Using Knowledge-Based and Extractive Text Summarization Techniques. A novel reinforcement learning architecture for continuous state and action spaces A Graph Cellular Automata Model to Study the Spreading of an Infectious Disease
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1