{"title":"Training a robot with limited computing resources to crawl using reinforcement learning","authors":"Moritz P. Heimbach, J. Weber, M. Schmidt","doi":"10.1109/IRC55401.2022.00051","DOIUrl":null,"url":null,"abstract":"In recent years, new successes in artificial intelligence and machine learning have been continuously achieved. However, this progress is largely based on the use of simulations as well as numerous powerful computers. Due to the volume taken up and the necessary power to run these components, this is not feasible for mobile robotics. Nevertheless, the use of machine learning in mobile robots is desirable in order to adapt to unknown or changing environmental conditions.This paper evaluates the performance of different reinforcement learning methods on a physical robot platform. This robot has an arm with two degrees of freedom that can be used to move across a surface. The goal is to learn the correct motion sequence of the arm to move the robot. The focus here is exclusively on using the robot’s onboard computer, a Raspberry Pi 4 Model B. To learn forward motion, Value Iteration and variants of Q-learning from the field of reinforcement learning are used.It is shown that since the structure of some problems can be described by a very limited problem space, even when using a physical robot relatively simple algorithms can yield sufficient learning results. Furthermore, hardware limitations may prevent using more complex algorithms.","PeriodicalId":282759,"journal":{"name":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 Sixth IEEE International Conference on Robotic Computing (IRC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IRC55401.2022.00051","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In recent years, new successes in artificial intelligence and machine learning have been continuously achieved. However, this progress is largely based on the use of simulations as well as numerous powerful computers. Due to the volume taken up and the necessary power to run these components, this is not feasible for mobile robotics. Nevertheless, the use of machine learning in mobile robots is desirable in order to adapt to unknown or changing environmental conditions.This paper evaluates the performance of different reinforcement learning methods on a physical robot platform. This robot has an arm with two degrees of freedom that can be used to move across a surface. The goal is to learn the correct motion sequence of the arm to move the robot. The focus here is exclusively on using the robot’s onboard computer, a Raspberry Pi 4 Model B. To learn forward motion, Value Iteration and variants of Q-learning from the field of reinforcement learning are used.It is shown that since the structure of some problems can be described by a very limited problem space, even when using a physical robot relatively simple algorithms can yield sufficient learning results. Furthermore, hardware limitations may prevent using more complex algorithms.
近年来,人工智能和机器学习领域不断取得新成就。然而,这一进展很大程度上是基于模拟的使用以及大量强大的计算机。由于占用的体积和运行这些组件所需的功率,这对于移动机器人来说是不可行的。然而,为了适应未知或不断变化的环境条件,在移动机器人中使用机器学习是可取的。本文评估了不同强化学习方法在物理机器人平台上的性能。这个机器人的手臂有两个自由度,可以用来在一个表面上移动。目标是学习手臂的正确运动顺序来移动机器人。这里的重点是专门使用机器人的机载计算机,Raspberry Pi 4 Model B.为了学习向前运动,使用了强化学习领域的值迭代和Q-learning的变体。研究表明,由于一些问题的结构可以用非常有限的问题空间来描述,即使使用物理机器人,相对简单的算法也可以产生足够的学习结果。此外,硬件限制可能妨碍使用更复杂的算法。