{"title":"Nonholonomic Robot Navigation of Mazes using Reinforcement Learning","authors":"Danielle K. Gleason, Michael Jenkin","doi":"10.5220/0011123600003271","DOIUrl":null,"url":null,"abstract":": Developing a navigation function for an unknown environment is a difficult task, made even more challenging when the environment has complex structure and the robot imposes nonholonomic constraints on the problem. Here we pose the problem of navigating an unknown environment as a reinforcement learning task for an Ackermann vehicle. We model environmental complexity using a standard characterization of mazes, and we show that training on complex maze architectures with loops (braid and partial braid mazes) results in an effective policy, but that for a more efficient policy, training on mazes without loops (perfect mazes) is to be preferred. Experimental results obtained in simulation are validated on a real robot operating both indoors and outdoors, assuming good localization and a 2D LIDAR to recover the local structure of the environment.","PeriodicalId":6436,"journal":{"name":"2010 2nd International Asia Conference on Informatics in Control, Automation and Robotics (CAR 2010)","volume":"23 1","pages":"369-376"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 2nd International Asia Conference on Informatics in Control, Automation and Robotics (CAR 2010)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5220/0011123600003271","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
: Developing a navigation function for an unknown environment is a difficult task, made even more challenging when the environment has complex structure and the robot imposes nonholonomic constraints on the problem. Here we pose the problem of navigating an unknown environment as a reinforcement learning task for an Ackermann vehicle. We model environmental complexity using a standard characterization of mazes, and we show that training on complex maze architectures with loops (braid and partial braid mazes) results in an effective policy, but that for a more efficient policy, training on mazes without loops (perfect mazes) is to be preferred. Experimental results obtained in simulation are validated on a real robot operating both indoors and outdoors, assuming good localization and a 2D LIDAR to recover the local structure of the environment.