Masashi Sugimoto, Ryunosuke Uchida, S. Tsuzuki, Hitoshi Sori, H. Inoue, K. Kurashige, S. Urushihara
{"title":"An Experimental Study for Tracking Ability of Deep Q-Network under the Multi-Objective Behaviour using a Mobile Robot with LiDAR","authors":"Masashi Sugimoto, Ryunosuke Uchida, S. Tsuzuki, Hitoshi Sori, H. Inoue, K. Kurashige, S. Urushihara","doi":"10.1145/3459104.3459120","DOIUrl":null,"url":null,"abstract":"The Reinforcement Learning (RL) had been attracting attention for a long time that because it can be easily applied to real robots. On the other hand, in Q-Learning one of RL methods, since it contains the Q-table and grind environment is updated, especially, a large amount of Q-tables are required to express continuous “states,” such as smooth movements of the robot arm. Moreover, there was a disadvantage that calculation could not be performed real-time in case of amount of states and actions. The Deep Q-Network (DQN), on the other hand, uses convolutional neural network to estimate the Q-value itself, so that it can obtain an approximate function of the Q-value. From this characteristic of calculation that ignoring the amount of discrete states, this method has attracted attention, in recent. However, it seems to the following of multitasking and moving goal point that Q-Learning was not good at has been inherited by DQN. In this paper, the authors have improvements the multi-purpose execution of DQN by changing the exploration ratio as known as epsilon dynamically, has been tried. As the verification experiment, in the actual environment, a mobile crawler that mounting the NVIDIA Jetson NX and 2D LiDAR with the improvements DQN has been applied, to verify the object tracking ability, as a moving target position. As the result, the authors have confirmed that the improve its weak point.","PeriodicalId":142284,"journal":{"name":"2021 International Symposium on Electrical, Electronics and Information Engineering","volume":"92 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Symposium on Electrical, Electronics and Information Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3459104.3459120","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The Reinforcement Learning (RL) had been attracting attention for a long time that because it can be easily applied to real robots. On the other hand, in Q-Learning one of RL methods, since it contains the Q-table and grind environment is updated, especially, a large amount of Q-tables are required to express continuous “states,” such as smooth movements of the robot arm. Moreover, there was a disadvantage that calculation could not be performed real-time in case of amount of states and actions. The Deep Q-Network (DQN), on the other hand, uses convolutional neural network to estimate the Q-value itself, so that it can obtain an approximate function of the Q-value. From this characteristic of calculation that ignoring the amount of discrete states, this method has attracted attention, in recent. However, it seems to the following of multitasking and moving goal point that Q-Learning was not good at has been inherited by DQN. In this paper, the authors have improvements the multi-purpose execution of DQN by changing the exploration ratio as known as epsilon dynamically, has been tried. As the verification experiment, in the actual environment, a mobile crawler that mounting the NVIDIA Jetson NX and 2D LiDAR with the improvements DQN has been applied, to verify the object tracking ability, as a moving target position. As the result, the authors have confirmed that the improve its weak point.