基于激光雷达的移动机器人多目标跟踪能力实验研究

Masashi Sugimoto, Ryunosuke Uchida, S. Tsuzuki, Hitoshi Sori, H. Inoue, K. Kurashige, S. Urushihara
{"title":"基于激光雷达的移动机器人多目标跟踪能力实验研究","authors":"Masashi Sugimoto, Ryunosuke Uchida, S. Tsuzuki, Hitoshi Sori, H. Inoue, K. Kurashige, S. Urushihara","doi":"10.1145/3459104.3459120","DOIUrl":null,"url":null,"abstract":"The Reinforcement Learning (RL) had been attracting attention for a long time that because it can be easily applied to real robots. On the other hand, in Q-Learning one of RL methods, since it contains the Q-table and grind environment is updated, especially, a large amount of Q-tables are required to express continuous “states,” such as smooth movements of the robot arm. Moreover, there was a disadvantage that calculation could not be performed real-time in case of amount of states and actions. The Deep Q-Network (DQN), on the other hand, uses convolutional neural network to estimate the Q-value itself, so that it can obtain an approximate function of the Q-value. From this characteristic of calculation that ignoring the amount of discrete states, this method has attracted attention, in recent. However, it seems to the following of multitasking and moving goal point that Q-Learning was not good at has been inherited by DQN. In this paper, the authors have improvements the multi-purpose execution of DQN by changing the exploration ratio as known as epsilon dynamically, has been tried. As the verification experiment, in the actual environment, a mobile crawler that mounting the NVIDIA Jetson NX and 2D LiDAR with the improvements DQN has been applied, to verify the object tracking ability, as a moving target position. As the result, the authors have confirmed that the improve its weak point.","PeriodicalId":142284,"journal":{"name":"2021 International Symposium on Electrical, Electronics and Information Engineering","volume":"92 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Experimental Study for Tracking Ability of Deep Q-Network under the Multi-Objective Behaviour using a Mobile Robot with LiDAR\",\"authors\":\"Masashi Sugimoto, Ryunosuke Uchida, S. Tsuzuki, Hitoshi Sori, H. Inoue, K. Kurashige, S. Urushihara\",\"doi\":\"10.1145/3459104.3459120\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The Reinforcement Learning (RL) had been attracting attention for a long time that because it can be easily applied to real robots. On the other hand, in Q-Learning one of RL methods, since it contains the Q-table and grind environment is updated, especially, a large amount of Q-tables are required to express continuous “states,” such as smooth movements of the robot arm. Moreover, there was a disadvantage that calculation could not be performed real-time in case of amount of states and actions. The Deep Q-Network (DQN), on the other hand, uses convolutional neural network to estimate the Q-value itself, so that it can obtain an approximate function of the Q-value. From this characteristic of calculation that ignoring the amount of discrete states, this method has attracted attention, in recent. However, it seems to the following of multitasking and moving goal point that Q-Learning was not good at has been inherited by DQN. In this paper, the authors have improvements the multi-purpose execution of DQN by changing the exploration ratio as known as epsilon dynamically, has been tried. As the verification experiment, in the actual environment, a mobile crawler that mounting the NVIDIA Jetson NX and 2D LiDAR with the improvements DQN has been applied, to verify the object tracking ability, as a moving target position. As the result, the authors have confirmed that the improve its weak point.\",\"PeriodicalId\":142284,\"journal\":{\"name\":\"2021 International Symposium on Electrical, Electronics and Information Engineering\",\"volume\":\"92 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-02-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Symposium on Electrical, Electronics and Information Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3459104.3459120\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Symposium on Electrical, Electronics and Information Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3459104.3459120","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

长期以来,强化学习(RL)一直受到人们的关注,因为它可以很容易地应用于真实的机器人。另一方面,在强化学习方法之一的Q-Learning中,由于它包含q表和磨削环境的更新,特别是需要大量的q表来表达连续的“状态”,例如机器人手臂的平滑运动。此外,存在一个缺点,即在状态和动作数量较多的情况下无法实时进行计算。而深度Q-Network (Deep Q-Network, DQN)则利用卷积神经网络对自身的q值进行估计,从而得到q值的近似函数。由于这种计算忽略了离散状态的数量的特点,近年来引起了人们的注意。然而,Q-Learning不擅长的多任务处理和移动目标点似乎被DQN继承了。在本文中,作者尝试通过动态改变探索比(即epsilon)来改进DQN的多用途执行。作为验证实验,在实际环境中,应用了安装NVIDIA Jetson NX和改进DQN的2D激光雷达的移动履带车,作为移动目标位置来验证目标跟踪能力。结果证实了该方法对其缺点的改进。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
An Experimental Study for Tracking Ability of Deep Q-Network under the Multi-Objective Behaviour using a Mobile Robot with LiDAR
The Reinforcement Learning (RL) had been attracting attention for a long time that because it can be easily applied to real robots. On the other hand, in Q-Learning one of RL methods, since it contains the Q-table and grind environment is updated, especially, a large amount of Q-tables are required to express continuous “states,” such as smooth movements of the robot arm. Moreover, there was a disadvantage that calculation could not be performed real-time in case of amount of states and actions. The Deep Q-Network (DQN), on the other hand, uses convolutional neural network to estimate the Q-value itself, so that it can obtain an approximate function of the Q-value. From this characteristic of calculation that ignoring the amount of discrete states, this method has attracted attention, in recent. However, it seems to the following of multitasking and moving goal point that Q-Learning was not good at has been inherited by DQN. In this paper, the authors have improvements the multi-purpose execution of DQN by changing the exploration ratio as known as epsilon dynamically, has been tried. As the verification experiment, in the actual environment, a mobile crawler that mounting the NVIDIA Jetson NX and 2D LiDAR with the improvements DQN has been applied, to verify the object tracking ability, as a moving target position. As the result, the authors have confirmed that the improve its weak point.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Exploring the Integration of Blockchain Technology and IoT in a Smart University Application Architecture 3D Moving Rigid Body Localization in the Presence of Anchor Position Errors RANS/LES Simulation of Low-Frequency Flow Oscillations on a NACA0012 Airfoil Near Stall Tuning Language Representation Models for Classification of Turkish News Improving Consumer Experience for Medical Information Using Text Analytics
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1