绕行环境下导航的深度强化学习新领域探索

Jian Jiang, Junzhe Xu, Jianhua Zhang, Shengyong Chen
{"title":"绕行环境下导航的深度强化学习新领域探索","authors":"Jian Jiang, Junzhe Xu, Jianhua Zhang, Shengyong Chen","doi":"10.1109/ICARM52023.2021.9536098","DOIUrl":null,"url":null,"abstract":"Deep Reinforcement Learning (DRL) has made a great progress in recent years with the development of many relative researching areas, such as Deep Learning. Researchers have trained agents to achieve human-level and even beyond human-level scores in video games by using DRL. In the field of robotics, DRL can also achieve satisfactory performance for the navigation task when the environment is relatively simple. However, when environments become complex, e.g., the detour ones, the DRL system often fails to attain good results. To tackle this problem, we propose an internal reward obtaining method called New-Field-Explore (NFE) mechanism which can navigate a robot from initial position to target position without collision in detour environments. We also present a benchmark suite based on the AI2-Thor environment for robot navigation in complex detour environments. The proposed method is evaluated in these environments by comparing the performance of state-of-the-art algorithms with or without the NFE mechanism1. Experimental results show the above reward is effective for mobile robot navigation tasks in detour indoor environments.","PeriodicalId":367307,"journal":{"name":"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep Reinforcement Learning with New-Field Exploration for Navigation in Detour Environment\",\"authors\":\"Jian Jiang, Junzhe Xu, Jianhua Zhang, Shengyong Chen\",\"doi\":\"10.1109/ICARM52023.2021.9536098\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Deep Reinforcement Learning (DRL) has made a great progress in recent years with the development of many relative researching areas, such as Deep Learning. Researchers have trained agents to achieve human-level and even beyond human-level scores in video games by using DRL. In the field of robotics, DRL can also achieve satisfactory performance for the navigation task when the environment is relatively simple. However, when environments become complex, e.g., the detour ones, the DRL system often fails to attain good results. To tackle this problem, we propose an internal reward obtaining method called New-Field-Explore (NFE) mechanism which can navigate a robot from initial position to target position without collision in detour environments. We also present a benchmark suite based on the AI2-Thor environment for robot navigation in complex detour environments. The proposed method is evaluated in these environments by comparing the performance of state-of-the-art algorithms with or without the NFE mechanism1. Experimental results show the above reward is effective for mobile robot navigation tasks in detour indoor environments.\",\"PeriodicalId\":367307,\"journal\":{\"name\":\"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)\",\"volume\":\"45 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICARM52023.2021.9536098\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 6th IEEE International Conference on Advanced Robotics and Mechatronics (ICARM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICARM52023.2021.9536098","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

近年来,随着深度学习等相关研究领域的发展,深度强化学习(Deep Reinforcement Learning, DRL)取得了长足的进步。研究人员已经通过使用DRL训练智能体在电子游戏中达到甚至超过人类水平的分数。在机器人领域,在环境相对简单的情况下,DRL对于导航任务也能取得令人满意的表现。然而,当环境变得复杂时,例如绕行时,DRL系统往往不能达到很好的效果。为了解决这个问题,我们提出了一种内部奖励获取方法,称为新领域探索(NFE)机制,该机制可以使机器人在绕行环境中从初始位置导航到目标位置而不会发生碰撞。我们还提出了一个基于AI2-Thor环境的基准套件,用于复杂绕路环境下的机器人导航。在这些环境中,通过比较具有或不具有NFE机制的最先进算法的性能来评估所提出的方法。实验结果表明,上述奖励对移动机器人在绕行室内环境下的导航任务是有效的。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Deep Reinforcement Learning with New-Field Exploration for Navigation in Detour Environment
Deep Reinforcement Learning (DRL) has made a great progress in recent years with the development of many relative researching areas, such as Deep Learning. Researchers have trained agents to achieve human-level and even beyond human-level scores in video games by using DRL. In the field of robotics, DRL can also achieve satisfactory performance for the navigation task when the environment is relatively simple. However, when environments become complex, e.g., the detour ones, the DRL system often fails to attain good results. To tackle this problem, we propose an internal reward obtaining method called New-Field-Explore (NFE) mechanism which can navigate a robot from initial position to target position without collision in detour environments. We also present a benchmark suite based on the AI2-Thor environment for robot navigation in complex detour environments. The proposed method is evaluated in these environments by comparing the performance of state-of-the-art algorithms with or without the NFE mechanism1. Experimental results show the above reward is effective for mobile robot navigation tasks in detour indoor environments.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Non-model Friction Disturbance Compensation of a Pan-tilt Based on MUAV for Aerial Remote Sensing Application Multi-Modal Attention Guided Real-Time Lane Detection Amphibious Robot with a Novel Composite Propulsion Mechanism Iterative Learning Control of Impedance Parameters for a Soft Exosuit Triple-step Nonlinear Controller with MLFNN for a Lower Limb Rehabilitation Robot
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1