Stable Jumping Control Based on Deep Reinforcement Learning for a Locust-Inspired Robot.

IF 3.4 3区 医学 Q1 ENGINEERING, MULTIDISCIPLINARY Biomimetics Pub Date : 2024-09-11 DOI:10.3390/biomimetics9090548
Qijie Zhou, Gangyang Li, Rui Tang, Yi Xu, Hao Wen, Qing Shi
{"title":"Stable Jumping Control Based on Deep Reinforcement Learning for a Locust-Inspired Robot.","authors":"Qijie Zhou, Gangyang Li, Rui Tang, Yi Xu, Hao Wen, Qing Shi","doi":"10.3390/biomimetics9090548","DOIUrl":null,"url":null,"abstract":"<p><p>Biologically inspired jumping robots exhibit exceptional movement capabilities and can quickly overcome obstacles. However, the stability and accuracy of jumping movements are significantly compromised by rapid changes in posture. Here, we propose a stable jumping control algorithm for a locust-inspired jumping robot based on deep reinforcement learning. The algorithm utilizes a training framework comprising two neural network modules (actor network and critic network) to enhance training performance. The framework can control jumping by directly mapping the robot's observations (robot position and velocity, obstacle position, target position, etc.) to its joint torques. The control policy increases randomness and exploration by introducing an entropy term to the policy function. Moreover, we designed a stage incentive mechanism to adjust the reward function dynamically, thereby improving the robot's jumping stability and accuracy. We established a locus-inspired jumping robot platform and conducted a series of jumping experiments in simulation. The results indicate that the robot could perform smooth and non-flip jumps, with the error of the distance from the target remaining below 3%. The robot consumed 44.6% less energy to travel the same distance by jumping compared with walking. Additionally, the proposed algorithm exhibited a faster convergence rate and improved convergence effects compared with other classical algorithms.</p>","PeriodicalId":8907,"journal":{"name":"Biomimetics","volume":"9 9","pages":""},"PeriodicalIF":3.4000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11430585/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomimetics","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.3390/biomimetics9090548","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

Biologically inspired jumping robots exhibit exceptional movement capabilities and can quickly overcome obstacles. However, the stability and accuracy of jumping movements are significantly compromised by rapid changes in posture. Here, we propose a stable jumping control algorithm for a locust-inspired jumping robot based on deep reinforcement learning. The algorithm utilizes a training framework comprising two neural network modules (actor network and critic network) to enhance training performance. The framework can control jumping by directly mapping the robot's observations (robot position and velocity, obstacle position, target position, etc.) to its joint torques. The control policy increases randomness and exploration by introducing an entropy term to the policy function. Moreover, we designed a stage incentive mechanism to adjust the reward function dynamically, thereby improving the robot's jumping stability and accuracy. We established a locus-inspired jumping robot platform and conducted a series of jumping experiments in simulation. The results indicate that the robot could perform smooth and non-flip jumps, with the error of the distance from the target remaining below 3%. The robot consumed 44.6% less energy to travel the same distance by jumping compared with walking. Additionally, the proposed algorithm exhibited a faster convergence rate and improved convergence effects compared with other classical algorithms.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于深度强化学习的蝗虫启发机器人稳定跳跃控制
受生物启发的跳跃机器人表现出非凡的运动能力,能够快速克服障碍。然而,由于姿态的快速变化,跳跃动作的稳定性和准确性大打折扣。在此,我们提出了一种基于深度强化学习的蝗虫启发跳跃机器人稳定跳跃控制算法。该算法利用由两个神经网络模块(行动者网络和批评者网络)组成的训练框架来提高训练性能。该框架可通过将机器人的观测值(机器人位置和速度、障碍物位置、目标位置等)直接映射到其关节扭矩来控制跳跃。控制策略通过在策略函数中引入熵项,增加了随机性和探索性。此外,我们还设计了一种阶段激励机制来动态调整奖励函数,从而提高机器人跳跃的稳定性和准确性。我们建立了一个受定位点启发的跳跃机器人平台,并在仿真中进行了一系列跳跃实验。结果表明,机器人可以平稳、不翻转地跳跃,与目标的距离误差保持在 3% 以下。与行走相比,机器人在相同距离内跳跃所消耗的能量减少了 44.6%。此外,与其他经典算法相比,所提出的算法收敛速度更快,收敛效果更好。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Biomimetics
Biomimetics Biochemistry, Genetics and Molecular Biology-Biotechnology
CiteScore
3.50
自引率
11.10%
发文量
189
审稿时长
11 weeks
期刊最新文献
Brain-Inspired Architecture for Spiking Neural Networks. Explorative Binary Gray Wolf Optimizer with Quadratic Interpolation for Feature Selection. Path Planning of an Unmanned Aerial Vehicle Based on a Multi-Strategy Improved Pelican Optimization Algorithm. Performance Comparison of Bio-Inspired Algorithms for Optimizing an ANN-Based MPPT Forecast for PV Systems. Clinical Applications of Micro/Nanobubble Technology in Neurological Diseases.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1