Efficient navigation in vortical flows based on reinforcement learning and flow field prediction

IF 5.5 2区 工程技术 Q1 ENGINEERING, CIVIL Ocean Engineering Pub Date : 2025-05-30 Epub Date: 2025-03-18 DOI:10.1016/j.oceaneng.2025.120937
Yuanpeng Zhang, Shizhan Zheng, Chao Xu, Shengze Cai
{"title":"Efficient navigation in vortical flows based on reinforcement learning and flow field prediction","authors":"Yuanpeng Zhang,&nbsp;Shizhan Zheng,&nbsp;Chao Xu,&nbsp;Shengze Cai","doi":"10.1016/j.oceaneng.2025.120937","DOIUrl":null,"url":null,"abstract":"<div><div>In this paper, we address the navigation problem of autonomous agents in complex, time-varying flow fields using Deep Reinforcement Learning (DRL). Specifically, the Proximal Policy Optimization (PPO) algorithm is used to solve Zermelo’s problem for point-to-point navigation tasks. The challenge of navigation in this article arises from the fact that the agent’s movement speed is slower than the surrounding flow velocity, requiring the agent to adapt to the flow dynamics rather than simply counteracting it. We propose the Look-Ahead State-Space (LASS) method as a novel approach to enhance navigation performance by enabling the agent to anticipate future states, which incorporate information from either true or predicted flow fields. A long short-term memory network combined with a transposed convolutional network is used to predict the future flow dynamics based solely on historical sensory data from the agent. Our results demonstrate that the LASS strategy improves the agent’s adaptability and significantly improves navigation success rates, even in dynamic environments. Additionally, we compare the PPO-based navigation method with an optimal control planner, revealing that while optimal control achieves marginally faster travel times, the DRL-based approach offers significant advantages in computational efficiency, making it more suitable for real-time applications.</div></div>","PeriodicalId":19403,"journal":{"name":"Ocean Engineering","volume":"327 ","pages":"Article 120937"},"PeriodicalIF":5.5000,"publicationDate":"2025-05-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ocean Engineering","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S002980182500650X","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/3/18 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"ENGINEERING, CIVIL","Score":null,"Total":0}
引用次数: 0

Abstract

In this paper, we address the navigation problem of autonomous agents in complex, time-varying flow fields using Deep Reinforcement Learning (DRL). Specifically, the Proximal Policy Optimization (PPO) algorithm is used to solve Zermelo’s problem for point-to-point navigation tasks. The challenge of navigation in this article arises from the fact that the agent’s movement speed is slower than the surrounding flow velocity, requiring the agent to adapt to the flow dynamics rather than simply counteracting it. We propose the Look-Ahead State-Space (LASS) method as a novel approach to enhance navigation performance by enabling the agent to anticipate future states, which incorporate information from either true or predicted flow fields. A long short-term memory network combined with a transposed convolutional network is used to predict the future flow dynamics based solely on historical sensory data from the agent. Our results demonstrate that the LASS strategy improves the agent’s adaptability and significantly improves navigation success rates, even in dynamic environments. Additionally, we compare the PPO-based navigation method with an optimal control planner, revealing that while optimal control achieves marginally faster travel times, the DRL-based approach offers significant advantages in computational efficiency, making it more suitable for real-time applications.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于强化学习和流场预测的高效涡旋导航
在本文中,我们使用深度强化学习(DRL)解决了复杂时变流场中自主代理的导航问题。具体地说,采用近端策略优化(PPO)算法来解决点到点导航任务中的Zermelo问题。本文中导航的挑战来自于agent的移动速度比周围的流速度慢,这要求agent适应流动力学而不是简单地抵消它。我们提出了前瞻状态空间(LASS)方法,作为一种新的方法,通过使智能体能够预测未来的状态来提高导航性能,这些状态包含来自真实或预测流场的信息。将长短期记忆网络与转置卷积网络相结合,仅根据智能体的历史感官数据来预测未来的流量动态。我们的研究结果表明,LASS策略提高了智能体的适应性,并显著提高了导航成功率,即使在动态环境中也是如此。此外,我们将基于ppo的导航方法与最优控制规划器进行了比较,结果表明,虽然最优控制可以实现略快的行程时间,但基于drl的方法在计算效率方面具有显著优势,使其更适合实时应用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Ocean Engineering
Ocean Engineering 工程技术-工程:大洋
CiteScore
7.30
自引率
34.00%
发文量
2379
审稿时长
8.1 months
期刊介绍: Ocean Engineering provides a medium for the publication of original research and development work in the field of ocean engineering. Ocean Engineering seeks papers in the following topics.
期刊最新文献
Incorporation of movement pattern analysis into route searching for ship ETA prediction Assessing wind-wave forces on smooth and rough horizontal cylinders by neural networks SPH-based numerical investigation into hydrodynamic performance and solitary-wave-based mooring screening of FPV array shielded by breakwaters Influence of earthquake duration on the non-linear slosh dynamics of chamfered bottom liquid tanks using Mixed-Eulerian Lagrangian approach Buckling of the constrained subsea FGP-GPLs liner under pressure loading and heating field based on multi-source data fusion scheme
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1