Yuanpeng Zhang, Shizhan Zheng, Chao Xu, Shengze Cai
{"title":"Efficient navigation in vortical flows based on reinforcement learning and flow field prediction","authors":"Yuanpeng Zhang, Shizhan Zheng, Chao Xu, Shengze Cai","doi":"10.1016/j.oceaneng.2025.120937","DOIUrl":null,"url":null,"abstract":"<div><div>In this paper, we address the navigation problem of autonomous agents in complex, time-varying flow fields using Deep Reinforcement Learning (DRL). Specifically, the Proximal Policy Optimization (PPO) algorithm is used to solve Zermelo’s problem for point-to-point navigation tasks. The challenge of navigation in this article arises from the fact that the agent’s movement speed is slower than the surrounding flow velocity, requiring the agent to adapt to the flow dynamics rather than simply counteracting it. We propose the Look-Ahead State-Space (LASS) method as a novel approach to enhance navigation performance by enabling the agent to anticipate future states, which incorporate information from either true or predicted flow fields. A long short-term memory network combined with a transposed convolutional network is used to predict the future flow dynamics based solely on historical sensory data from the agent. Our results demonstrate that the LASS strategy improves the agent’s adaptability and significantly improves navigation success rates, even in dynamic environments. Additionally, we compare the PPO-based navigation method with an optimal control planner, revealing that while optimal control achieves marginally faster travel times, the DRL-based approach offers significant advantages in computational efficiency, making it more suitable for real-time applications.</div></div>","PeriodicalId":19403,"journal":{"name":"Ocean Engineering","volume":"327 ","pages":"Article 120937"},"PeriodicalIF":4.6000,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ocean Engineering","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S002980182500650X","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, CIVIL","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we address the navigation problem of autonomous agents in complex, time-varying flow fields using Deep Reinforcement Learning (DRL). Specifically, the Proximal Policy Optimization (PPO) algorithm is used to solve Zermelo’s problem for point-to-point navigation tasks. The challenge of navigation in this article arises from the fact that the agent’s movement speed is slower than the surrounding flow velocity, requiring the agent to adapt to the flow dynamics rather than simply counteracting it. We propose the Look-Ahead State-Space (LASS) method as a novel approach to enhance navigation performance by enabling the agent to anticipate future states, which incorporate information from either true or predicted flow fields. A long short-term memory network combined with a transposed convolutional network is used to predict the future flow dynamics based solely on historical sensory data from the agent. Our results demonstrate that the LASS strategy improves the agent’s adaptability and significantly improves navigation success rates, even in dynamic environments. Additionally, we compare the PPO-based navigation method with an optimal control planner, revealing that while optimal control achieves marginally faster travel times, the DRL-based approach offers significant advantages in computational efficiency, making it more suitable for real-time applications.
期刊介绍:
Ocean Engineering provides a medium for the publication of original research and development work in the field of ocean engineering. Ocean Engineering seeks papers in the following topics.