A reinforcement learning-based evolutionary algorithm for the unmanned aerial vehicles maritime search and rescue path planning problem considering multiple rescue centers
{"title":"A reinforcement learning-based evolutionary algorithm for the unmanned aerial vehicles maritime search and rescue path planning problem considering multiple rescue centers","authors":"Haowen Zhan, Yue Zhang, Jingbo Huang, Yanjie Song, Lining Xing, Jie Wu, Zengyun Gao","doi":"10.1007/s12293-024-00420-8","DOIUrl":null,"url":null,"abstract":"<p>In the realm of maritime emergencies, unmanned aerial vehicles (UAVs) play a crucial role in enhancing search and rescue (SAR) operations. They help in efficiently rescuing distressed crews, strengthening maritime surveillance, and maintaining national security due to their cost-effectiveness, versatility, and effectiveness. However, the vast expanse of sea territories and the rapid changes in maritime conditions make a single SAR center insufficient for handling complex emergencies. Thus, it is vital to develop strategies for quickly deploying UAV resources from multiple SAR centers for area reconnaissance and supporting maritime rescue operations. This study introduces a graph-structured planning model for the maritime SAR path planning problem, considering multiple rescue centers (MSARPPP-MRC). It incorporates workload distribution among SAR centers and UAV operational constraints. We propose a reinforcement learning-based genetic algorithm (GA-RL) to tackle the MSARPPP-MRC problem. GA-RL uses heuristic rules to initialize the population and employs the Q-learning method to manage the progeny during each generation, including their retention, storage, or disposal. When the elite repository’s capacity is reached, a decision is made on the utilization of these members to refresh the population. Additionally, adaptive crossover and perturbation strategies are applied to develop a more effective SAR scheme. Extensive testing proves that GA-RL surpasses other algorithms in optimization efficacy and efficiency, highlighting the benefits of reinforcement learning in population management.</p>","PeriodicalId":3,"journal":{"name":"ACS Applied Electronic Materials","volume":null,"pages":null},"PeriodicalIF":4.3000,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACS Applied Electronic Materials","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s12293-024-00420-8","RegionNum":3,"RegionCategory":"材料科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
In the realm of maritime emergencies, unmanned aerial vehicles (UAVs) play a crucial role in enhancing search and rescue (SAR) operations. They help in efficiently rescuing distressed crews, strengthening maritime surveillance, and maintaining national security due to their cost-effectiveness, versatility, and effectiveness. However, the vast expanse of sea territories and the rapid changes in maritime conditions make a single SAR center insufficient for handling complex emergencies. Thus, it is vital to develop strategies for quickly deploying UAV resources from multiple SAR centers for area reconnaissance and supporting maritime rescue operations. This study introduces a graph-structured planning model for the maritime SAR path planning problem, considering multiple rescue centers (MSARPPP-MRC). It incorporates workload distribution among SAR centers and UAV operational constraints. We propose a reinforcement learning-based genetic algorithm (GA-RL) to tackle the MSARPPP-MRC problem. GA-RL uses heuristic rules to initialize the population and employs the Q-learning method to manage the progeny during each generation, including their retention, storage, or disposal. When the elite repository’s capacity is reached, a decision is made on the utilization of these members to refresh the population. Additionally, adaptive crossover and perturbation strategies are applied to develop a more effective SAR scheme. Extensive testing proves that GA-RL surpasses other algorithms in optimization efficacy and efficiency, highlighting the benefits of reinforcement learning in population management.
在海上紧急情况领域,无人驾驶飞行器(UAV)在加强搜救(SAR)行动方面发挥着至关重要的作用。由于其成本效益高、用途广泛且效果显著,它们有助于有效营救遇险船员、加强海上监视和维护国家安全。然而,幅员辽阔的海域和瞬息万变的海况使得单一的 SAR 中心不足以应对复杂的紧急情况。因此,制定从多个 SAR 中心快速部署无人机资源的战略,用于区域侦察和支持海上救援行动至关重要。本研究针对考虑多个救援中心的海上搜救路径规划问题引入了图结构规划模型(MSARPPP-MRC)。该模型纳入了 SAR 中心之间的工作量分配和无人机操作约束。我们提出了一种基于强化学习的遗传算法(GA-RL)来解决 MSARPPP-MRC 问题。GA-RL 使用启发式规则来初始化种群,并采用 Q-learning 方法来管理每一代的后代,包括保留、存储或处置。当精英库的容量达到一定程度时,就会决定是否利用这些成员来刷新种群。此外,还应用了自适应交叉和扰动策略来开发更有效的 SAR 方案。广泛的测试证明,GA-RL 在优化效果和效率方面超越了其他算法,突出了强化学习在种群管理方面的优势。