Adaptive Multiagent Reinforcement Learning Solver for Tactical Conflict Resolution in Diverse Urban Airspace Configurations

IF 5.7 2区 计算机科学 Q1 ENGINEERING, AEROSPACE IEEE Transactions on Aerospace and Electronic Systems Pub Date : 2024-10-30 DOI:10.1109/TAES.2024.3479191
Rodolphe Fremond;Yan Xu;Gokhan Inalhan
{"title":"Adaptive Multiagent Reinforcement Learning Solver for Tactical Conflict Resolution in Diverse Urban Airspace Configurations","authors":"Rodolphe Fremond;Yan Xu;Gokhan Inalhan","doi":"10.1109/TAES.2024.3479191","DOIUrl":null,"url":null,"abstract":"This article discusses the development of an adaptive deep reinforcement learning solver, designed for centralized on-ground tactical conflict resolution and applied within diverse in-cruise configurations of urban airspace. Our approach utilizes a multiagent reinforcement learning algorithm with a shared policy framework, employing a proximal policy optimization baseline. The solver aims to ensure tactical conflict resolution through centralized decision-making and distributed instructions via speed calibration and flight-level assignment. Each operation adheres to a designated flight plan and operational volume tolerances. Our methodology first enhances the solver's adaptability by interpreting a broad spectrum of in-cruise conflict configurations. This is achieved through a cautious observation and action space design, a normalization strategy, and a recurrent neural network that generalizes conflict resolution for any number of intruders. We also apply existing standards from airborne collision avoidance systems to evaluate the safety performance of our machine-learning-based solver, addressing the gap in machine learning model evaluation as a conventional system. We explore eight case studies of varying difficulty to investigate the safety impact of specific conflict configurations, from single to multiple intersections, considering shared routes or individual operation volumes, among other factors. Finally, we develop a ninth case study that maps all possible configurations under cruise conditions. We explore different trained models and assess their compatibility with other case studies, demonstrating the solver's adaptability in performing tactical conflict resolution tasks for operations in urban airspace.","PeriodicalId":13157,"journal":{"name":"IEEE Transactions on Aerospace and Electronic Systems","volume":"61 2","pages":"2802-2820"},"PeriodicalIF":5.7000,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Aerospace and Electronic Systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10739911/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, AEROSPACE","Score":null,"Total":0}
引用次数: 0

Abstract

This article discusses the development of an adaptive deep reinforcement learning solver, designed for centralized on-ground tactical conflict resolution and applied within diverse in-cruise configurations of urban airspace. Our approach utilizes a multiagent reinforcement learning algorithm with a shared policy framework, employing a proximal policy optimization baseline. The solver aims to ensure tactical conflict resolution through centralized decision-making and distributed instructions via speed calibration and flight-level assignment. Each operation adheres to a designated flight plan and operational volume tolerances. Our methodology first enhances the solver's adaptability by interpreting a broad spectrum of in-cruise conflict configurations. This is achieved through a cautious observation and action space design, a normalization strategy, and a recurrent neural network that generalizes conflict resolution for any number of intruders. We also apply existing standards from airborne collision avoidance systems to evaluate the safety performance of our machine-learning-based solver, addressing the gap in machine learning model evaluation as a conventional system. We explore eight case studies of varying difficulty to investigate the safety impact of specific conflict configurations, from single to multiple intersections, considering shared routes or individual operation volumes, among other factors. Finally, we develop a ninth case study that maps all possible configurations under cruise conditions. We explore different trained models and assess their compatibility with other case studies, demonstrating the solver's adaptability in performing tactical conflict resolution tasks for operations in urban airspace.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
用于解决多样化城市空域配置中战术冲突的自适应多代理强化学习求解器
本文讨论了一种自适应深度强化学习求解器的开发,该求解器设计用于集中的地面战术冲突解决,并应用于城市空域的各种巡航配置。我们的方法利用具有共享策略框架的多智能体强化学习算法,采用近端策略优化基线。求解器的目标是通过速度校准和飞行水平分配的集中决策和分布式指令来确保战术冲突的解决。每次操作都遵守指定的飞行计划和操作容积公差。我们的方法首先通过解释巡航中广泛的冲突配置来增强求解器的适应性。这是通过谨慎的观察和行动空间设计、规范化策略和循环神经网络来实现的,该神经网络可以针对任意数量的入侵者推广冲突解决方案。我们还应用了机载避碰系统的现有标准来评估我们基于机器学习的求解器的安全性能,解决了机器学习模型评估与传统系统的差距。我们探索了八个不同难度的案例研究,以调查从单个到多个交叉口的特定冲突配置对安全的影响,考虑共享路线或单独的运行量,以及其他因素。最后,我们开发了第九个案例研究,在巡航条件下映射所有可能的配置。我们探索了不同的训练模型,并评估了它们与其他案例研究的兼容性,展示了求解器在执行城市空域作战战术冲突解决任务时的适应性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
7.80
自引率
13.60%
发文量
433
审稿时长
8.7 months
期刊介绍: IEEE Transactions on Aerospace and Electronic Systems focuses on the organization, design, development, integration, and operation of complex systems for space, air, ocean, or ground environment. These systems include, but are not limited to, navigation, avionics, spacecraft, aerospace power, radar, sonar, telemetry, defense, transportation, automated testing, and command and control.
期刊最新文献
Online Trajectory Planning for Hypersonic Glide Vehicle under Multiple No-Fly Zones: An Attention Mechanism-based BiGRU Framework Rapid Indirect Diagnosis of MEMS Gyroscope Initial Bias With On-Board Capability for Guided Missiles Dual Event-Triggered Remote Information-based State Estimation with Measurement Outliers Robust Model-Based Reinforcement Learning for Rocket Landing Deep Learning–Assisted UAV Localization Framework for Post-Disaster Search and Rescue Missions
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1