R. Allen, Y. Rachlin, J. Ruprecht, Sean Loughran, J. Varey, H. Viggh
{"title":"SpaceGym: Discrete and Differential Games in Non-Cooperative Space Operations","authors":"R. Allen, Y. Rachlin, J. Ruprecht, Sean Loughran, J. Varey, H. Viggh","doi":"10.1109/AERO55745.2023.10115968","DOIUrl":null,"url":null,"abstract":"This paper introduces a collection of non-cooperative game environments that are intended to spur development and act as proving grounds for autonomous and AI decision-makers in the orbital domain. SpaceGym comprises two distinct suites of game environments: OrbitDefender2D (OD2D) and the Ker-bal Space Program Differential Games suite (KSPDG). OrbitDe-fender2D consists of discrete, chess-like, two-player gridworld games. OD2D game mechanics are loosely based on orbital motion and players compete to maintain control of orbital slots. The KSPDG suite consists of multi-agent pursuit-evasion differential games constructed within the Kerbal Space Program (KSP) game engine. In comparison to the very limited set of comparable environments in the existing literature, KSPDG represents a much more configurable, extensible, and higher-fidelity aerospace environment suite that leverages a mature game engine to incorporate physics models for phenomenon such as collision mechanics, kinematic chains for deformable bodies, atmospheric drag, variable-mass propulsion, solar irradiance, and thermal models. Both the discrete and differential game suites are built with standardized input/output interfaces based on OpenAI Gym and PettingZoo specifications. This standardization enables-but does not enforce-the use of rein-forcement learning agents within the SpaceGym environments. As a comparison point for future research, we provide baseline agents that employ techniques of model predictive control, numerical differential game solvers, and reinforcement learning-along with their respective performance metrics-for a subset of the SpaceGym environments. The SpaceGym software libraries can be found at https://github.com/mit-II/spacegym-od2d and https://github.com/mit-II/spacegym-kspdg.","PeriodicalId":344285,"journal":{"name":"2023 IEEE Aerospace Conference","volume":"366 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE Aerospace Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/AERO55745.2023.10115968","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper introduces a collection of non-cooperative game environments that are intended to spur development and act as proving grounds for autonomous and AI decision-makers in the orbital domain. SpaceGym comprises two distinct suites of game environments: OrbitDefender2D (OD2D) and the Ker-bal Space Program Differential Games suite (KSPDG). OrbitDe-fender2D consists of discrete, chess-like, two-player gridworld games. OD2D game mechanics are loosely based on orbital motion and players compete to maintain control of orbital slots. The KSPDG suite consists of multi-agent pursuit-evasion differential games constructed within the Kerbal Space Program (KSP) game engine. In comparison to the very limited set of comparable environments in the existing literature, KSPDG represents a much more configurable, extensible, and higher-fidelity aerospace environment suite that leverages a mature game engine to incorporate physics models for phenomenon such as collision mechanics, kinematic chains for deformable bodies, atmospheric drag, variable-mass propulsion, solar irradiance, and thermal models. Both the discrete and differential game suites are built with standardized input/output interfaces based on OpenAI Gym and PettingZoo specifications. This standardization enables-but does not enforce-the use of rein-forcement learning agents within the SpaceGym environments. As a comparison point for future research, we provide baseline agents that employ techniques of model predictive control, numerical differential game solvers, and reinforcement learning-along with their respective performance metrics-for a subset of the SpaceGym environments. The SpaceGym software libraries can be found at https://github.com/mit-II/spacegym-od2d and https://github.com/mit-II/spacegym-kspdg.
本文介绍了一系列非合作游戏环境,旨在刺激开发,并作为轨道领域自主和人工智能决策者的试验场。SpaceGym包含两个不同的游戏环境套件:OrbitDefender2D (OD2D)和Ker-bal Space Program Differential Games套件(KSPDG)。《OrbitDe-fender2D》是一款类似国际象棋的双人网格世界游戏。OD2D游戏机制基于轨道运动,玩家通过竞争来保持对轨道槽的控制。KSPDG套件由在Kerbal空间计划(KSP)游戏引擎中构建的多智能体追击-逃避微分博弈组成。与现有文献中非常有限的可比环境相比,KSPDG代表了一个更具可配置性、可扩展性和更高保真度的航空航天环境套件,它利用成熟的游戏引擎将碰撞力学、可变形物体的运动链、大气阻力、变质量推进、太阳辐照度和热模型等现象的物理模型结合起来。离散和差分游戏套件都是基于OpenAI Gym和PettingZoo规范的标准化输入/输出接口构建的。这种标准化支持(但不强制)在SpaceGym环境中使用强化学习代理。作为未来研究的比较点,我们为SpaceGym环境的一个子集提供了采用模型预测控制、数值微分博弈求解器和强化学习技术的基线代理,以及它们各自的性能指标。SpaceGym软件库可以在https://github.com/mit-II/spacegym-od2d和https://github.com/mit-II/spacegym-kspdg上找到。