F. D. Kronewitter, Sumner Lee, Kenneth. Oliphant, Dell. Kronewitter, Kenneth. Oliphant
{"title":"A Cognitive ML Agent for Airborne Networking","authors":"F. D. Kronewitter, Sumner Lee, Kenneth. Oliphant, Dell. Kronewitter, Kenneth. Oliphant","doi":"10.1109/MILCOM47813.2019.9020975","DOIUrl":null,"url":null,"abstract":"This conference note describes a Deep Reinforcement Learning architecture specifically designed to improve wireless network performance over a heterogeneous airborne wireless network consisting of multiple waveforms, antennas, platforms, link protocols, frequencies, spatial transmission, and codes. The cooperative optimization of this high dimensional space is a difficult problem which obviously has a highly correlated characterization where human network operators cannot possibly capture these correlations. Model-free Reinforcement Learning techniques represent a potential solution to our problem. Specifically, we use Deep Q-Learning Networks (DQN) to improve networking performance. We have developed a high-fidelity network simulation tool we call Tactical Airborne Network Simulator (TANS) which we use to train our neural network before deploying to the field where the asset is deployed to some mission which is hopefully somewhat similar to the scenarios used for training. By utilizing the model developed under the TANS training scenarios for the target mission scenario our learning technique gets a head start, rather than using a truly model-free approach. Our technique is codified in the ML community as “Deep Transfer Learning” [4] where terms and metrics have been examined. This paper represents an initial investigation into both the decision support agent architecture and the ML technique. Upcoming research will be described below including our vision for an expanded agent architecture as well as ideas for improved ML techniques which ultimately will result in better wireless network performance. Here we demonstrate a minor throughput performance improvement of 4% using a proof of concept agent over the use of a standard unassisted network. We improved the throughput from 309kbps to 324 kbps.","PeriodicalId":371812,"journal":{"name":"MILCOM 2019 - 2019 IEEE Military Communications Conference (MILCOM)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"MILCOM 2019 - 2019 IEEE Military Communications Conference (MILCOM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MILCOM47813.2019.9020975","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This conference note describes a Deep Reinforcement Learning architecture specifically designed to improve wireless network performance over a heterogeneous airborne wireless network consisting of multiple waveforms, antennas, platforms, link protocols, frequencies, spatial transmission, and codes. The cooperative optimization of this high dimensional space is a difficult problem which obviously has a highly correlated characterization where human network operators cannot possibly capture these correlations. Model-free Reinforcement Learning techniques represent a potential solution to our problem. Specifically, we use Deep Q-Learning Networks (DQN) to improve networking performance. We have developed a high-fidelity network simulation tool we call Tactical Airborne Network Simulator (TANS) which we use to train our neural network before deploying to the field where the asset is deployed to some mission which is hopefully somewhat similar to the scenarios used for training. By utilizing the model developed under the TANS training scenarios for the target mission scenario our learning technique gets a head start, rather than using a truly model-free approach. Our technique is codified in the ML community as “Deep Transfer Learning” [4] where terms and metrics have been examined. This paper represents an initial investigation into both the decision support agent architecture and the ML technique. Upcoming research will be described below including our vision for an expanded agent architecture as well as ideas for improved ML techniques which ultimately will result in better wireless network performance. Here we demonstrate a minor throughput performance improvement of 4% using a proof of concept agent over the use of a standard unassisted network. We improved the throughput from 309kbps to 324 kbps.