{"title":"基于数字孪生的快速优化飞行基站鲁棒q学习","authors":"T. Guo","doi":"10.1109/VTC2022-Fall57202.2022.10012827","DOIUrl":null,"url":null,"abstract":"This paper studies a use case for quickly connecting users with an aerial bass station (BS) in emergency by leveraging digital twin (DT) and robust reinforcement learning (RL). Scattered communities of users are assigned a limited number of channels, and the flying BS is autonomously and optimally placed according to predefined criteria. Q-learning, a common type of RL, is employed as a solution to optimization of BS placement. Two optimization objectives are considered to maximize the per-user data rate in the worst condition and minimize the total BS transmitted power, respectively. To overcome resource limitations of the aerial BS, the RL training with many iterations is done in the DT virtual space connected to the physical space via an aerial BS. In particular, a practically sound max-min technique is proposed to handle the measurement and prediction uncertainties. It is shown that, even if the models used in DT are imperfect and measurements are inaccurate, nearly-optimal results can be obtained in DT. Compared to RL training 100% in the physical space, a huge number of BS moves can be avoided and a significant amount of time and energy can be saved. The assessment results suggest that model and measurement errors, especially when applying DT, should be seriously considered, and the robust optimization technique has potential to handle the uncertainties.","PeriodicalId":326047,"journal":{"name":"2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall)","volume":"97 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Robust Q-learning for Fast And Optimal Flying Base Station Placement Aided By Digital Twin For Emergency Use\",\"authors\":\"T. Guo\",\"doi\":\"10.1109/VTC2022-Fall57202.2022.10012827\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper studies a use case for quickly connecting users with an aerial bass station (BS) in emergency by leveraging digital twin (DT) and robust reinforcement learning (RL). Scattered communities of users are assigned a limited number of channels, and the flying BS is autonomously and optimally placed according to predefined criteria. Q-learning, a common type of RL, is employed as a solution to optimization of BS placement. Two optimization objectives are considered to maximize the per-user data rate in the worst condition and minimize the total BS transmitted power, respectively. To overcome resource limitations of the aerial BS, the RL training with many iterations is done in the DT virtual space connected to the physical space via an aerial BS. In particular, a practically sound max-min technique is proposed to handle the measurement and prediction uncertainties. It is shown that, even if the models used in DT are imperfect and measurements are inaccurate, nearly-optimal results can be obtained in DT. Compared to RL training 100% in the physical space, a huge number of BS moves can be avoided and a significant amount of time and energy can be saved. The assessment results suggest that model and measurement errors, especially when applying DT, should be seriously considered, and the robust optimization technique has potential to handle the uncertainties.\",\"PeriodicalId\":326047,\"journal\":{\"name\":\"2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall)\",\"volume\":\"97 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/VTC2022-Fall57202.2022.10012827\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 96th Vehicular Technology Conference (VTC2022-Fall)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/VTC2022-Fall57202.2022.10012827","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Robust Q-learning for Fast And Optimal Flying Base Station Placement Aided By Digital Twin For Emergency Use
This paper studies a use case for quickly connecting users with an aerial bass station (BS) in emergency by leveraging digital twin (DT) and robust reinforcement learning (RL). Scattered communities of users are assigned a limited number of channels, and the flying BS is autonomously and optimally placed according to predefined criteria. Q-learning, a common type of RL, is employed as a solution to optimization of BS placement. Two optimization objectives are considered to maximize the per-user data rate in the worst condition and minimize the total BS transmitted power, respectively. To overcome resource limitations of the aerial BS, the RL training with many iterations is done in the DT virtual space connected to the physical space via an aerial BS. In particular, a practically sound max-min technique is proposed to handle the measurement and prediction uncertainties. It is shown that, even if the models used in DT are imperfect and measurements are inaccurate, nearly-optimal results can be obtained in DT. Compared to RL training 100% in the physical space, a huge number of BS moves can be avoided and a significant amount of time and energy can be saved. The assessment results suggest that model and measurement errors, especially when applying DT, should be seriously considered, and the robust optimization technique has potential to handle the uncertainties.