{"title":"混合电源下SWIPT系统的最优能量分配与多用户调度","authors":"Delin Guo, Lan Tang, Xinggan Zhang","doi":"10.1109/GCWkshps45667.2019.9024326","DOIUrl":null,"url":null,"abstract":"This paper studies the utilization and transfer of renewable green energy in a multiuser downlink communication network. In the considered multiuser system, the base station (BS) is powered by both harvested energy and grid. When the BS transmits data to one user terminal, other terminals can replenish energy opportunistically from received radio-frequency (RF) signals, which is called simultaneous wireless information and power transfer (SWIPT). Our objective is to maximize the average throughput by multiuser scheduling and energy allocation utilizing causal channel state information while satisfying the requirement for harvested energy and the average power constraint of the grid. With channel dynamics and energy arrival modeled as Markov processes, we characterize the problem as a Markov decision process (MDP). The standard reinforcement learning framework is considered as an effective solution to MDP. If the transition probability of MDP is known, the policy iteration (PI) algorithm is used to solve the problem, otherwise, the R-learning algorithm is adopted. Simulation results show that the proposed algorithm can improve the average throughput of the system and increase the energy harvested by idle user terminals compared with the existing work. And R-learning can achieve performance close to the PI algorithm under the condition that the channel transition probability is unknown.","PeriodicalId":210825,"journal":{"name":"2019 IEEE Globecom Workshops (GC Wkshps)","volume":"209 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Optimal Energy Allocation and Multiuser Scheduling in SWIPT Systems with Hybrid Power Supply\",\"authors\":\"Delin Guo, Lan Tang, Xinggan Zhang\",\"doi\":\"10.1109/GCWkshps45667.2019.9024326\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper studies the utilization and transfer of renewable green energy in a multiuser downlink communication network. In the considered multiuser system, the base station (BS) is powered by both harvested energy and grid. When the BS transmits data to one user terminal, other terminals can replenish energy opportunistically from received radio-frequency (RF) signals, which is called simultaneous wireless information and power transfer (SWIPT). Our objective is to maximize the average throughput by multiuser scheduling and energy allocation utilizing causal channel state information while satisfying the requirement for harvested energy and the average power constraint of the grid. With channel dynamics and energy arrival modeled as Markov processes, we characterize the problem as a Markov decision process (MDP). The standard reinforcement learning framework is considered as an effective solution to MDP. If the transition probability of MDP is known, the policy iteration (PI) algorithm is used to solve the problem, otherwise, the R-learning algorithm is adopted. Simulation results show that the proposed algorithm can improve the average throughput of the system and increase the energy harvested by idle user terminals compared with the existing work. And R-learning can achieve performance close to the PI algorithm under the condition that the channel transition probability is unknown.\",\"PeriodicalId\":210825,\"journal\":{\"name\":\"2019 IEEE Globecom Workshops (GC Wkshps)\",\"volume\":\"209 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Globecom Workshops (GC Wkshps)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/GCWkshps45667.2019.9024326\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Globecom Workshops (GC Wkshps)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GCWkshps45667.2019.9024326","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Optimal Energy Allocation and Multiuser Scheduling in SWIPT Systems with Hybrid Power Supply
This paper studies the utilization and transfer of renewable green energy in a multiuser downlink communication network. In the considered multiuser system, the base station (BS) is powered by both harvested energy and grid. When the BS transmits data to one user terminal, other terminals can replenish energy opportunistically from received radio-frequency (RF) signals, which is called simultaneous wireless information and power transfer (SWIPT). Our objective is to maximize the average throughput by multiuser scheduling and energy allocation utilizing causal channel state information while satisfying the requirement for harvested energy and the average power constraint of the grid. With channel dynamics and energy arrival modeled as Markov processes, we characterize the problem as a Markov decision process (MDP). The standard reinforcement learning framework is considered as an effective solution to MDP. If the transition probability of MDP is known, the policy iteration (PI) algorithm is used to solve the problem, otherwise, the R-learning algorithm is adopted. Simulation results show that the proposed algorithm can improve the average throughput of the system and increase the energy harvested by idle user terminals compared with the existing work. And R-learning can achieve performance close to the PI algorithm under the condition that the channel transition probability is unknown.