Yutong Xie, Zhengzhuo Xu, Yuxing Zhong, Jing Xu, Shimin Gong, Yi Wang
{"title":"基于策略的深度强化学习的能量收集物联网设备的反向散射辅助计算卸载","authors":"Yutong Xie, Zhengzhuo Xu, Yuxing Zhong, Jing Xu, Shimin Gong, Yi Wang","doi":"10.1109/ICCChinaW.2019.8849964","DOIUrl":null,"url":null,"abstract":"Wireless Internet of Things (IoT) devices can be deployed for data acquisition and decision making, e.g., the wearable sensors used for healthcare monitoring. Due to limited computation capability, the low-power IoT devices can optionally offload power-consuming computation to a nearby computing server. To balance power consumption in data offloading and computation, we propose a novel hybrid data offloading scheme that allows each device to offload data via either the conventional RF communications or low-power backscatter communications. Such a flexibility makes it more complicated to optimize the offloading strategy with uncertain workload and energy supply at each device. As such, we propose the deep reinforcement learning (DRL) to learn the optimal offloading policy from past experience. In particular, we rely on the policy-based DRL approach for continuous control problems in the actor-critic framework. By interacting with the network environment, we can optimize each user's energy harvesting time and the workload allocation among different offloading schemes. The numerical results show that the proposed DRL approach can achieve much higher reward and learning speed compared to the conventional deep Q-network method.","PeriodicalId":252172,"journal":{"name":"2019 IEEE/CIC International Conference on Communications Workshops in China (ICCC Workshops)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"22","resultStr":"{\"title\":\"Backscatter-Assisted Computation Offloading for Energy Harvesting IoT Devices via Policy-based Deep Reinforcement Learning\",\"authors\":\"Yutong Xie, Zhengzhuo Xu, Yuxing Zhong, Jing Xu, Shimin Gong, Yi Wang\",\"doi\":\"10.1109/ICCChinaW.2019.8849964\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Wireless Internet of Things (IoT) devices can be deployed for data acquisition and decision making, e.g., the wearable sensors used for healthcare monitoring. Due to limited computation capability, the low-power IoT devices can optionally offload power-consuming computation to a nearby computing server. To balance power consumption in data offloading and computation, we propose a novel hybrid data offloading scheme that allows each device to offload data via either the conventional RF communications or low-power backscatter communications. Such a flexibility makes it more complicated to optimize the offloading strategy with uncertain workload and energy supply at each device. As such, we propose the deep reinforcement learning (DRL) to learn the optimal offloading policy from past experience. In particular, we rely on the policy-based DRL approach for continuous control problems in the actor-critic framework. By interacting with the network environment, we can optimize each user's energy harvesting time and the workload allocation among different offloading schemes. The numerical results show that the proposed DRL approach can achieve much higher reward and learning speed compared to the conventional deep Q-network method.\",\"PeriodicalId\":252172,\"journal\":{\"name\":\"2019 IEEE/CIC International Conference on Communications Workshops in China (ICCC Workshops)\",\"volume\":\"61 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"22\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE/CIC International Conference on Communications Workshops in China (ICCC Workshops)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCChinaW.2019.8849964\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE/CIC International Conference on Communications Workshops in China (ICCC Workshops)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCChinaW.2019.8849964","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Backscatter-Assisted Computation Offloading for Energy Harvesting IoT Devices via Policy-based Deep Reinforcement Learning
Wireless Internet of Things (IoT) devices can be deployed for data acquisition and decision making, e.g., the wearable sensors used for healthcare monitoring. Due to limited computation capability, the low-power IoT devices can optionally offload power-consuming computation to a nearby computing server. To balance power consumption in data offloading and computation, we propose a novel hybrid data offloading scheme that allows each device to offload data via either the conventional RF communications or low-power backscatter communications. Such a flexibility makes it more complicated to optimize the offloading strategy with uncertain workload and energy supply at each device. As such, we propose the deep reinforcement learning (DRL) to learn the optimal offloading policy from past experience. In particular, we rely on the policy-based DRL approach for continuous control problems in the actor-critic framework. By interacting with the network environment, we can optimize each user's energy harvesting time and the workload allocation among different offloading schemes. The numerical results show that the proposed DRL approach can achieve much higher reward and learning speed compared to the conventional deep Q-network method.