Computational Resource Sharing in a Vehicular Cloud Network via Deep Reinforcement Learning

Shilin Xu, Caili Guo, R. Hu, Y. Qian
{"title":"Computational Resource Sharing in a Vehicular Cloud Network via Deep Reinforcement Learning","authors":"Shilin Xu, Caili Guo, R. Hu, Y. Qian","doi":"10.1109/GLOBECOM42002.2020.9348250","DOIUrl":null,"url":null,"abstract":"With the explosive growth of the computation intensive vehicular applications, the demand for computational resource in vehicular networks has increased dramatically. However some vehicular networks may be deployed in an environment that lack resource-rich facilities to support computationally expensive vehicular applications. In this work we propose a new scheme that enables computational resource sharing among vehicles in vehicular cloud network (VCN), which can be formulated as a complex multi-knapsack problem. In order to solve it, a deep reinforcement learning (DRL) algorithm is developed. Considering the non-stationary behavior brought in by the parallel learning and exploring processes among vehicles, computational resource sharing in such a vehicular network is a typical multiagent problem, therefore we model the problem with a Markov game problem. In addition, to tackle the heterogeneity property of the computational resources, a multi-hot encoding scheme is designed to standardize the action space in DRL. Furthermore, we propose a centralized training and decentralized execution framework that can be solved by a multi-agent deep deterministic policy gradient (MADDPG) algorithm. The numerical simulation results demonstrate the effectiveness of the proposed scheme.","PeriodicalId":12759,"journal":{"name":"GLOBECOM 2020 - 2020 IEEE Global Communications Conference","volume":"59 1","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"GLOBECOM 2020 - 2020 IEEE Global Communications Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GLOBECOM42002.2020.9348250","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

With the explosive growth of the computation intensive vehicular applications, the demand for computational resource in vehicular networks has increased dramatically. However some vehicular networks may be deployed in an environment that lack resource-rich facilities to support computationally expensive vehicular applications. In this work we propose a new scheme that enables computational resource sharing among vehicles in vehicular cloud network (VCN), which can be formulated as a complex multi-knapsack problem. In order to solve it, a deep reinforcement learning (DRL) algorithm is developed. Considering the non-stationary behavior brought in by the parallel learning and exploring processes among vehicles, computational resource sharing in such a vehicular network is a typical multiagent problem, therefore we model the problem with a Markov game problem. In addition, to tackle the heterogeneity property of the computational resources, a multi-hot encoding scheme is designed to standardize the action space in DRL. Furthermore, we propose a centralized training and decentralized execution framework that can be solved by a multi-agent deep deterministic policy gradient (MADDPG) algorithm. The numerical simulation results demonstrate the effectiveness of the proposed scheme.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于深度强化学习的车载云网络计算资源共享
随着计算密集型车辆应用的爆炸式增长,车载网络对计算资源的需求急剧增加。然而,一些车载网络可能部署在缺乏资源丰富的设施的环境中,以支持计算昂贵的车载应用。在这项工作中,我们提出了一种新的方案,使车辆云网络(VCN)中车辆之间的计算资源共享,可以表述为一个复杂的多背包问题。为了解决这一问题,提出了一种深度强化学习(DRL)算法。考虑到车辆间并行学习和探索过程所带来的非平稳行为,这种车辆网络中的计算资源共享是一个典型的多智能体问题,因此我们将该问题建模为马尔可夫博弈问题。此外,针对计算资源的异构性,设计了多热编码方案,对DRL中的操作空间进行了标准化。此外,我们提出了一个集中训练和分散执行的框架,该框架可以通过多智能体深度确定性策略梯度(madpg)算法来解决。数值仿真结果验证了该方案的有效性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
AirID: Injecting a Custom RF Fingerprint for Enhanced UAV Identification using Deep Learning Oversampling Algorithm based on Reinforcement Learning in Imbalanced Problems FAST-RAM: A Fast AI-assistant Solution for Task Offloading and Resource Allocation in MEC Achieving Privacy-Preserving Vehicle Selection for Effective Content Dissemination in Smart Cities Age-optimal Transmission Policy for Markov Source with Differential Encoding
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1