Green Computation Offloading With DRL in Multi-Access Edge Computing

IF 2.5 4区 计算机科学 Q3 TELECOMMUNICATIONS Transactions on Emerging Telecommunications Technologies Pub Date : 2024-11-08 DOI:10.1002/ett.70003
Changkui Yin, Yingchi Mao, Meng Chen, Yi Rong, Yinqiu Liu, Xiaoming He
{"title":"Green Computation Offloading With DRL in Multi-Access Edge Computing","authors":"Changkui Yin,&nbsp;Yingchi Mao,&nbsp;Meng Chen,&nbsp;Yi Rong,&nbsp;Yinqiu Liu,&nbsp;Xiaoming He","doi":"10.1002/ett.70003","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>In multi-access edge computing (MEC), computational task offloading of mobile terminals (MT) is expected to provide the green applications with the restriction of energy consumption and service latency. Nevertheless, the diverse statuses of a range of edge servers and mobile terminals, along with the fluctuating offloading routes, present a challenge in the realm of computational task offloading. In order to bolster green applications, we present an innovative computational task offloading model as our initial approach. In particular, the nascent model is constrained by energy consumption and service latency considerations: (1) Smart mobile terminals with computational capabilities could serve as carriers; (2) The diverse computational and communication capacities of edge servers have the potential to enhance the offloading process; (3) The unpredictable routing paths of mobile terminals and edge servers could result in varied information transmissions. We then propose an improved deep reinforcement learning (DRL) algorithm named PS-DDPG with the prioritized experience replay (PER) and the stochastic weight averaging (SWA) mechanisms based on deep deterministic policy gradients (DDPG) to seek an optimal offloading mode, saving energy consumption. Next, we introduce an enhanced deep reinforcement learning (DRL) algorithm named PS-DDPG, incorporating the prioritized experience replay (PER) and stochastic weight averaging (SWA) techniques rooted in deep deterministic policy gradients (DDPG). This approach aims to identify an efficient offloading strategy, thereby reducing energy consumption. Fortunately, <span></span><math></math> algorithm is proposed for each MT, which is responsible for making decisions regarding task partition, channel allocation, and power transmission control. Our developed approach achieves the ultimate estimation of observed values and enhances memory via write operations. The replay buffer holds data from previous <span></span><math></math> time slots to upgrade both the actor and critic networks, followed by a buffer reset. Comprehensive experiments validate the superior performance, including stability and convergence, of our algorithm when juxtaposed with prior studies.</p>\n </div>","PeriodicalId":23282,"journal":{"name":"Transactions on Emerging Telecommunications Technologies","volume":"35 11","pages":""},"PeriodicalIF":2.5000,"publicationDate":"2024-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Transactions on Emerging Telecommunications Technologies","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/ett.70003","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

In multi-access edge computing (MEC), computational task offloading of mobile terminals (MT) is expected to provide the green applications with the restriction of energy consumption and service latency. Nevertheless, the diverse statuses of a range of edge servers and mobile terminals, along with the fluctuating offloading routes, present a challenge in the realm of computational task offloading. In order to bolster green applications, we present an innovative computational task offloading model as our initial approach. In particular, the nascent model is constrained by energy consumption and service latency considerations: (1) Smart mobile terminals with computational capabilities could serve as carriers; (2) The diverse computational and communication capacities of edge servers have the potential to enhance the offloading process; (3) The unpredictable routing paths of mobile terminals and edge servers could result in varied information transmissions. We then propose an improved deep reinforcement learning (DRL) algorithm named PS-DDPG with the prioritized experience replay (PER) and the stochastic weight averaging (SWA) mechanisms based on deep deterministic policy gradients (DDPG) to seek an optimal offloading mode, saving energy consumption. Next, we introduce an enhanced deep reinforcement learning (DRL) algorithm named PS-DDPG, incorporating the prioritized experience replay (PER) and stochastic weight averaging (SWA) techniques rooted in deep deterministic policy gradients (DDPG). This approach aims to identify an efficient offloading strategy, thereby reducing energy consumption. Fortunately, algorithm is proposed for each MT, which is responsible for making decisions regarding task partition, channel allocation, and power transmission control. Our developed approach achieves the ultimate estimation of observed values and enhances memory via write operations. The replay buffer holds data from previous time slots to upgrade both the actor and critic networks, followed by a buffer reset. Comprehensive experiments validate the superior performance, including stability and convergence, of our algorithm when juxtaposed with prior studies.

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
在多接入边缘计算中利用 DRL 实现绿色计算卸载
在多接入边缘计算(MEC)中,移动终端(MT)的计算任务卸载有望为绿色应用提供能耗和服务延迟限制。然而,一系列边缘服务器和移动终端的状态各不相同,而且卸载路线也不固定,这给计算任务卸载领域带来了挑战。为了支持绿色应用,我们提出了一种创新的计算任务卸载模型作为初步方法。具体而言,该新生模型受到能耗和服务延迟因素的制约:(1)具有计算能力的智能移动终端可作为载体;(2)边缘服务器的不同计算和通信能力有可能增强卸载过程;(3)移动终端和边缘服务器的路由路径不可预测,可能导致信息传输的变化。因此,我们提出了一种名为 PS-DDPG 的改进型深度强化学习(DRL)算法,该算法具有基于深度确定性策略梯度(DDPG)的优先经验重放(PER)和随机权重平均(SWA)机制,可寻求最佳卸载模式,从而节省能耗。接下来,我们介绍了一种名为 PS-DDPG 的增强型深度强化学习(DRL)算法,它将优先经验重放(PER)和随机权重平均(SWA)技术融入了深度确定性策略梯度(DDPG)。这种方法旨在确定有效的卸载策略,从而降低能耗。幸运的是,我们为每个 MT 提出了算法,由其负责做出任务分区、信道分配和功率传输控制方面的决策。我们开发的方法实现了对观测值的最终估算,并通过写操作增强了内存。重放缓冲区保存前一时间段的数据,以升级行动者和批评者网络,然后重置缓冲区。综合实验验证了我们的算法在稳定性和收敛性等方面的优越性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
8.90
自引率
13.90%
发文量
249
期刊介绍: ransactions on Emerging Telecommunications Technologies (ETT), formerly known as European Transactions on Telecommunications (ETT), has the following aims: - to attract cutting-edge publications from leading researchers and research groups around the world - to become a highly cited source of timely research findings in emerging fields of telecommunications - to limit revision and publication cycles to a few months and thus significantly increase attractiveness to publish - to become the leading journal for publishing the latest developments in telecommunications
期刊最新文献
Issue Information Research and Implementation of a Classification Method of Industrial Big Data for Security Management Moving Target Detection in Clutter Environment Based on Track Posture Hypothesis Testing Spiking Quantum Fire Hawk Network Based Reliable Scheduling for Lifetime Maximization of Wireless Sensor Network Optimized Data Replication in Cloud Using Hybrid Optimization Approach
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1