Contextual Deep Reinforcement Learning for Flow and Energy Management in Wireless Sensor and IoT Networks

IF 5.3 2区 计算机科学 Q1 TELECOMMUNICATIONS IEEE Transactions on Green Communications and Networking Pub Date : 2024-01-24 DOI:10.1109/TGCN.2024.3358230
Hrishikesh Dutta;Amit Kumar Bhuyan;Subir Biswas
{"title":"Contextual Deep Reinforcement Learning for Flow and Energy Management in Wireless Sensor and IoT Networks","authors":"Hrishikesh Dutta;Amit Kumar Bhuyan;Subir Biswas","doi":"10.1109/TGCN.2024.3358230","DOIUrl":null,"url":null,"abstract":"Efficient slot allocation and transmit-sleep scheduling is an effective access control mechanism for improving communication performance and network lifetime in resource-constrained wireless networks. In this paper, a decentralized and multi-tier framework is presented for joint slot allocation and transmit-sleep scheduling in wireless network nodes with thin energy budget. The key learning objectives of this architecture are: collision-free transmission scheduling, reducing energy consumption, and improving network performance. This is achieved using a cooperative and decentralized learning behavior of multiple Reinforcement Learning (RL) agents. The resulting architecture provides throughput-sustainable support for data flows while minimizing energy expenditure and sleep-induced packet losses. To achieve this, a concept of Context is introduced to the RL framework in order to capture network traffic dynamics. The resulting Contextual Deep Q-Learning (CDQL) model makes the system adaptive to dynamic and heterogeneous network load. It also improves energy efficiency when compared with the traditional tabular Q-learning-based approaches. The results demonstrate how this framework can be used for prioritizing application-specific requirements, namely, energy saving and communication reliability. The trade-offs among packet drop, energy expenditure, and learning convergence are studied, and an application-specific solution is proposed for managing them. The performance is compared against an existing state-of-the-art scheduling approach. Moreover, an analytical model of the system dynamics is developed and validated using simulation for arbitrary mesh topologies and traffic patterns.","PeriodicalId":13052,"journal":{"name":"IEEE Transactions on Green Communications and Networking","volume":null,"pages":null},"PeriodicalIF":5.3000,"publicationDate":"2024-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Green Communications and Networking","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10413520/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

Efficient slot allocation and transmit-sleep scheduling is an effective access control mechanism for improving communication performance and network lifetime in resource-constrained wireless networks. In this paper, a decentralized and multi-tier framework is presented for joint slot allocation and transmit-sleep scheduling in wireless network nodes with thin energy budget. The key learning objectives of this architecture are: collision-free transmission scheduling, reducing energy consumption, and improving network performance. This is achieved using a cooperative and decentralized learning behavior of multiple Reinforcement Learning (RL) agents. The resulting architecture provides throughput-sustainable support for data flows while minimizing energy expenditure and sleep-induced packet losses. To achieve this, a concept of Context is introduced to the RL framework in order to capture network traffic dynamics. The resulting Contextual Deep Q-Learning (CDQL) model makes the system adaptive to dynamic and heterogeneous network load. It also improves energy efficiency when compared with the traditional tabular Q-learning-based approaches. The results demonstrate how this framework can be used for prioritizing application-specific requirements, namely, energy saving and communication reliability. The trade-offs among packet drop, energy expenditure, and learning convergence are studied, and an application-specific solution is proposed for managing them. The performance is compared against an existing state-of-the-art scheduling approach. Moreover, an analytical model of the system dynamics is developed and validated using simulation for arbitrary mesh topologies and traffic patterns.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
针对无线传感器和物联网网络中流量和能量管理的情境深度强化学习
在资源受限的无线网络中,高效的时隙分配和发送-休眠调度是提高通信性能和网络寿命的有效访问控制机制。本文提出了一种去中心化的多层框架,用于在能量预算较低的无线网络节点中进行联合时隙分配和发送-休眠调度。该架构的主要学习目标是:无碰撞传输调度、降低能耗和提高网络性能。这是通过多个强化学习(RL)代理的合作和分散学习行为来实现的。由此产生的架构可为数据流提供吞吐量可持续的支持,同时最大限度地减少能源消耗和睡眠引起的数据包丢失。为此,RL 框架引入了 "情境 "概念,以捕捉网络流量动态。由此产生的上下文深度 Q 学习(CDQL)模型使系统能够适应动态和异构网络负载。与传统的基于表格的 Q 学习方法相比,它还提高了能效。研究结果表明,该框架可用于优先满足特定应用的要求,即节能和通信可靠性。研究了丢包、能量消耗和学习收敛之间的权衡,并提出了管理这些问题的特定应用解决方案。将其性能与现有的最先进调度方法进行了比较。此外,还开发了一个系统动态分析模型,并针对任意网状拓扑和流量模式进行了仿真验证。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Transactions on Green Communications and Networking
IEEE Transactions on Green Communications and Networking Computer Science-Computer Networks and Communications
CiteScore
9.30
自引率
6.20%
发文量
181
期刊最新文献
Table of Contents Guest Editorial Special Issue on Green Open Radio Access Networks: Architecture, Challenges, Opportunities, and Use Cases IEEE Transactions on Green Communications and Networking IEEE Communications Society Information HSADR: A New Highly Secure Aggregation and Dropout-Resilient Federated Learning Scheme for Radio Access Networks With Edge Computing Systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1