Deep Reinforcement Learning-Based Dynamic MultiChannel Access for Heterogeneous Wireless Networks with DenseNet

K. Zong
{"title":"Deep Reinforcement Learning-Based Dynamic MultiChannel Access for Heterogeneous Wireless Networks with DenseNet","authors":"K. Zong","doi":"10.1109/ICCCWorkshops52231.2021.9538886","DOIUrl":null,"url":null,"abstract":"In this paper, we consider the problem of dynamic multi-channel access in the heterogeneous wireless networks, where multiple independent channels are shared by multiple nodes which have different types. The objective is to find a strategy to maximize the expected long-term probability of successful transmissions. The problem of dynamic multi-channel access can be formulated as a partially observable Markov decision process (POMDP). In order to deal with this problem, we apply the deep reinforcement learning (DRL) approach to provide a model-free access method, where the nodes don’t have a prior knowledge of the wireless networks or the ability to exchange messages with other nodes. Specially, we take advantage of the double deep Q-network (DDQN) with DenseNet to learn the wireless network environment and to select the optimal channel at the beginning of each time slot. We investigate the proposed DDQN approach in different environments for both the fixed-pattern scenarios and the time-varying scenarios. The experimental results show that the proposed DDQN with DenseNet can efficiently learn the pattern of channel switch and choose the near optimal action to avoid the collision for every slot. Besides, the proposed DDQN approach can also achieve satisfactory performance to adapt the time-varying scenarios.","PeriodicalId":335240,"journal":{"name":"2021 IEEE/CIC International Conference on Communications in China (ICCC Workshops)","volume":"212 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE/CIC International Conference on Communications in China (ICCC Workshops)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCWorkshops52231.2021.9538886","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

In this paper, we consider the problem of dynamic multi-channel access in the heterogeneous wireless networks, where multiple independent channels are shared by multiple nodes which have different types. The objective is to find a strategy to maximize the expected long-term probability of successful transmissions. The problem of dynamic multi-channel access can be formulated as a partially observable Markov decision process (POMDP). In order to deal with this problem, we apply the deep reinforcement learning (DRL) approach to provide a model-free access method, where the nodes don’t have a prior knowledge of the wireless networks or the ability to exchange messages with other nodes. Specially, we take advantage of the double deep Q-network (DDQN) with DenseNet to learn the wireless network environment and to select the optimal channel at the beginning of each time slot. We investigate the proposed DDQN approach in different environments for both the fixed-pattern scenarios and the time-varying scenarios. The experimental results show that the proposed DDQN with DenseNet can efficiently learn the pattern of channel switch and choose the near optimal action to avoid the collision for every slot. Besides, the proposed DDQN approach can also achieve satisfactory performance to adapt the time-varying scenarios.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于深度强化学习的DenseNet异构无线网络动态多通道接入
本文研究了异构无线网络中多个独立信道由不同类型的多个节点共享的动态多信道接入问题。目标是找到一种策略,以最大限度地提高预期的成功传输的长期概率。动态多通道接入问题可以表示为部分可观察马尔可夫决策过程(POMDP)。为了解决这个问题,我们应用深度强化学习(DRL)方法来提供一种无模型访问方法,其中节点不具有无线网络的先验知识或与其他节点交换消息的能力。特别地,我们利用DenseNet的双深度q网络(DDQN)来学习无线网络环境,并在每个时隙的开始选择最优信道。在固定模式和时变场景下,研究了不同环境下提出的DDQN方法。实验结果表明,基于DenseNet的DDQN可以有效地学习信道切换模式,并选择接近最优的动作来避免每个槽的碰撞。此外,所提出的DDQN方法在适应时变场景时也能取得令人满意的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Link Reliability Prediction for Long-range Underwater Acoustic Communications between Gliders A Review of 3GPP Release 18 on Smart Energy and Infrastructure Analysis on Power Configuration in 5G Co-construction and Sharing Network Application of Passive Acoustic Technology in the Monitoring of Abalone’s Feeding Behavior Ultra-Compact Dual-Polarized Dipole Antenna for Ultra-Massive MIMO Systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1