On the Training of Reinforcement Learning-based Algorithms in 5G and Beyond Radio Access Networks

Irene Vilà Muñoz, J. Pérez-Romero, O. Sallent
{"title":"On the Training of Reinforcement Learning-based Algorithms in 5G and Beyond Radio Access Networks","authors":"Irene Vilà Muñoz, J. Pérez-Romero, O. Sallent","doi":"10.1109/NetSoft54395.2022.9844032","DOIUrl":null,"url":null,"abstract":"Reinforcement Learning (RL)-based algorithmic solutions have been profusely proposed in recent years for addressing multiple problems in the Radio Access Network (RAN). However, how RL algorithms have to be trained for a successful exploitation has not received sufficient attention. To address this limitation, which is particularly relevant given the peculiarities of wireless communications, this paper proposes a functional framework for training RL strategies in the RAN. The framework is aligned with the O-RAN Alliance machine learning workflow and introduces specific functionalities for RL, such as the way of specifying the training datasets, the mechanisms to monitor the performance of the trained policies during inference in the real network, and the capability to conduct a retraining if necessary. The proposed framework is illustrated with a relevant use case in 5G, namely RAN slicing, by considering a Deep Q-Network algorithm for capacity sharing. Finally, insights on other possible applicability examples of the proposed framework are provided.","PeriodicalId":125799,"journal":{"name":"2022 IEEE 8th International Conference on Network Softwarization (NetSoft)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 8th International Conference on Network Softwarization (NetSoft)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NetSoft54395.2022.9844032","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Reinforcement Learning (RL)-based algorithmic solutions have been profusely proposed in recent years for addressing multiple problems in the Radio Access Network (RAN). However, how RL algorithms have to be trained for a successful exploitation has not received sufficient attention. To address this limitation, which is particularly relevant given the peculiarities of wireless communications, this paper proposes a functional framework for training RL strategies in the RAN. The framework is aligned with the O-RAN Alliance machine learning workflow and introduces specific functionalities for RL, such as the way of specifying the training datasets, the mechanisms to monitor the performance of the trained policies during inference in the real network, and the capability to conduct a retraining if necessary. The proposed framework is illustrated with a relevant use case in 5G, namely RAN slicing, by considering a Deep Q-Network algorithm for capacity sharing. Finally, insights on other possible applicability examples of the proposed framework are provided.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
5G及以上无线接入网络中基于强化学习的算法训练研究
近年来,基于强化学习(RL)的算法解决方案被大量提出,用于解决无线接入网(RAN)中的多种问题。然而,如何训练强化学习算法才能成功利用还没有得到足够的重视。考虑到无线通信的特殊性,为了解决这一限制,本文提出了一个在RAN中训练强化学习策略的功能框架。该框架与O-RAN联盟机器学习工作流程保持一致,并为强化学习引入了特定的功能,例如指定训练数据集的方式,在真实网络中进行推理期间监控训练策略性能的机制,以及在必要时进行再训练的能力。通过考虑用于容量共享的Deep Q-Network算法,用5G中的相关用例(即RAN切片)说明了所提出的框架。最后,对所提出的框架的其他可能的适用性示例提供了见解。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Flexible Measurement Testbed for Evaluating Time-Sensitive Networking in Industrial Automation Applications Latency-aware Topology Discovery in SDN-based Time-Sensitive Networks NLP4: An Architecture for Intent-Driven Data Plane Programmability CHIMA: a Framework for Network Services Deployment and Performance Assurance Encrypted Network Traffic Classification in SDN using Self-supervised Learning
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1