{"title":"Multi-Agent Reinforcement Learning for Multi-Cell Spectrum and Power Allocation","authors":"Yiming Zhang;Dongning Guo","doi":"10.1109/TCOMM.2025.3534565","DOIUrl":null,"url":null,"abstract":"Efficient and scalable radio resource allocation is essential for the success of wireless cellular networks. This paper presents a fully scalable multi-agent reinforcement learning (MARL) framework, where each agent manages spectrum, power allocation, and scheduling within a cell, using only locally available information. The objective is to minimize packet delays under stochastic traffic arrivals, applicable to both conflict graph models and cellular network configurations. This is formulated as a distributed learning problem and implemented using a multi-agent proximal policy optimization (MAPPO) algorithm. This traffic-driven MARL approach enables fully decentralized training and execution, ensuring scalability to arbitrarily large networks. Extensive simulations demonstrate that the proposed methods achieve quality of service (QoS) performance comparable to centralized algorithms that require global information, while the trained policies show robust scalability across diverse network sizes and traffic conditions.","PeriodicalId":13041,"journal":{"name":"IEEE Transactions on Communications","volume":"73 8","pages":"5980-5992"},"PeriodicalIF":8.3000,"publicationDate":"2025-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Communications","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10854577/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Efficient and scalable radio resource allocation is essential for the success of wireless cellular networks. This paper presents a fully scalable multi-agent reinforcement learning (MARL) framework, where each agent manages spectrum, power allocation, and scheduling within a cell, using only locally available information. The objective is to minimize packet delays under stochastic traffic arrivals, applicable to both conflict graph models and cellular network configurations. This is formulated as a distributed learning problem and implemented using a multi-agent proximal policy optimization (MAPPO) algorithm. This traffic-driven MARL approach enables fully decentralized training and execution, ensuring scalability to arbitrarily large networks. Extensive simulations demonstrate that the proposed methods achieve quality of service (QoS) performance comparable to centralized algorithms that require global information, while the trained policies show robust scalability across diverse network sizes and traffic conditions.
期刊介绍:
The IEEE Transactions on Communications is dedicated to publishing high-quality manuscripts that showcase advancements in the state-of-the-art of telecommunications. Our scope encompasses all aspects of telecommunications, including telephone, telegraphy, facsimile, and television, facilitated by electromagnetic propagation methods such as radio, wire, aerial, underground, coaxial, and submarine cables, as well as waveguides, communication satellites, and lasers. We cover telecommunications in various settings, including marine, aeronautical, space, and fixed station services, addressing topics such as repeaters, radio relaying, signal storage, regeneration, error detection and correction, multiplexing, carrier techniques, communication switching systems, data communications, and communication theory. Join us in advancing the field of telecommunications through groundbreaking research and innovation.