Pub Date : 2025-03-30DOI: 10.1109/TMLCN.2025.3575368
Kazi Hasan;Khaleda Papry;Thomas Trappenberg;Israat Haque
Radio Link Failure (RLF) prediction system in Radio Access Networks (RANs) is critical for ensuring seamless communication and meeting the stringent requirements of high data rates, low latency, and improved reliability in 5G networks. However, weather conditions such as precipitation, humidity, temperature, and wind impact these communication links. Usually, historical radio link Key Performance Indicators (KPIs) and their surrounding weather station observations are utilized for building learning-based RLF prediction models. However, such models must be capable of learning the spatial weather context in a dynamic RAN and effectively encoding time series KPIs with the weather observation data. Existing work utilizes a heuristic-based and non-generalizable weather station aggregation method that uses Long Short-Term Memory (LSTM) for non-weighted sequence modeling. This paper fills the gap by proposing GenTrap, a novel RLF prediction framework that introduces a Graph Neural Network (GNN)-based learnable weather effect aggregation module and employs state-of-the-art time series transformer as the temporal feature extractor for radio link failure prediction. The GNN module encodes surrounding weather station data of each radio site while the transformer module encodes historical radio and weather observation features. The proposed aggregation method of GenTrap can be integrated into any existing prediction model to achieve better performance and generalizability. We evaluate GenTrap on two real-world datasets (rural and urban) with 2.6 million KPI data points and show that GenTrap offers a significantly higher F1-score of 0.93 for rural and 0.79 for urban, an increase of 29% and 21% respectively, compared to the state-of-the-art LSTM-based solutions while offering a 20% increased generalization capability.
无线接入网(RANs)中的RLF (Radio Link Failure)预测系统对于确保无缝通信,满足5G网络对高数据速率、低延迟和高可靠性的严格要求至关重要。然而,诸如降水、湿度、温度和风等天气条件会影响这些通信链路。通常,利用历史无线电链路关键性能指标(kpi)及其周围气象站观测数据建立基于学习的RLF预测模型。然而,这种模型必须能够在动态RAN中学习空间天气环境,并有效地用天气观测数据编码时间序列kpi。现有工作采用基于启发式的非一般化气象站聚合方法,该方法使用长短期记忆(LSTM)进行非加权序列建模。本文提出了一种新的RLF预测框架GenTrap,该框架引入了基于图神经网络(GNN)的可学习天气效应聚合模块,并采用最先进的时间序列变压器作为无线电链路故障预测的时间特征提取器,填补了这一空白。GNN模块对每个无线电站点周围气象站数据进行编码,变压器模块对历史无线电和天气观测特征进行编码。所提出的GenTrap聚合方法可以集成到任何现有的预测模型中,以获得更好的性能和泛化性。我们在两个真实世界的数据集(农村和城市)上使用260万KPI数据点对GenTrap进行了评估,结果表明,与最先进的基于lstm的解决方案相比,GenTrap在农村和城市的f1得分分别为0.93和0.79,分别提高了29%和21%,同时泛化能力提高了20%。
{"title":"A Generalized GNN-Transformer-Based Radio Link Failure Prediction Framework in 5G RAN","authors":"Kazi Hasan;Khaleda Papry;Thomas Trappenberg;Israat Haque","doi":"10.1109/TMLCN.2025.3575368","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3575368","url":null,"abstract":"Radio Link Failure (RLF) prediction system in Radio Access Networks (RANs) is critical for ensuring seamless communication and meeting the stringent requirements of high data rates, low latency, and improved reliability in 5G networks. However, weather conditions such as precipitation, humidity, temperature, and wind impact these communication links. Usually, historical radio link Key Performance Indicators (KPIs) and their surrounding weather station observations are utilized for building learning-based RLF prediction models. However, such models must be capable of learning the spatial weather context in a dynamic RAN and effectively encoding time series KPIs with the weather observation data. Existing work utilizes a heuristic-based and non-generalizable weather station aggregation method that uses Long Short-Term Memory (LSTM) for non-weighted sequence modeling. This paper fills the gap by proposing GenTrap, a novel RLF prediction framework that introduces a Graph Neural Network (GNN)-based learnable weather effect aggregation module and employs state-of-the-art time series transformer as the temporal feature extractor for radio link failure prediction. The GNN module encodes surrounding weather station data of each radio site while the transformer module encodes historical radio and weather observation features. The proposed aggregation method of GenTrap can be integrated into any existing prediction model to achieve better performance and generalizability. We evaluate GenTrap on two real-world datasets (rural and urban) with 2.6 million KPI data points and show that GenTrap offers a significantly higher F1-score of 0.93 for rural and 0.79 for urban, an increase of 29% and 21% respectively, compared to the state-of-the-art LSTM-based solutions while offering a 20% increased generalization capability.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"710-724"},"PeriodicalIF":0.0,"publicationDate":"2025-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11018489","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144308387","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-20DOI: 10.1109/TMLCN.2025.3553100
Berend J. D. Gort;Godfrey M. Kibalya;Angelos Antonopoulos
Effective resource management in edge-cloud networks is crucial for meeting Quality of Service (QoS) requirements while minimizing operational costs. However, dynamic and fluctuating workloads pose significant challenges for accurate workload prediction and efficient resource allocation, particularly in resource-constrained edge environments. In this paper, we introduce AERO (Adaptive Edge-cloud Resource Orchestration), a novel lightweight forecasting model designed to address these challenges. AERO features an adaptive period detection mechanism that dynamically identifies dominant periodicities in multivariate workload data, allowing it to adjust to varying patterns and abrupt changes. With fewer than 1,000 parameters, AERO is highly suitable for deployment on edge devices with limited computational capacity. We formalize our approach through a comprehensive system model and extend an existing simulation framework with predictor modules to evaluate AERO’s performance in realistic cloud-edge environments. Our extensive evaluations on real-world cloud workload datasets demonstrate that AERO achieves comparable prediction accuracy to complex state-of-the-art models with millions of parameters, while significantly reducing model size and computational overhead. In addition, simulations show that AERO improves orchestration performance, reducing energy consumption and response times compared to existing proactive and reactive approaches. Our live deployment experiments further validate these findings, demonstrating that AERO consistently delivers superior performance. These results highlight AERO as an effective solution for improving resource management and reducing operational costs in dynamic cloud-edge environments.
{"title":"AERO: Adaptive Edge-Cloud Orchestration With a Sub-1K-Parameter Forecasting Model","authors":"Berend J. D. Gort;Godfrey M. Kibalya;Angelos Antonopoulos","doi":"10.1109/TMLCN.2025.3553100","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3553100","url":null,"abstract":"Effective resource management in edge-cloud networks is crucial for meeting Quality of Service (QoS) requirements while minimizing operational costs. However, dynamic and fluctuating workloads pose significant challenges for accurate workload prediction and efficient resource allocation, particularly in resource-constrained edge environments. In this paper, we introduce AERO (Adaptive Edge-cloud Resource Orchestration), a novel lightweight forecasting model designed to address these challenges. AERO features an adaptive period detection mechanism that dynamically identifies dominant periodicities in multivariate workload data, allowing it to adjust to varying patterns and abrupt changes. With fewer than 1,000 parameters, AERO is highly suitable for deployment on edge devices with limited computational capacity. We formalize our approach through a comprehensive system model and extend an existing simulation framework with predictor modules to evaluate AERO’s performance in realistic cloud-edge environments. Our extensive evaluations on real-world cloud workload datasets demonstrate that AERO achieves comparable prediction accuracy to complex state-of-the-art models with millions of parameters, while significantly reducing model size and computational overhead. In addition, simulations show that AERO improves orchestration performance, reducing energy consumption and response times compared to existing proactive and reactive approaches. Our live deployment experiments further validate these findings, demonstrating that AERO consistently delivers superior performance. These results highlight AERO as an effective solution for improving resource management and reducing operational costs in dynamic cloud-edge environments.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"463-478"},"PeriodicalIF":0.0,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10935743","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143740375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-19DOI: 10.1109/TMLCN.2025.3571026
Michael Baur;Nurettin Turan;Simon Wallner;Wolfgang Utschick
Generative models are typically evaluated by direct inspection of their generated samples, e.g., by visual inspection in the case of images. Further evaluation metrics like the Fréchet inception distance or maximum mean discrepancy are intricate to interpret and lack physical motivation. These observations make evaluating generative models in the wireless PHY layer non-trivial. This work establishes a framework consisting of evaluation metrics and methods for generative models applied to the wireless PHY layer. The proposed metrics and methods are motivated by wireless applications, facilitating interpretation and understandability for the wireless community. In particular, we propose a spectral efficiency analysis for validating the generated channel norms and a codebook fingerprinting method to validate the generated channel directions. Moreover, we propose an application cross-check to evaluate the generative model’s samples for training machine learning-based models in relevant downstream tasks. Our analysis is based on real-world measurement data and includes the Gaussian mixture model, variational autoencoder, diffusion model, and generative adversarial network. Our results indicate that solely relying on metrics like the maximum mean discrepancy produces inconsistent and uninterpretable evaluation outcomes. In contrast, the proposed metrics and methods exhibit consistent and explainable behavior.
{"title":"Evaluation Metrics and Methods for Generative Models in the Wireless PHY Layer","authors":"Michael Baur;Nurettin Turan;Simon Wallner;Wolfgang Utschick","doi":"10.1109/TMLCN.2025.3571026","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3571026","url":null,"abstract":"Generative models are typically evaluated by direct inspection of their generated samples, e.g., by visual inspection in the case of images. Further evaluation metrics like the Fréchet inception distance or maximum mean discrepancy are intricate to interpret and lack physical motivation. These observations make evaluating generative models in the wireless PHY layer non-trivial. This work establishes a framework consisting of evaluation metrics and methods for generative models applied to the wireless PHY layer. The proposed metrics and methods are motivated by wireless applications, facilitating interpretation and understandability for the wireless community. In particular, we propose a spectral efficiency analysis for validating the generated channel norms and a codebook fingerprinting method to validate the generated channel directions. Moreover, we propose an application cross-check to evaluate the generative model’s samples for training machine learning-based models in relevant downstream tasks. Our analysis is based on real-world measurement data and includes the Gaussian mixture model, variational autoencoder, diffusion model, and generative adversarial network. Our results indicate that solely relying on metrics like the maximum mean discrepancy produces inconsistent and uninterpretable evaluation outcomes. In contrast, the proposed metrics and methods exhibit consistent and explainable behavior.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"677-689"},"PeriodicalIF":0.0,"publicationDate":"2025-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11007069","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144219745","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-18DOI: 10.1109/TMLCN.2025.3551689
Sepideh Afshar;Reza Razavi;Mohammad Moshirpour
Accurate throughput forecasting is essential for ensuring the seamless operation of Real-Time Communication (RTC) applications. These demands for accurate throughput forecasting become particularly challenging when dealing with wireless access links, as they inherently exhibit fluctuating bandwidth. Ensuring an exceptional user Quality of Experience (QoE) in this scenario depends on accurately predicting available bandwidth in the short term since it plays a pivotal role in guiding video rate adaptation. Yet, current methodologies for short-term bandwidth prediction (SBP) struggle to perform adequately in dynamically changing real-world network environments and lack generalizability to adapt across varied network conditions. Also, acquiring long and representative traces that capture real-world network complexity is challenging. To overcome these challenges, we propose closed-loop clustering-based Global Forecasting Models (GFMs) for SBP. Unlike local models, GFMs apply the same function to all traces enabling cross-learning, and leveraging relationships among traces to address the performance issues seen in current SBP algorithms. To address potential heterogeneity within the data and improve prediction quality, a clustered-wise GFM is utilized to group similar traces based on prediction accuracy. Finally, the proposed method is validated using real-world datasets of HSDPA 3G, NYC LTE, and Irish 5G data demonstrating significant improvements in accuracy and generalizability.
{"title":"Closed-Loop Clustering-Based Global Bandwidth Prediction in Real-Time Video Streaming","authors":"Sepideh Afshar;Reza Razavi;Mohammad Moshirpour","doi":"10.1109/TMLCN.2025.3551689","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3551689","url":null,"abstract":"Accurate throughput forecasting is essential for ensuring the seamless operation of Real-Time Communication (RTC) applications. These demands for accurate throughput forecasting become particularly challenging when dealing with wireless access links, as they inherently exhibit fluctuating bandwidth. Ensuring an exceptional user Quality of Experience (QoE) in this scenario depends on accurately predicting available bandwidth in the short term since it plays a pivotal role in guiding video rate adaptation. Yet, current methodologies for short-term bandwidth prediction (SBP) struggle to perform adequately in dynamically changing real-world network environments and lack generalizability to adapt across varied network conditions. Also, acquiring long and representative traces that capture real-world network complexity is challenging. To overcome these challenges, we propose closed-loop clustering-based Global Forecasting Models (GFMs) for SBP. Unlike local models, GFMs apply the same function to all traces enabling cross-learning, and leveraging relationships among traces to address the performance issues seen in current SBP algorithms. To address potential heterogeneity within the data and improve prediction quality, a clustered-wise GFM is utilized to group similar traces based on prediction accuracy. Finally, the proposed method is validated using real-world datasets of HSDPA 3G, NYC LTE, and Irish 5G data demonstrating significant improvements in accuracy and generalizability.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"448-462"},"PeriodicalIF":0.0,"publicationDate":"2025-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10929655","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143716486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-10DOI: 10.1109/TMLCN.2025.3550119
Ce Feng;Parv Venkitasubramaniam
The increasing adoption of machine learning at the edge (ML-at-the-edge) and federated learning (FL) presents a dual challenge: ensuring data privacy as well as addressing resource constraints such as limited computational power, memory, and communication bandwidth. Traditional approaches typically apply differentially private stochastic gradient descent (DP-SGD) to preserve privacy, followed by quantization techniques as a post-processing step to reduce model size and communication overhead. However, this sequential framework introduces inherent drawbacks, as quantization alone lacks privacy guarantees and often introduces errors that degrade model performance. In this work, we propose randomized quantization as an integrated solution to address these dual challenges by embedding randomness directly into the quantization process. This approach enhances privacy while simultaneously reducing communication and computational overhead. To achieve this, we introduce Randomized Quantizer Projection Stochastic Gradient Descent (RQP-SGD), a method designed for ML-at-the-edge that embeds DP-SGD within a randomized quantization-based projection during model training. For federated learning, we develop Gaussian Sampling Quantization (GSQ), which integrates discrete Gaussian sampling into the quantization process to ensure local differential privacy (LDP). Unlike conventional methods that rely on Gaussian noise addition, GSQ achieves privacy through discrete Gaussian sampling while improving communication efficiency and model utility across distributed systems. Through rigorous theoretical analysis and extensive experiments on benchmark datasets, we demonstrate that these methods significantly enhance the utility-privacy trade-off and computational efficiency in both ML-at-the-edge and FL systems. RQP-SGD is evaluated on MNIST and the Breast Cancer Diagnostic dataset, showing an average 10.62% utility improvement over the deterministic quantization-based projected DP-SGD while maintaining (1.0, 0)-DP. In federated learning tasks, GSQ-FL improves accuracy by an average 11.52% over DP-FedPAQ across MNIST and FashionMNIST under non-IID conditions. Additionally, GSQ-FL outperforms DP-FedPAQ by 16.54% on CIFAR-10 and 8.7% on FEMNIST.
{"title":"Randomized Quantization for Privacy in Resource Constrained Machine Learning at-the-Edge and Federated Learning","authors":"Ce Feng;Parv Venkitasubramaniam","doi":"10.1109/TMLCN.2025.3550119","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3550119","url":null,"abstract":"The increasing adoption of machine learning at the edge (ML-at-the-edge) and federated learning (FL) presents a dual challenge: ensuring data privacy as well as addressing resource constraints such as limited computational power, memory, and communication bandwidth. Traditional approaches typically apply differentially private stochastic gradient descent (DP-SGD) to preserve privacy, followed by quantization techniques as a post-processing step to reduce model size and communication overhead. However, this sequential framework introduces inherent drawbacks, as quantization alone lacks privacy guarantees and often introduces errors that degrade model performance. In this work, we propose randomized quantization as an integrated solution to address these dual challenges by embedding randomness directly into the quantization process. This approach enhances privacy while simultaneously reducing communication and computational overhead. To achieve this, we introduce Randomized Quantizer Projection Stochastic Gradient Descent (RQP-SGD), a method designed for ML-at-the-edge that embeds DP-SGD within a randomized quantization-based projection during model training. For federated learning, we develop Gaussian Sampling Quantization (GSQ), which integrates discrete Gaussian sampling into the quantization process to ensure local differential privacy (LDP). Unlike conventional methods that rely on Gaussian noise addition, GSQ achieves privacy through discrete Gaussian sampling while improving communication efficiency and model utility across distributed systems. Through rigorous theoretical analysis and extensive experiments on benchmark datasets, we demonstrate that these methods significantly enhance the utility-privacy trade-off and computational efficiency in both ML-at-the-edge and FL systems. RQP-SGD is evaluated on MNIST and the Breast Cancer Diagnostic dataset, showing an average 10.62% utility improvement over the deterministic quantization-based projected DP-SGD while maintaining (1.0, 0)-DP. In federated learning tasks, GSQ-FL improves accuracy by an average 11.52% over DP-FedPAQ across MNIST and FashionMNIST under non-IID conditions. Additionally, GSQ-FL outperforms DP-FedPAQ by 16.54% on CIFAR-10 and 8.7% on FEMNIST.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"395-419"},"PeriodicalIF":0.0,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10919124","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143645337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-03-06DOI: 10.1109/TMLCN.2025.3567370
Seda Dogan-Tusha;Faissal El Bouanani;Marwa Qaraqe
Federated Learning (FL) has attracted the interest of researchers since it hinders inefficient resource utilization by developing a global learning model based on local model parameters (LMP). This study introduces a novel optimal stopping theory (OST) based online node selection scheme for low complex and multi-parameter FL procedure in IoT networks. Global model accuracy (GMA) in FL depends on the accuracy of the LMP received by the central entity (CE). It is therefore essential to choose trusty nodes to guarantee a certain level of global model accuracy without inducing additional system complexity. For this reason, the proposed technique in this study utilizes the secretary problem (SP) approach as an OST to perform node selection considering both received signal strength (RSS) and local model accuracy (LMA) of available nodes. By leveraging the SP, the proposed technique employs a stopping rule that maximizes the probability of selecting the node with the best quality, and thereby avoids testing all candidate nodes. To this end, this work provides a mathematical framework for maximizing the selection probability of the best node amongst candidate nodes. Specifically, the developed framework has been used to calculate the weighting coefficients of the RSS and LMA to define the node quality. Comprehensive analysis and simulation results illustrate that the OST based proposed technique outperforms state-of-the-art methods including the random node selection and the offline node selection (exhaustive search methods) in terms of GMA and computational complexity, respectively.
{"title":"Optimal Stopping Theory-Based Online Node Selection in IoT Networks for Multi-Parameter Federated Learning","authors":"Seda Dogan-Tusha;Faissal El Bouanani;Marwa Qaraqe","doi":"10.1109/TMLCN.2025.3567370","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3567370","url":null,"abstract":"Federated Learning (FL) has attracted the interest of researchers since it hinders inefficient resource utilization by developing a global learning model based on local model parameters (LMP). This study introduces a novel optimal stopping theory (OST) based online node selection scheme for low complex and multi-parameter FL procedure in IoT networks. Global model accuracy (GMA) in FL depends on the accuracy of the LMP received by the central entity (CE). It is therefore essential to choose trusty nodes to guarantee a certain level of global model accuracy without inducing additional system complexity. For this reason, the proposed technique in this study utilizes the secretary problem (SP) approach as an OST to perform node selection considering both received signal strength (RSS) and local model accuracy (LMA) of available nodes. By leveraging the SP, the proposed technique employs a stopping rule that maximizes the probability of selecting the node with the best quality, and thereby avoids testing all candidate nodes. To this end, this work provides a mathematical framework for maximizing the selection probability of the best node amongst candidate nodes. Specifically, the developed framework has been used to calculate the weighting coefficients of the RSS and LMA to define the node quality. Comprehensive analysis and simulation results illustrate that the OST based proposed technique outperforms state-of-the-art methods including the random node selection and the offline node selection (exhaustive search methods) in terms of GMA and computational complexity, respectively.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"659-676"},"PeriodicalIF":0.0,"publicationDate":"2025-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10988901","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144100011","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In massive MIMO systems, achieving optimal end-to-end transmission encompasses various aspects such as power control, modulation schemes, path selection, and accurate channel estimation. Nonetheless, optimizing resource allocation remains a significant challenge. In path selection, the direct link is a straightforward link between the transmitter and the receiver. On the other hand, the indirect link involves reflections, diffraction, or scattering, often due to interactions with objects or obstacles. Relying exclusively on one type of link can lead to suboptimal and limited performance. Link management (LM) is emerging as a viable solution, and accurate channel estimation provides essential information to make informed decisions about transmission parameters. In this paper, we study LM and channel estimation that flexibly adjust the transmission ratio of direct and indirect links to improve generalization, using a denoising variational autoencoder with attention modules (DVAE-ATT) to enhance sum rate. Our experiments show significant improvements in IRS-assisted millimeter-wave MIMO systems. Incorporating LM increased the sum rate and reduced MSE by approximately 9%. Variational autoencoders (VAE) outperformed traditional autoencoders in the spatial domain, as confirmed by heatmap analysis. Additionally, our investigation of DVAE-ATT reveals notable differences in the temporal domain with and without attention mechanisms. Finally, we analyze performance across varying numbers of users and ranges. Across various distances—5m, 15m, 25m, and 35m—performance improvements averaged 6%, 11%, 16%, and 22%, respectively.
{"title":"Paths Optimization by Jointing Link Management and Channel Estimation Using Variational Autoencoder With Attention for IRS-MIMO Systems","authors":"Meng-Hsun Wu;Hong-Yunn Chen;Ta-Wei Yang;Chih-Chuan Hsu;Chih-Wei Huang;Cheng-Fu Chou","doi":"10.1109/TMLCN.2025.3547689","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3547689","url":null,"abstract":"In massive MIMO systems, achieving optimal end-to-end transmission encompasses various aspects such as power control, modulation schemes, path selection, and accurate channel estimation. Nonetheless, optimizing resource allocation remains a significant challenge. In path selection, the direct link is a straightforward link between the transmitter and the receiver. On the other hand, the indirect link involves reflections, diffraction, or scattering, often due to interactions with objects or obstacles. Relying exclusively on one type of link can lead to suboptimal and limited performance. Link management (LM) is emerging as a viable solution, and accurate channel estimation provides essential information to make informed decisions about transmission parameters. In this paper, we study LM and channel estimation that flexibly adjust the transmission ratio of direct and indirect links to improve generalization, using a denoising variational autoencoder with attention modules (DVAE-ATT) to enhance sum rate. Our experiments show significant improvements in IRS-assisted millimeter-wave MIMO systems. Incorporating LM increased the sum rate and reduced MSE by approximately 9%. Variational autoencoders (VAE) outperformed traditional autoencoders in the spatial domain, as confirmed by heatmap analysis. Additionally, our investigation of DVAE-ATT reveals notable differences in the temporal domain with and without attention mechanisms. Finally, we analyze performance across varying numbers of users and ranges. Across various distances—5m, 15m, 25m, and 35m—performance improvements averaged 6%, 11%, 16%, and 22%, respectively.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"381-394"},"PeriodicalIF":0.0,"publicationDate":"2025-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10909334","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143583165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The Metaverse holds the potential to revolutionize digital interactions through the establishment of a highly dynamic and immersive virtual realm over wireless communications systems, offering services such as massive twinning and telepresence. This landscape presents novel challenges, particularly efficient management of multiple access to the frequency spectrum, for which numerous adaptive Deep Reinforcement Learning (DRL) approaches have been explored. However, challenges persist in adapting agents to heterogeneous and non-stationary wireless environments. In this paper, we present a novel approach that leverages Continual Learning (CL) to enhance intelligent Medium Access Control (MAC) protocols, featuring an intelligent agent coexisting with legacy User Equipments (UEs) with varying numbers, protocols, and transmission profiles unknown to the agent for the sake of backward compatibility and privacy. We introduce an adaptive Double and Dueling Deep Q-Learning (D3QL)-based MAC protocol, enriched by a symmetry-aware CL mechanism, which maximizes intelligent agent throughput while ensuring fairness. Mathematical analysis validates the efficiency of our proposed scheme, showcasing superiority over conventional DRL-based techniques in terms of throughput, collision rate, and fairness, coupled with real-time responsiveness in highly dynamic scenarios.
{"title":"A Novel Multiple Access Scheme for Heterogeneous Wireless Communications Using Symmetry-Aware Continual Deep Reinforcement Learning","authors":"Hamidreza Mazandarani;Masoud Shokrnezhad;Tarik Taleb","doi":"10.1109/TMLCN.2025.3546183","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3546183","url":null,"abstract":"The Metaverse holds the potential to revolutionize digital interactions through the establishment of a highly dynamic and immersive virtual realm over wireless communications systems, offering services such as massive twinning and telepresence. This landscape presents novel challenges, particularly efficient management of multiple access to the frequency spectrum, for which numerous adaptive Deep Reinforcement Learning (DRL) approaches have been explored. However, challenges persist in adapting agents to heterogeneous and non-stationary wireless environments. In this paper, we present a novel approach that leverages Continual Learning (CL) to enhance intelligent Medium Access Control (MAC) protocols, featuring an intelligent agent coexisting with legacy User Equipments (UEs) with varying numbers, protocols, and transmission profiles unknown to the agent for the sake of backward compatibility and privacy. We introduce an adaptive Double and Dueling Deep Q-Learning (D3QL)-based MAC protocol, enriched by a symmetry-aware CL mechanism, which maximizes intelligent agent throughput while ensuring fairness. Mathematical analysis validates the efficiency of our proposed scheme, showcasing superiority over conventional DRL-based techniques in terms of throughput, collision rate, and fairness, coupled with real-time responsiveness in highly dynamic scenarios.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"353-368"},"PeriodicalIF":0.0,"publicationDate":"2025-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10908203","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-26DOI: 10.1109/TMLCN.2025.3546181
Ruslan Zhagypar;Nour Kouzayha;Hesham ElSawy;Hayssam Dahrouj;Tareq Y. Al-Naffouri
The development of the sixth-generation (6G) of wireless networks is driving computation toward the network edge, where Hierarchical Federated Learning (HFL) plays a pivotal role in distributing learning across edge devices. In HFL, edge devices train local models and send updates to an edge server for local aggregation, which are then forwarded to a central server for global aggregation. However, the unreliability of communication channels at the edge and backhaul links poses a significant bottleneck for HFL-enabled systems. To address this challenge, this paper proposes an unbiased HFL algorithm for Uncrewed Aerial Vehicle (UAV)-assisted wireless networks. While applicable to terrestrial base stations (BSs), the proposed algorithm relies on UAVs for local model aggregation thanks to their ability to enhance wireless channels with lower latency and improved coverage. The proposed algorithm adjusts update weights during local and global aggregations at UAVs to mitigate the impact of unreliable channels. To quantify channel unreliability in HFL, stochastic geometry tools are employed to assess success probabilities of local and global model parameter transmissions. Incorporating these metrics aims to mitigate biases towards devices with better channel conditions in UAV-assisted networks. The paper further examines the theoretical convergence of the proposed unbiased UAV-assisted HFL algorithm under adverse channel conditions and highlights the impact of the limited battery capacity of the UAV on the efficiency of the HFL algorithm. Additionally, the algorithm facilitates optimization of system parameters such as UAV count, altitude, battery capacity, etc. The simulation results underscore the effectiveness of the proposed unbiased HFL scheme, demonstrating a 5.5% higher accuracy and approximately 85% faster convergence compared to conventional HFL algorithms. We make our code available at the following GitHub repository: $texttt {UAV-assisted Unbiased HFL Code}$ .
{"title":"UAV-Assisted Unbiased Hierarchical Federated Learning: Performance and Convergence Analysis","authors":"Ruslan Zhagypar;Nour Kouzayha;Hesham ElSawy;Hayssam Dahrouj;Tareq Y. Al-Naffouri","doi":"10.1109/TMLCN.2025.3546181","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3546181","url":null,"abstract":"The development of the sixth-generation (6G) of wireless networks is driving computation toward the network edge, where Hierarchical Federated Learning (HFL) plays a pivotal role in distributing learning across edge devices. In HFL, edge devices train local models and send updates to an edge server for local aggregation, which are then forwarded to a central server for global aggregation. However, the unreliability of communication channels at the edge and backhaul links poses a significant bottleneck for HFL-enabled systems. To address this challenge, this paper proposes an unbiased HFL algorithm for Uncrewed Aerial Vehicle (UAV)-assisted wireless networks. While applicable to terrestrial base stations (BSs), the proposed algorithm relies on UAVs for local model aggregation thanks to their ability to enhance wireless channels with lower latency and improved coverage. The proposed algorithm adjusts update weights during local and global aggregations at UAVs to mitigate the impact of unreliable channels. To quantify channel unreliability in HFL, stochastic geometry tools are employed to assess success probabilities of local and global model parameter transmissions. Incorporating these metrics aims to mitigate biases towards devices with better channel conditions in UAV-assisted networks. The paper further examines the theoretical convergence of the proposed unbiased UAV-assisted HFL algorithm under adverse channel conditions and highlights the impact of the limited battery capacity of the UAV on the efficiency of the HFL algorithm. Additionally, the algorithm facilitates optimization of system parameters such as UAV count, altitude, battery capacity, etc. The simulation results underscore the effectiveness of the proposed unbiased HFL scheme, demonstrating a 5.5% higher accuracy and approximately 85% faster convergence compared to conventional HFL algorithms. We make our code available at the following GitHub repository: <inline-formula> <tex-math>$texttt {UAV-assisted Unbiased HFL Code}$ </tex-math></inline-formula>.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"420-447"},"PeriodicalIF":0.0,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10904929","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143645156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-02-26DOI: 10.1109/TMLCN.2025.3545777
Yuwen Qian;Tianyang Qiu;Chuan Ma;Yiyang Ni;Long Yuan;Xiangwei Zhou;Jun Li
Intelligent transportation systems grapple with the formidable task of precisely forecasting real-time traffic conditions, where the traffic dynamics exhibit intricacies arising from spatial and temporal dependencies. The urban road network presents a complex web of interconnected roads, where the state of traffic on one road can influence the conditions of others. Moreover, the prediction of traffic conditions necessitates the consideration of diverse temporal factors. Notably, the proximity of a time point to the present moment wields a more substantial impact on subsequent states. In this paper, we propose the knowledge-driven graph convolutional network (KGCN) aided by the gated recurrent unit with a selected attention mechanism (GSAM) to predict traffic flow. In particular, KGCN is employed to capture the correlation of the external knowledge factors for the road and the spatial dependencies, and the gated recurrent unit (GRU) is used to cope with temporal dependence. Furthermore, to improve traffic prediction accuracy, we propose the GRU combined with a selected attention mechanism with Gumble-Max to predict traffic at the temporal dimension, where a selector is chosen to dynamically assign the feature in various time intervals with different weights. Experimental results with real-life data show the proposed KGCN with GSAM can achieve high accuracy in traffic prediction. Compared to the traditional traffic prediction method, the proposed KGCN with GSAM can achieve higher efficacy and robustness when capturing global dynamic temporal dependencies, external knowledge factor correlations, and spatial correlations.
{"title":"On Traffic Prediction With Knowledge-Driven Spatial–Temporal Graph Convolutional Network Aided by Selected Attention Mechanism","authors":"Yuwen Qian;Tianyang Qiu;Chuan Ma;Yiyang Ni;Long Yuan;Xiangwei Zhou;Jun Li","doi":"10.1109/TMLCN.2025.3545777","DOIUrl":"https://doi.org/10.1109/TMLCN.2025.3545777","url":null,"abstract":"Intelligent transportation systems grapple with the formidable task of precisely forecasting real-time traffic conditions, where the traffic dynamics exhibit intricacies arising from spatial and temporal dependencies. The urban road network presents a complex web of interconnected roads, where the state of traffic on one road can influence the conditions of others. Moreover, the prediction of traffic conditions necessitates the consideration of diverse temporal factors. Notably, the proximity of a time point to the present moment wields a more substantial impact on subsequent states. In this paper, we propose the knowledge-driven graph convolutional network (KGCN) aided by the gated recurrent unit with a selected attention mechanism (GSAM) to predict traffic flow. In particular, KGCN is employed to capture the correlation of the external knowledge factors for the road and the spatial dependencies, and the gated recurrent unit (GRU) is used to cope with temporal dependence. Furthermore, to improve traffic prediction accuracy, we propose the GRU combined with a selected attention mechanism with Gumble-Max to predict traffic at the temporal dimension, where a selector is chosen to dynamically assign the feature in various time intervals with different weights. Experimental results with real-life data show the proposed KGCN with GSAM can achieve high accuracy in traffic prediction. Compared to the traditional traffic prediction method, the proposed KGCN with GSAM can achieve higher efficacy and robustness when capturing global dynamic temporal dependencies, external knowledge factor correlations, and spatial correlations.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"3 ","pages":"369-380"},"PeriodicalIF":0.0,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10904899","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143570620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}