Pub Date : 2024-08-09DOI: 10.1109/TMLCN.2024.3441521
Anastasios Foliadis;Mario H. Castañeda Garcia;Richard A. Stirling-Gallacher;Reiner S. Thomä
Deep learning (DL) methods have been shown to improve the performance of several use cases for the fifth-generation (5G) New radio (NR) air interface. In this paper we investigate user equipment (UE) positioning using the channel state information (CSI) fingerprints between a UE and multiple base stations (BSs). In such a setup, we consider two different fusion techniques: early and late fusion. With early fusion, a single DL model can be trained for UE positioning by combining the CSI fingerprints of the multiple BSs as input. With late fusion, a separate DL model is trained at each BS using the CSI specific to that BS and the outputs of these individual models are then combined to determine the UE’s position. In this work we compare these different fusion techniques and show that fusing the outputs of separate models achieves higher positioning accuracy, especially in a dynamic scenario. We also show that the combination of multiple outputs further benefits from considering the uncertainty of the output of the DL model at each BS. For a more efficient training of the DL model across BSs, we additionally propose a multi-task learning (MTL) scheme by sharing some parameters across the models while jointly training all models. This method, not only improves the accuracy of the individual models, but also of the final combined estimate. Lastly, we evaluate the reliability of the uncertainty estimation to determine which of the fusion methods provides the highest quality of uncertainty estimates.
{"title":"Deep Learning-Based Positioning With Multi-Task Learning and Uncertainty-Based Fusion","authors":"Anastasios Foliadis;Mario H. Castañeda Garcia;Richard A. Stirling-Gallacher;Reiner S. Thomä","doi":"10.1109/TMLCN.2024.3441521","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3441521","url":null,"abstract":"Deep learning (DL) methods have been shown to improve the performance of several use cases for the fifth-generation (5G) New radio (NR) air interface. In this paper we investigate user equipment (UE) positioning using the channel state information (CSI) fingerprints between a UE and multiple base stations (BSs). In such a setup, we consider two different fusion techniques: early and late fusion. With early fusion, a single DL model can be trained for UE positioning by combining the CSI fingerprints of the multiple BSs as input. With late fusion, a separate DL model is trained at each BS using the CSI specific to that BS and the outputs of these individual models are then combined to determine the UE’s position. In this work we compare these different fusion techniques and show that fusing the outputs of separate models achieves higher positioning accuracy, especially in a dynamic scenario. We also show that the combination of multiple outputs further benefits from considering the uncertainty of the output of the DL model at each BS. For a more efficient training of the DL model across BSs, we additionally propose a multi-task learning (MTL) scheme by sharing some parameters across the models while jointly training all models. This method, not only improves the accuracy of the individual models, but also of the final combined estimate. Lastly, we evaluate the reliability of the uncertainty estimation to determine which of the fusion methods provides the highest quality of uncertainty estimates.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1127-1141"},"PeriodicalIF":0.0,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10632202","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142083901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This article addresses the problem of Ultra Reliable Low Latency Communications (URLLC) in wireless networks, a framework with particularly stringent constraints imposed by many Internet of Things (IoT) applications from diverse sectors. We propose a novel Deep Reinforcement Learning (DRL) scheduling algorithm, named NOMA-PPO, to solve the Non-Orthogonal Multiple Access (NOMA) uplink URLLC scheduling problem involving strict deadlines. The challenge of addressing uplink URLLC requirements in NOMA systems is related to the combinatorial complexity of the action space due to the possibility to schedule multiple devices, and to the partial observability constraint that we impose to our algorithm in order to meet the IoT communication constraints and be scalable. Our approach involves 1) formulating the NOMA-URLLC problem as a Partially Observable Markov Decision Process (POMDP) and the introduction of an agent state, serving as a sufficient statistic of past observations and actions, enabling a transformation of the POMDP into a Markov Decision Process (MDP); 2) adapting the Proximal Policy Optimization (PPO) algorithm to handle the combinatorial action space; 3) incorporating prior knowledge into the learning agent with the introduction of a Bayesian policy. Numerical results reveal that not only does our approach outperform traditional multiple access protocols and DRL benchmarks on 3GPP scenarios, but also proves to be robust under various channel and traffic configurations, efficiently exploiting inherent time correlations.
{"title":"Deep Reinforcement Learning for Uplink Scheduling in NOMA-URLLC Networks","authors":"Benoît-Marie Robaglia;Marceau Coupechoux;Dimitrios Tsilimantos","doi":"10.1109/TMLCN.2024.3437351","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3437351","url":null,"abstract":"This article addresses the problem of Ultra Reliable Low Latency Communications (URLLC) in wireless networks, a framework with particularly stringent constraints imposed by many Internet of Things (IoT) applications from diverse sectors. We propose a novel Deep Reinforcement Learning (DRL) scheduling algorithm, named NOMA-PPO, to solve the Non-Orthogonal Multiple Access (NOMA) uplink URLLC scheduling problem involving strict deadlines. The challenge of addressing uplink URLLC requirements in NOMA systems is related to the combinatorial complexity of the action space due to the possibility to schedule multiple devices, and to the partial observability constraint that we impose to our algorithm in order to meet the IoT communication constraints and be scalable. Our approach involves 1) formulating the NOMA-URLLC problem as a Partially Observable Markov Decision Process (POMDP) and the introduction of an agent state, serving as a sufficient statistic of past observations and actions, enabling a transformation of the POMDP into a Markov Decision Process (MDP); 2) adapting the Proximal Policy Optimization (PPO) algorithm to handle the combinatorial action space; 3) incorporating prior knowledge into the learning agent with the introduction of a Bayesian policy. Numerical results reveal that not only does our approach outperform traditional multiple access protocols and DRL benchmarks on 3GPP scenarios, but also proves to be robust under various channel and traffic configurations, efficiently exploiting inherent time correlations.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1142-1158"},"PeriodicalIF":0.0,"publicationDate":"2024-08-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10621640","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142099787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-31DOI: 10.1109/TMLCN.2024.3436472
Stéfani Pires;Adriana Ribeiro;Leobino N. Sampaio
In-network cache architectures, such as Information-centric networks (ICNs), have proven to be an efficient alternative to deal with the growing content consumption on networks. In caching networks, any device can potentially act as a caching node. In practice, real cache networks may employ different caching replacement policies by a node. The reason is that the policies may vary in efficiency according to unbounded context factors, such as cache size, content request pattern, content distribution popularity, and the relative cache location. The lack of suitable policies for all nodes and scenarios undermines the efficient use of available cache resources. Therefore, a new model for choosing caching policies appropriately to cache contexts on-demand and over time becomes necessary. In this direction, we propose a new caching meta-policy strategy capable of learning the most appropriate policy for cache online and dynamically adapting to context variations that leads to changes in which policy is best. The meta-policy decouples the eviction strategy from managing the context information used by the policy, and models the choice of suitable policies as online learning with a bandit feedback problem. The meta-policy supports deploying a diverse set of self-contained caching policies in different scenarios, including adaptive policies. Experimental results with single and multiple caches have shown the meta-policy effectiveness and adaptability to different content request models in synthetic and trace-driven simulations. Moreover, we compared the meta-policy adaptive behavior with the Adaptive Replacement Policy (ARC) behavior.
{"title":"On Learning Suitable Caching Policies for In-Network Caching","authors":"Stéfani Pires;Adriana Ribeiro;Leobino N. Sampaio","doi":"10.1109/TMLCN.2024.3436472","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3436472","url":null,"abstract":"In-network cache architectures, such as Information-centric networks (ICNs), have proven to be an efficient alternative to deal with the growing content consumption on networks. In caching networks, any device can potentially act as a caching node. In practice, real cache networks may employ different caching replacement policies by a node. The reason is that the policies may vary in efficiency according to unbounded context factors, such as cache size, content request pattern, content distribution popularity, and the relative cache location. The lack of suitable policies for all nodes and scenarios undermines the efficient use of available cache resources. Therefore, a new model for choosing caching policies appropriately to cache contexts on-demand and over time becomes necessary. In this direction, we propose a new caching meta-policy strategy capable of learning the most appropriate policy for cache online and dynamically adapting to context variations that leads to changes in which policy is best. The meta-policy decouples the eviction strategy from managing the context information used by the policy, and models the choice of suitable policies as online learning with a bandit feedback problem. The meta-policy supports deploying a diverse set of self-contained caching policies in different scenarios, including adaptive policies. Experimental results with single and multiple caches have shown the meta-policy effectiveness and adaptability to different content request models in synthetic and trace-driven simulations. Moreover, we compared the meta-policy adaptive behavior with the Adaptive Replacement Policy (ARC) behavior.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1076-1092"},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10616152","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965679","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-29DOI: 10.1109/TMLCN.2024.3435168
Zhen Gao;Shuang Liu;Junbo Zhao;Xiaofei Wang;Yu Wang;Zhu Han
Convolutional Neural Networks (CNNs) have been applied in wide areas of computer vision, and edge intelligence is expected to provide instant AI service with the support of broadband mobile networks. However, the deployment of CNNs on network edge faces severe challenges. First, edge or embedded devices are usually not reliable, and hardware failures can corrupt the CNN system, which is unacceptable for critical applications, such as autonomous driving and object detection on space platforms. Second, edge or embedded devices are usually resource-limited, and therefore traditional redundancy-based protection methods are not applicable due to huge overhead. Although network pruning is effective to reduce the complexity of CNNs, we cannot have sufficient data for performance recovery in many scenarios due to privacy and security concerns. To enhance the reliability of CNNs on resource-limited devices with the few-shot constraint, we propose to construct an ensemble system with weak base CNNs pruned from the original strong CNN. To improve the ensemble performance with diverse base CNNs, we first propose a novel filter importance evaluation method by combining the amplitude and gradient information of the filter. Since the gradient part is related to the input data, different subsets of data are used for layer sensitivity analysis for different base CNNs, so that the different pruning configurations can be obtained for each base CNN. On this basis, a modified ReLU function is proposed to determine the final pruning rate of each layer in each base CNN. Extensive experiments prove that the proposed solution can effectively improve the reliability of CNNs with much less resource requirement for each edge server.
{"title":"Ensemble-Based Reliability Enhancement for Edge-Deployed CNNs in Few-Shot Scenarios","authors":"Zhen Gao;Shuang Liu;Junbo Zhao;Xiaofei Wang;Yu Wang;Zhu Han","doi":"10.1109/TMLCN.2024.3435168","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3435168","url":null,"abstract":"Convolutional Neural Networks (CNNs) have been applied in wide areas of computer vision, and edge intelligence is expected to provide instant AI service with the support of broadband mobile networks. However, the deployment of CNNs on network edge faces severe challenges. First, edge or embedded devices are usually not reliable, and hardware failures can corrupt the CNN system, which is unacceptable for critical applications, such as autonomous driving and object detection on space platforms. Second, edge or embedded devices are usually resource-limited, and therefore traditional redundancy-based protection methods are not applicable due to huge overhead. Although network pruning is effective to reduce the complexity of CNNs, we cannot have sufficient data for performance recovery in many scenarios due to privacy and security concerns. To enhance the reliability of CNNs on resource-limited devices with the few-shot constraint, we propose to construct an ensemble system with weak base CNNs pruned from the original strong CNN. To improve the ensemble performance with diverse base CNNs, we first propose a novel filter importance evaluation method by combining the amplitude and gradient information of the filter. Since the gradient part is related to the input data, different subsets of data are used for layer sensitivity analysis for different base CNNs, so that the different pruning configurations can be obtained for each base CNN. On this basis, a modified ReLU function is proposed to determine the final pruning rate of each layer in each base CNN. Extensive experiments prove that the proposed solution can effectively improve the reliability of CNNs with much less resource requirement for each edge server.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1062-1075"},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10614218","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141965372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-25DOI: 10.1109/TMLCN.2024.3433620
Swapnil Sadashiv Shinde;Daniele Tarchi
Vehicular Edge Computing (VEC) represents a novel advancement within the Internet of Vehicles (IoV). Despite its implementation through Road Side Units (RSUs), VEC frequently falls short of satisfying the escalating demands of Vehicle Users (VUs) for new services, necessitating supplementary computational and communication resources. Non-Terrestrial Networks (NTN) with onboard Edge Computing (EC) facilities are gaining a central place in the 6G vision, allowing one to extend future services also to uncovered areas. This scenario, composed of a multitude of VUs, terrestrial and non-terrestrial nodes, and characterized by mobility and stringent requirements, brings in a very high complexity. Machine Learning (ML) represents a perfect tool for solving these types of problems. Integrated Terrestrial and Non-terrestrial (T-NT) EC, supported by innovative intelligent solutions enabled through ML technology, can boost the VEC capacity, coverage range, and resource utilization. Therefore, by exploring the integrated T-NT EC platforms, we design a multi-EC-enabled vehicular networking platform with a heterogeneous set of services. Next, we model the latency and energy requirements for processing the VU tasks through partial computation offloading operations. We aim to optimize the overall latency and energy requirements for processing the VU data by selecting the appropriate edge nodes and the offloading amount. The problem is defined as a multi-layer sequential decision-making problem through the Markov Decision Processes (MDP). The Hierarchical Reinforcement Learning (HRL) method, implemented through a Deep Q network, is used to optimize the network selection and offloading policies. Simulation results are compared with different benchmark methods to show performance gains in terms of overall cost requirements and reliability.
车载边缘计算(VEC)是车联网(IoV)的一项新进展。尽管通过路侧单元(RSU)实现了 VEC,但 VEC 经常无法满足车辆用户(VU)对新服务不断升级的需求,因此需要补充计算和通信资源。带有车载边缘计算(EC)设施的非地面网络(NTN)在 6G 愿景中占据了重要位置,使未来的服务也能扩展到未覆盖区域。这种场景由众多 VU、地面和非地面节点组成,具有移动性和严格要求的特点,因此复杂度非常高。机器学习 (ML) 是解决此类问题的完美工具。地面和非地面(T-NT)集成电子通信系统在通过 ML 技术实现的创新智能解决方案的支持下,可以提高 VEC 容量、覆盖范围和资源利用率。因此,通过探索集成的 T-NT 电子通信平台,我们设计了一个具有异构服务集的多电子通信车联网平台。接下来,我们通过部分计算卸载操作对处理 VU 任务的延迟和能源需求进行建模。我们的目标是通过选择合适的边缘节点和卸载量,优化处理 VU 数据的整体延迟和能源需求。通过马尔可夫决策过程(Markov Decision Processes,MDP),该问题被定义为多层顺序决策问题。通过深度 Q 网络实现的分层强化学习(HRL)方法用于优化网络选择和卸载策略。仿真结果与不同的基准方法进行了比较,以显示在总体成本要求和可靠性方面的性能提升。
{"title":"Hierarchical Reinforcement Learning for Multi-Layer Multi-Service Non-Terrestrial Vehicular Edge Computing","authors":"Swapnil Sadashiv Shinde;Daniele Tarchi","doi":"10.1109/TMLCN.2024.3433620","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3433620","url":null,"abstract":"Vehicular Edge Computing (VEC) represents a novel advancement within the Internet of Vehicles (IoV). Despite its implementation through Road Side Units (RSUs), VEC frequently falls short of satisfying the escalating demands of Vehicle Users (VUs) for new services, necessitating supplementary computational and communication resources. Non-Terrestrial Networks (NTN) with onboard Edge Computing (EC) facilities are gaining a central place in the 6G vision, allowing one to extend future services also to uncovered areas. This scenario, composed of a multitude of VUs, terrestrial and non-terrestrial nodes, and characterized by mobility and stringent requirements, brings in a very high complexity. Machine Learning (ML) represents a perfect tool for solving these types of problems. Integrated Terrestrial and Non-terrestrial (T-NT) EC, supported by innovative intelligent solutions enabled through ML technology, can boost the VEC capacity, coverage range, and resource utilization. Therefore, by exploring the integrated T-NT EC platforms, we design a multi-EC-enabled vehicular networking platform with a heterogeneous set of services. Next, we model the latency and energy requirements for processing the VU tasks through partial computation offloading operations. We aim to optimize the overall latency and energy requirements for processing the VU data by selecting the appropriate edge nodes and the offloading amount. The problem is defined as a multi-layer sequential decision-making problem through the Markov Decision Processes (MDP). The Hierarchical Reinforcement Learning (HRL) method, implemented through a Deep Q network, is used to optimize the network selection and offloading policies. Simulation results are compared with different benchmark methods to show performance gains in terms of overall cost requirements and reliability.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1045-1061"},"PeriodicalIF":0.0,"publicationDate":"2024-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10609447","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141964883","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-24DOI: 10.1109/TMLCN.2024.3432865
Mehdi Sattari;Hao Guo;Deniz Gündüz;Ashkan Panahi;Tommy Svensson
Millimeter wave (mmWave) multiple-input-multi-output (MIMO) is now a reality with great potential for further improvement. We study full-duplex transmissions as an effective way to improve mmWave MIMO systems. Compared to half-duplex systems, full-duplex transmissions may offer higher data rates and lower latency. However, full-duplex transmission is hindered by self-interference (SI) at the receive antennas, and SI channel estimation becomes a crucial step to make the full-duplex systems feasible. In this paper, we address the problem of channel estimation in full-duplex mmWave MIMO systems using neural networks (NNs). Our approach involves sharing pilot resources between user equipments (UEs) and transmit antennas at the base station (BS), aiming to reduce the pilot overhead in full-duplex systems and to achieve a comparable level to that of a half-duplex system. Additionally, in the case of separate antenna configurations in a full-duplex BS, providing channel estimates of transmit antenna (TX) arrays to the downlink UEs poses another challenge, as the TX arrays are not capable of receiving pilot signals. To address this, we employ an NN to map the channel from the downlink UEs to the receive antenna (RX) arrays to the channel from the TX arrays to the downlink UEs. We further elaborate on how NNs perform the estimation with different architectures, (e.g., different numbers of hidden layers), the introduction of non-linear distortion (e.g., with a 1-bit analog-to-digital converter (ADC)), and different channel conditions (e.g., low-correlated and high-correlated channels). Our work provides novel insights into NN-based channel estimators.
毫米波(mmWave)多输入多输出(MIMO)现已成为现实,并具有进一步改进的巨大潜力。我们将全双工传输作为改进毫米波多输入多输出系统的有效方法进行研究。与半双工系统相比,全双工传输可提供更高的数据传输速率和更低的延迟。然而,全双工传输会受到接收天线自干扰(SI)的阻碍,因此 SI 信道估计成为使全双工系统可行的关键步骤。在本文中,我们利用神经网络(NN)解决了全双工毫米波 MIMO 系统中的信道估计问题。我们的方法涉及在用户设备(UE)和基站(BS)的发射天线之间共享先导资源,旨在减少全双工系统中的先导开销,并达到与半双工系统相当的水平。此外,在全双工基站采用独立天线配置的情况下,向下行链路 UE 提供发射天线(TX)阵列的信道估计也是一个挑战,因为 TX 阵列无法接收先导信号。为了解决这个问题,我们采用了一种 NN,将下行链路 UE 到接收天线 (RX) 阵列的信道映射到从发射天线阵列到下行链路 UE 的信道。我们进一步阐述了 NN 如何在不同架构(如不同数量的隐藏层)、引入非线性失真(如使用 1 位模数转换器 (ADC))和不同信道条件(如低相关和高相关信道)下执行估计。我们的工作为基于 NN 的信道估计器提供了新的见解。
{"title":"Full-Duplex Millimeter Wave MIMO Channel Estimation: A Neural Network Approach","authors":"Mehdi Sattari;Hao Guo;Deniz Gündüz;Ashkan Panahi;Tommy Svensson","doi":"10.1109/TMLCN.2024.3432865","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3432865","url":null,"abstract":"Millimeter wave (mmWave) multiple-input-multi-output (MIMO) is now a reality with great potential for further improvement. We study full-duplex transmissions as an effective way to improve mmWave MIMO systems. Compared to half-duplex systems, full-duplex transmissions may offer higher data rates and lower latency. However, full-duplex transmission is hindered by self-interference (SI) at the receive antennas, and SI channel estimation becomes a crucial step to make the full-duplex systems feasible. In this paper, we address the problem of channel estimation in full-duplex mmWave MIMO systems using neural networks (NNs). Our approach involves sharing pilot resources between user equipments (UEs) and transmit antennas at the base station (BS), aiming to reduce the pilot overhead in full-duplex systems and to achieve a comparable level to that of a half-duplex system. Additionally, in the case of separate antenna configurations in a full-duplex BS, providing channel estimates of transmit antenna (TX) arrays to the downlink UEs poses another challenge, as the TX arrays are not capable of receiving pilot signals. To address this, we employ an NN to map the channel from the downlink UEs to the receive antenna (RX) arrays to the channel from the TX arrays to the downlink UEs. We further elaborate on how NNs perform the estimation with different architectures, (e.g., different numbers of hidden layers), the introduction of non-linear distortion (e.g., with a 1-bit analog-to-digital converter (ADC)), and different channel conditions (e.g., low-correlated and high-correlated channels). Our work provides novel insights into NN-based channel estimators.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1093-1108"},"PeriodicalIF":0.0,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10608175","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141994029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-24DOI: 10.1109/TMLCN.2024.3432861
Qiyun Guo;Haotai Liang;Zhicheng Bao;Chen Dong;Xiaodong Xu;Zhongzheng Tang;Yue Bei
Endogenous intelligence has emerged as a crucial aspect of next-generation communication networks. This concept is closely intertwined with artificial intelligence (AI), with its primary components being data, algorithms, and computility. Data collection remains a critical concern that warrants focused attention. To address the challenge of data expansion and forwarding, the intellicise router is proposed. It extends the local dataset and continuously enhances the local model through a specifically crafted algorithm, which enhances AI performance, as exemplified by its application in image recognition tasks. Service capability is employed to gauge the router’s ability to provide services and the upper bounds are derived. To analyze the algorithm’s effectiveness, a category-increase model is developed to calculate the probability of categories rising under both equal and unequal probabilities of image communication categories. The numerical analysis results align with simulation results, affirming the validity of the category-increase model. To assess the performance of the intellicise router, a communication system is simulated. A comparative analysis of these experimental results demonstrates that the intellicise router can continuously improve its performance to provide better service.
{"title":"Intellicise Router Promotes Endogenous Intelligence in Communication Network","authors":"Qiyun Guo;Haotai Liang;Zhicheng Bao;Chen Dong;Xiaodong Xu;Zhongzheng Tang;Yue Bei","doi":"10.1109/TMLCN.2024.3432861","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3432861","url":null,"abstract":"Endogenous intelligence has emerged as a crucial aspect of next-generation communication networks. This concept is closely intertwined with artificial intelligence (AI), with its primary components being data, algorithms, and computility. Data collection remains a critical concern that warrants focused attention. To address the challenge of data expansion and forwarding, the intellicise router is proposed. It extends the local dataset and continuously enhances the local model through a specifically crafted algorithm, which enhances AI performance, as exemplified by its application in image recognition tasks. Service capability is employed to gauge the router’s ability to provide services and the upper bounds are derived. To analyze the algorithm’s effectiveness, a category-increase model is developed to calculate the probability of categories rising under both equal and unequal probabilities of image communication categories. The numerical analysis results align with simulation results, affirming the validity of the category-increase model. To assess the performance of the intellicise router, a communication system is simulated. A comparative analysis of these experimental results demonstrates that the intellicise router can continuously improve its performance to provide better service.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1509-1526"},"PeriodicalIF":0.0,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10608170","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142397285","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-07-04DOI: 10.1109/TMLCN.2024.3423648
Li Ju;Tianru Zhang;Salman Toor;Andreas Hellander
Federated learning is a distributed and privacy-preserving approach to train a statistical model collaboratively from decentralized data held by different parties. However, when the datasets are not independent and identically distributed, models trained by naive federated algorithms may be biased towards certain participants, and model performance across participants is non-uniform. This is known as the fairness problem in federated learning. In this paper, we formulate fairness-controlled federated learning as a dynamical multi-objective optimization problem to ensure the fairness and convergence with theoretical guarantee. To solve the problem efficiently, we study the convergence and bias of Adam