Pub Date : 2024-06-05DOI: 10.1109/TMLCN.2024.3410208
Halit Bugra Tulay;Can Emre Koksal
With the growing adoption of vehicular networks, ensuring the security of these networks is becoming increasingly crucial. However, the broadcast nature of communication in these networks creates numerous privacy and security concerns. In particular, the Sybil attack, where attackers can use multiple identities to disseminate false messages, cause service delays, or gain control of the network, poses a significant threat. To combat this attack, we propose a novel approach utilizing the channel state information (CSI) of vehicles. Our approach leverages the distinct spatio-temporal variations of CSI samples obtained in vehicular communication signals to detect these attacks. We conduct extensive real-world experiments using vehicle-to-everything (V2X) data, gathered from dedicated short-range communications (DSRC) in vehicular networks. Our results demonstrate a high detection rate of over 98% in the real-world experiments, showcasing the practicality and effectiveness of our method in realistic vehicular scenarios. Furthermore, we rigorously test our approach through advanced ray-tracing simulations in urban environments, which demonstrates high efficacy even in complex scenarios involving various vehicles. This makes our approach a valuable, hardware-independent solution for the V2X technologies at major intersections.
{"title":"Sybil Attack Detection Based on Signal Clustering in Vehicular Networks","authors":"Halit Bugra Tulay;Can Emre Koksal","doi":"10.1109/TMLCN.2024.3410208","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3410208","url":null,"abstract":"With the growing adoption of vehicular networks, ensuring the security of these networks is becoming increasingly crucial. However, the broadcast nature of communication in these networks creates numerous privacy and security concerns. In particular, the Sybil attack, where attackers can use multiple identities to disseminate false messages, cause service delays, or gain control of the network, poses a significant threat. To combat this attack, we propose a novel approach utilizing the channel state information (CSI) of vehicles. Our approach leverages the distinct spatio-temporal variations of CSI samples obtained in vehicular communication signals to detect these attacks. We conduct extensive real-world experiments using vehicle-to-everything (V2X) data, gathered from dedicated short-range communications (DSRC) in vehicular networks. Our results demonstrate a high detection rate of over 98% in the real-world experiments, showcasing the practicality and effectiveness of our method in realistic vehicular scenarios. Furthermore, we rigorously test our approach through advanced ray-tracing simulations in urban environments, which demonstrates high efficacy even in complex scenarios involving various vehicles. This makes our approach a valuable, hardware-independent solution for the V2X technologies at major intersections.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"753-765"},"PeriodicalIF":0.0,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10550012","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141319672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-05DOI: 10.1109/TMLCN.2024.3410211
Mohammed Saif;Md. Zoheb Hassan;Md. Jahangir Hossain
It is anticipated that aerial-terrestrial integrated networks incorporating unmanned aerial vehicles (UAVs) mounted relays will offer improved coverage and connectivity in the beyond 5G era. Meanwhile, federated learning (FL) is a promising distributed machine learning technique for building inference models over wireless networks due to its ability to maintain user privacy and reduce communication overhead. However, off-the-shelf FL models aggregate global parameters at a central parameter server (CPS), increasing energy consumption and latency, as well as inefficiently utilizing radio resource blocks (RRBs) for distributed user devices (UDs). This paper presents a resource-efficient and decentralized FL framework called FedMoD (federated learning with model dissemination), for millimeter-wave (mmWave) aerial-terrestrial integrated networks with the following two unique characteristics. Firstly, FedMoD incorporates a novel decentralized model dissemination scheme that uses UAVs as local model aggregators through UAV-to-UAV and device-to-device (D2D) communications. As a result, FedMoD 1) increases the number of participant UDs in developing the FL model; and 2) achieves global model aggregation without involving CPS. Secondly, FedMoD reduces FL’s energy consumption using radio resource management (RRM) under the constraints of over-the-air learning latency. To achieve this, by leveraging graph theory, FedMoD optimizes the scheduling of line-of-sight (LOS) UDs to suitable UAVs and RRBs over mmWave links and non-LOS UDs to available LOS UDs via overlay D2D communications. Extensive simulations reveal that FedMoD, despite being decentralized, offers the same convergence performance to the conventional centralized FL frameworks.
{"title":"Decentralized Aggregation for Energy-Efficient Federated Learning in mmWave Aerial-Terrestrial Integrated Networks","authors":"Mohammed Saif;Md. Zoheb Hassan;Md. Jahangir Hossain","doi":"10.1109/TMLCN.2024.3410211","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3410211","url":null,"abstract":"It is anticipated that aerial-terrestrial integrated networks incorporating unmanned aerial vehicles (UAVs) mounted relays will offer improved coverage and connectivity in the beyond 5G era. Meanwhile, federated learning (FL) is a promising distributed machine learning technique for building inference models over wireless networks due to its ability to maintain user privacy and reduce communication overhead. However, off-the-shelf FL models aggregate global parameters at a central parameter server (CPS), increasing energy consumption and latency, as well as inefficiently utilizing radio resource blocks (RRBs) for distributed user devices (UDs). This paper presents a resource-efficient and decentralized FL framework called FedMoD (federated learning with model dissemination), for millimeter-wave (mmWave) aerial-terrestrial integrated networks with the following two unique characteristics. Firstly, FedMoD incorporates a novel decentralized model dissemination scheme that uses UAVs as local model aggregators through UAV-to-UAV and device-to-device (D2D) communications. As a result, FedMoD 1) increases the number of participant UDs in developing the FL model; and 2) achieves global model aggregation without involving CPS. Secondly, FedMoD reduces FL’s energy consumption using radio resource management (RRM) under the constraints of over-the-air learning latency. To achieve this, by leveraging graph theory, FedMoD optimizes the scheduling of line-of-sight (LOS) UDs to suitable UAVs and RRBs over mmWave links and non-LOS UDs to available LOS UDs via overlay D2D communications. Extensive simulations reveal that FedMoD, despite being decentralized, offers the same convergence performance to the conventional centralized FL frameworks.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1283-1304"},"PeriodicalIF":0.0,"publicationDate":"2024-06-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10550002","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142230893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-04DOI: 10.1109/TMLCN.2024.3409205
Eric Samikwa;Antonio Di Maio;Torsten Braun
Federated Learning (FL) in edge Internet of Things (IoT) environments is challenging due to the heterogeneous nature of the learning environment, mainly embodied in two aspects. Firstly, the statistically heterogeneous data, usually non-independent identically distributed (non-IID), from geographically distributed clients can deteriorate the FL training accuracy. Secondly, the heterogeneous computing and communication resources in IoT devices often result in unstable training processes that slow down the training of a global model and affect energy consumption. Most existing solutions address only the unilateral side of the heterogeneity issue but neglect the joint problem of resources and data heterogeneity for the resource-constrained IoT. In this article, we propose Dynamic Federated split Learning (DFL) to address the joint problem of data and resource heterogeneity for distributed training in IoT. DFL enhances training efficiency in heterogeneous dynamic IoT through resource-aware split computing of deep neural networks and dynamic clustering of training participants based on the similarity of their sub-model layers. We evaluate DFL on a real testbed comprising heterogeneous IoT devices using two widely-adopted datasets, in various non-IID settings. Results show that DFL improves training performance in terms of training time by up to 48%, accuracy by up to 32%, and energy consumption by up to 62.8% compared to classic FL and Federated Split Learning in scenarios with both data and resource heterogeneity.
{"title":"DFL: Dynamic Federated Split Learning in Heterogeneous IoT","authors":"Eric Samikwa;Antonio Di Maio;Torsten Braun","doi":"10.1109/TMLCN.2024.3409205","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3409205","url":null,"abstract":"Federated Learning (FL) in edge Internet of Things (IoT) environments is challenging due to the heterogeneous nature of the learning environment, mainly embodied in two aspects. Firstly, the statistically heterogeneous data, usually non-independent identically distributed (non-IID), from geographically distributed clients can deteriorate the FL training accuracy. Secondly, the heterogeneous computing and communication resources in IoT devices often result in unstable training processes that slow down the training of a global model and affect energy consumption. Most existing solutions address only the unilateral side of the heterogeneity issue but neglect the joint problem of resources and data heterogeneity for the resource-constrained IoT. In this article, we propose Dynamic Federated split Learning (DFL) to address the joint problem of data and resource heterogeneity for distributed training in IoT. DFL enhances training efficiency in heterogeneous dynamic IoT through resource-aware split computing of deep neural networks and dynamic clustering of training participants based on the similarity of their sub-model layers. We evaluate DFL on a real testbed comprising heterogeneous IoT devices using two widely-adopted datasets, in various non-IID settings. Results show that DFL improves training performance in terms of training time by up to 48%, accuracy by up to 32%, and energy consumption by up to 62.8% compared to classic FL and Federated Split Learning in scenarios with both data and resource heterogeneity.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"733-752"},"PeriodicalIF":0.0,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547401","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141319638","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-04DOI: 10.1109/TMLCN.2024.3409200
Shusen Jing;Songyang Zhang;Zhi Ding
Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.
{"title":"Reinforcement Learning for Robust Header Compression (ROHC) Under Model Uncertainty","authors":"Shusen Jing;Songyang Zhang;Zhi Ding","doi":"10.1109/TMLCN.2024.3409200","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3409200","url":null,"abstract":"Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1033-1044"},"PeriodicalIF":0.0,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547320","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141725562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-03DOI: 10.1109/TMLCN.2024.3408723
Maher Marwani;Georges Kaddoum
The proliferation of wireless technologies and the escalating performance requirements of wireless applications have led to diverse and dynamic wireless environments, presenting formidable challenges to existing radio resource management (RRM) frameworks. Researchers have proposed utilizing deep learning (DL) models to address these challenges to learn patterns from wireless data and leverage the extracted information to resolve multiple RRM tasks, such as channel allocation and power control. However, it is noteworthy that the majority of existing DL architectures are designed to operate on Euclidean data, thereby disregarding a substantial amount of information about the topological structure of wireless networks. As a result, the performance of DL models may be suboptimal when applied to wireless environments due to the failure to capture the network’s non-Euclidean geometry. This study presents a novel approach to address the challenge of power control and spectrum allocation in an N-link interference environment with shared channels, utilizing a graph neural network (GNN) based framework. In this type of wireless environment, the available bandwidth can be divided into blocks, offering greater flexibility in allocating bandwidth to communication links, but also requiring effective management of interference. One potential solution to mitigate the impact of interference is to control the transmission power of each link while ensuring the network’s data rate performance. Therefore, the power control and spectrum allocation problems are inherently coupled and should be solved jointly. The proposed GNN-based framework presents a promising avenue for tackling this complex challenge. Our experimental results demonstrate that our proposed approach yields significant improvements compared to other existing methods in terms of convergence, generalization, performance, and robustness, particularly in the context of an imperfect channel.
无线技术的普及和无线应用对性能要求的不断提高,导致了无线环境的多样化和动态化,给现有的无线资源管理(RRM)框架带来了严峻的挑战。研究人员提出利用深度学习(DL)模型来应对这些挑战,从无线数据中学习模式,并利用提取的信息来解决多种 RRM 任务,如信道分配和功率控制。然而,值得注意的是,大多数现有的深度学习架构都是针对欧几里得数据设计的,因此忽略了大量有关无线网络拓扑结构的信息。因此,在无线环境中应用 DL 模型时,由于无法捕捉网络的非欧几里得几何结构,其性能可能无法达到最佳。本研究提出了一种新方法,利用基于图神经网络(GNN)的框架来解决共享信道的 N 链路干扰环境中的功率控制和频谱分配难题。在这类无线环境中,可用带宽可被划分为多个区块,从而为通信链路的带宽分配提供了更大的灵活性,但同时也要求对干扰进行有效管理。减轻干扰影响的一个潜在解决方案是控制每个链路的传输功率,同时确保网络的数据速率性能。因此,功率控制和频谱分配问题本质上是耦合的,应联合解决。所提出的基于 GNN 的框架为应对这一复杂挑战提供了一个很有前景的途径。我们的实验结果表明,与其他现有方法相比,我们提出的方法在收敛性、泛化、性能和鲁棒性方面都有显著改进,尤其是在信道不完善的情况下。
{"title":"Graph Neural Networks Approach for Joint Wireless Power Control and Spectrum Allocation","authors":"Maher Marwani;Georges Kaddoum","doi":"10.1109/TMLCN.2024.3408723","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3408723","url":null,"abstract":"The proliferation of wireless technologies and the escalating performance requirements of wireless applications have led to diverse and dynamic wireless environments, presenting formidable challenges to existing radio resource management (RRM) frameworks. Researchers have proposed utilizing deep learning (DL) models to address these challenges to learn patterns from wireless data and leverage the extracted information to resolve multiple RRM tasks, such as channel allocation and power control. However, it is noteworthy that the majority of existing DL architectures are designed to operate on Euclidean data, thereby disregarding a substantial amount of information about the topological structure of wireless networks. As a result, the performance of DL models may be suboptimal when applied to wireless environments due to the failure to capture the network’s non-Euclidean geometry. This study presents a novel approach to address the challenge of power control and spectrum allocation in an N-link interference environment with shared channels, utilizing a graph neural network (GNN) based framework. In this type of wireless environment, the available bandwidth can be divided into blocks, offering greater flexibility in allocating bandwidth to communication links, but also requiring effective management of interference. One potential solution to mitigate the impact of interference is to control the transmission power of each link while ensuring the network’s data rate performance. Therefore, the power control and spectrum allocation problems are inherently coupled and should be solved jointly. The proposed GNN-based framework presents a promising avenue for tackling this complex challenge. Our experimental results demonstrate that our proposed approach yields significant improvements compared to other existing methods in terms of convergence, generalization, performance, and robustness, particularly in the context of an imperfect channel.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"717-732"},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10545547","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141298352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-06-03DOI: 10.1109/TMLCN.2024.3407691
David López-Pérez;Antonio De Domenico;Nicola Piovesan;Mérouane Debbah
The energy consumption of mobile networks poses a critical challenge. Mitigating this concern necessitates the deployment and optimization of network energy-saving solutions, such as carrier shutdown, to dynamically manage network resources. Traditional optimization approaches encounter complexity due to factors like the large number of cells, stochastic traffic, channel variations, and intricate trade-offs. This paper introduces the simulated reality of communication networks (SRCON) framework, a novel, data-driven modeling paradigm that harnesses live network data and employs a blend of machine learning (ML)- and expert-based models. These mix of models accurately characterizes the functioning of network components, and predicts network energy efficiency and user equipment (UE) quality of service for any energy carrier shutdown configuration in a specific network. Distinguishing itself from existing methods, SRCON eliminates the reliance on expensive expert knowledge, drive testing, or incomplete maps for predicting network performance. This paper details the pipeline employed by SRCON to decompose the large network energy efficiency modeling problem into ML- and expert-based submodels. It demonstrates how, by embracing stochasticity, and carefully crafting the relationship between such submodels, the overall computational complexity can be reduced and prediction accuracy enhanced. Results derived from real network data underscore the paradigm shift introduced by SRCON, showcasing significant gains over a state-of-the-art method used by a operator for network energy efficiency modeling. The reliability of this local, data-driven modeling of the network proves to be a key asset for network energy-saving optimization.
移动网络的能耗是一个严峻的挑战。为缓解这一问题,有必要部署和优化网络节能解决方案,如载波关闭,以动态管理网络资源。由于存在大量小区、随机流量、信道变化和复杂的权衡等因素,传统的优化方法非常复杂。本文介绍了通信网络模拟现实(SRCON)框架,这是一种新颖的数据驱动建模范例,它利用实时网络数据,并融合了基于机器学习(ML)和专家的模型。这些混合模型能准确描述网络组件的功能,并预测特定网络中任何能源载体关断配置的网络能效和用户设备(UE)的服务质量。有别于现有方法,SRCON 无需依赖昂贵的专家知识、驱动测试或不完整的地图来预测网络性能。本文详细介绍了 SRCON 将大型网络能效建模问题分解为基于 ML 和专家的子模型的过程。它展示了如何通过采用随机性和精心设计这些子模型之间的关系来降低总体计算复杂度和提高预测准确性。从真实网络数据中得出的结果突显了 SRCON 带来的模式转变,与一家运营商用于网络能效建模的最先进方法相比,SRCON 取得了显著的进步。事实证明,这种由数据驱动的本地网络建模的可靠性是网络节能优化的关键资产。
{"title":"Data-Driven Energy Efficiency Modeling in Large-Scale Networks: An Expert Knowledge and ML-Based Approach","authors":"David López-Pérez;Antonio De Domenico;Nicola Piovesan;Mérouane Debbah","doi":"10.1109/TMLCN.2024.3407691","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3407691","url":null,"abstract":"The energy consumption of mobile networks poses a critical challenge. Mitigating this concern necessitates the deployment and optimization of network energy-saving solutions, such as carrier shutdown, to dynamically manage network resources. Traditional optimization approaches encounter complexity due to factors like the large number of cells, stochastic traffic, channel variations, and intricate trade-offs. This paper introduces the simulated reality of communication networks (SRCON) framework, a novel, data-driven modeling paradigm that harnesses live network data and employs a blend of machine learning (ML)- and expert-based models. These mix of models accurately characterizes the functioning of network components, and predicts network energy efficiency and user equipment (UE) quality of service for any energy carrier shutdown configuration in a specific network. Distinguishing itself from existing methods, SRCON eliminates the reliance on expensive expert knowledge, drive testing, or incomplete maps for predicting network performance. This paper details the pipeline employed by SRCON to decompose the large network energy efficiency modeling problem into ML- and expert-based submodels. It demonstrates how, by embracing stochasticity, and carefully crafting the relationship between such submodels, the overall computational complexity can be reduced and prediction accuracy enhanced. Results derived from real network data underscore the paradigm shift introduced by SRCON, showcasing significant gains over a state-of-the-art method used by a operator for network energy efficiency modeling. The reliability of this local, data-driven modeling of the network proves to be a key asset for network energy-saving optimization.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"780-804"},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547043","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141429918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-30DOI: 10.1109/TMLCN.2024.3395419
Makhduma F. Saiyedand;Irfan Al-Anbagi
The upsurge of Internet of Things (IoT) devices has increased their vulnerability to Distributed Denial of Service (DDoS) attacks. DDoS attacks have evolved into complex multi-vector threats that high-volume and low-volume attack strategies, posing challenges for detection using traditional methods. These challenges highlight the importance of reliable detection and prevention measures. This paper introduces a novel Deep Ensemble learning with Pruning (DEEPShield) system, to efficiently detect both high- and low-volume DDoS attacks in resource-constrained environments. The DEEPShield system uses ensemble learning by integrating a Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) network with a network traffic analysis system. This system analyzes and preprocesses network traffic while being data-agnostic, resulting in high detection accuracy. In addition, the DEEPShield system applies unit pruning to refine ensemble models, optimizing them for deployment on edge devices while maintaining a balance between accuracy and computational efficiency. To address the lack of a detailed dataset for high- and low-volume DDoS attacks, this paper also introduces a dataset named HL-IoT, which includes both attack types. Furthermore, the testbed evaluation of the DEEPShield system under various load scenarios and network traffic loads showcases its effectiveness and robustness. Compared to the state-of-the-art deep ensembles and deep learning methods across various datasets, including HL-IoT, ToN-IoT, CICIDS-17, and ISCX-12, the DEEPShield system consistently achieves an accuracy over 90% for both DDoS attack types. Furthermore, the DEEPShield system achieves this performance with reduced memory and processing requirements, underscoring its adaptability for edge computing scenarios.
{"title":"Deep Ensemble Learning With Pruning for DDoS Attack Detection in IoT Networks","authors":"Makhduma F. Saiyedand;Irfan Al-Anbagi","doi":"10.1109/TMLCN.2024.3395419","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3395419","url":null,"abstract":"The upsurge of Internet of Things (IoT) devices has increased their vulnerability to Distributed Denial of Service (DDoS) attacks. DDoS attacks have evolved into complex multi-vector threats that high-volume and low-volume attack strategies, posing challenges for detection using traditional methods. These challenges highlight the importance of reliable detection and prevention measures. This paper introduces a novel Deep Ensemble learning with Pruning (DEEPShield) system, to efficiently detect both high- and low-volume DDoS attacks in resource-constrained environments. The DEEPShield system uses ensemble learning by integrating a Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) network with a network traffic analysis system. This system analyzes and preprocesses network traffic while being data-agnostic, resulting in high detection accuracy. In addition, the DEEPShield system applies unit pruning to refine ensemble models, optimizing them for deployment on edge devices while maintaining a balance between accuracy and computational efficiency. To address the lack of a detailed dataset for high- and low-volume DDoS attacks, this paper also introduces a dataset named HL-IoT, which includes both attack types. Furthermore, the testbed evaluation of the DEEPShield system under various load scenarios and network traffic loads showcases its effectiveness and robustness. Compared to the state-of-the-art deep ensembles and deep learning methods across various datasets, including HL-IoT, ToN-IoT, CICIDS-17, and ISCX-12, the DEEPShield system consistently achieves an accuracy over 90% for both DDoS attack types. Furthermore, the DEEPShield system achieves this performance with reduced memory and processing requirements, underscoring its adaptability for edge computing scenarios.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"596-616"},"PeriodicalIF":0.0,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10513369","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140895040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Federated Domain Generalization (FDG) aims to train a global model that generalizes well to new clients in a privacy-conscious manner, even when domain shifts are encountered. The increasing concerns of knowledge generalization and data privacy also challenge the traditional gather-and-analyze paradigm in networks. Recent investigations mainly focus on aggregation optimization and domain-invariant representations. However, without directly considering the data augmentation and leveraging the knowledge among existing domains, the domain-only data cannot guarantee the generalization ability of the FDG model when testing on the unseen domain. To overcome the problem, this paper proposes a distributed data augmentation method which combines Generative Adversarial Networks (GANs) and Federated Analytics (FA) to enhance the generalization ability of the trained FDG model, called FA-FDG. First, FA-FDG integrates GAN data generators from each Federated Learning (FL) client. Second, an evaluation index called generalization ability of domain (GAD) is proposed in the FA server. Then, the targeted data augmentation is implemented in each FL client with the GAD index and the integrated data generators. Extensive experiments on several data sets have shown the effectiveness of FA-FDG. Specifically, the accuracy of the FDG model improves up to 5.12% in classification problems, and the R-squared index of the FDG model advances up to 0.22 in the regression problem.
联合领域泛化(Federated Domain Generalization,FDG)旨在训练一个全局模型,即使在遇到领域转移时,该模型也能以注重隐私的方式很好地泛化到新客户。人们对知识泛化和数据隐私的关注与日俱增,这也对网络中传统的 "收集-分析 "模式提出了挑战。最近的研究主要集中在聚合优化和领域不变表示法上。然而,如果不直接考虑数据增强和利用现有领域间的知识,纯领域数据就无法保证 FDG 模型在未见领域进行测试时的泛化能力。为了克服这一问题,本文提出了一种结合生成对抗网络(GANs)和联合分析(FA)的分布式数据增强方法,以增强训练好的 FDG 模型的泛化能力,称为 FA-FDG。首先,FA-FDG 整合了每个联邦学习(FL)客户端的 GAN 数据生成器。其次,在 FA 服务器中提出了一个名为领域泛化能力(GAD)的评价指标。然后,利用 GAD 指数和集成的数据生成器,在每个 FL 客户端实施有针对性的数据增强。在多个数据集上进行的大量实验证明了 FA-FDG 的有效性。具体来说,在分类问题上,FDG 模型的准确率提高了 5.12%,在回归问题上,FDG 模型的 R 平方指数提高了 0.22。
{"title":"Federated Analytics With Data Augmentation in Domain Generalization Toward Future Networks","authors":"Xunzheng Zhang;Juan Marcelo Parra-Ullauri;Shadi Moazzeni;Xenofon Vasilakos;Reza Nejabati;Dimitra Simeonidou","doi":"10.1109/TMLCN.2024.3393892","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3393892","url":null,"abstract":"Federated Domain Generalization (FDG) aims to train a global model that generalizes well to new clients in a privacy-conscious manner, even when domain shifts are encountered. The increasing concerns of knowledge generalization and data privacy also challenge the traditional gather-and-analyze paradigm in networks. Recent investigations mainly focus on aggregation optimization and domain-invariant representations. However, without directly considering the data augmentation and leveraging the knowledge among existing domains, the domain-only data cannot guarantee the generalization ability of the FDG model when testing on the unseen domain. To overcome the problem, this paper proposes a distributed data augmentation method which combines Generative Adversarial Networks (GANs) and Federated Analytics (FA) to enhance the generalization ability of the trained FDG model, called FA-FDG. First, FA-FDG integrates GAN data generators from each Federated Learning (FL) client. Second, an evaluation index called generalization ability of domain (GAD) is proposed in the FA server. Then, the targeted data augmentation is implemented in each FL client with the GAD index and the integrated data generators. Extensive experiments on several data sets have shown the effectiveness of FA-FDG. Specifically, the accuracy of the FDG model improves up to 5.12% in classification problems, and the R-squared index of the FDG model advances up to 0.22 in the regression problem.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"560-579"},"PeriodicalIF":0.0,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10508396","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140818800","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-18DOI: 10.1109/TMLCN.2024.3391216
Zain Ali;Zouheir Rezki;Hamid Sadjadpour
Underlay Cognitive Radio (CR) systems were introduced to resolve the issue of spectrum scarcity in wireless communication. In CR systems, an unlicensed Secondary Transmitter (ST) shares the channel with a licensed Primary Transmitter (PT). Spectral efficiency of the CR systems can be further increased if multiple STs share the same channel. In underlay CR systems, the STs are required to keep interference at a low level to avoid outage at the primary system. The restriction on interference in underlay CR prevents some STs from transmitting while other STs may achieve high data rates, thus making the underlay CR network unfair. In this work, we consider the problem of achieving fairness in the rates of the STs. The considered optimization problem is non-convex in nature. The conventional iteration-based optimizers are time-consuming and may not converge when the considered problem is non-convex. To deal with the problem, we propose a deep-Q reinforcement learning (DQ-RL) framework that employs two separate deep neural networks for the computation and estimation of the Q-values which provides a fast solution and is robust to channel dynamic. The proposed technique achieves near optimal values of fairness while offering primary outage probability of less than 4%. Further, increasing the number of STs results in a linear increase in the computational complexity of the proposed framework. A comparison of several variants of the proposed scheme with the optimal solution is also presented. Finally, we present a novel cumulative reward framework and discuss how the combined-reward approach improves the performance of the communication system.
底层认知无线电(CR)系统的出现是为了解决无线通信中频谱稀缺的问题。在认知无线电系统中,未获得许可的二级发射机(ST)与获得许可的一级发射机(PT)共享信道。如果多个 ST 共享同一信道,则可进一步提高 CR 系统的频谱效率。在下层 CR 系统中,ST 必须将干扰控制在较低水平,以避免主系统中断。在底层 CR 中,对干扰的限制使一些 ST 无法进行传输,而其他 ST 则可能实现很高的数据传输速率,从而使底层 CR 网络变得不公平。在这项工作中,我们考虑了如何实现 ST 速率公平性的问题。所考虑的优化问题在本质上是非凸的。传统的基于迭代的优化器非常耗时,而且当所考虑的问题是非凸问题时可能无法收敛。为解决这一问题,我们提出了一种深度 Q 强化学习(DQ-RL)框架,该框架采用两个独立的深度神经网络来计算和估计 Q 值,可提供快速解决方案,并对信道动态具有鲁棒性。所提出的技术可实现接近最优的公平值,同时提供小于 4% 的主中断概率。此外,增加 ST 的数量会导致拟议框架的计算复杂度线性增加。此外,我们还对所提方案的几种变体与最优解决方案进行了比较。最后,我们提出了一个新颖的累积奖励框架,并讨论了综合奖励方法如何提高通信系统的性能。
{"title":"Multi-Agent Double Deep Q-Learning for Fairness in Multiple-Access Underlay Cognitive Radio Networks","authors":"Zain Ali;Zouheir Rezki;Hamid Sadjadpour","doi":"10.1109/TMLCN.2024.3391216","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3391216","url":null,"abstract":"Underlay Cognitive Radio (CR) systems were introduced to resolve the issue of spectrum scarcity in wireless communication. In CR systems, an unlicensed Secondary Transmitter (ST) shares the channel with a licensed Primary Transmitter (PT). Spectral efficiency of the CR systems can be further increased if multiple STs share the same channel. In underlay CR systems, the STs are required to keep interference at a low level to avoid outage at the primary system. The restriction on interference in underlay CR prevents some STs from transmitting while other STs may achieve high data rates, thus making the underlay CR network unfair. In this work, we consider the problem of achieving fairness in the rates of the STs. The considered optimization problem is non-convex in nature. The conventional iteration-based optimizers are time-consuming and may not converge when the considered problem is non-convex. To deal with the problem, we propose a deep-Q reinforcement learning (DQ-RL) framework that employs two separate deep neural networks for the computation and estimation of the Q-values which provides a fast solution and is robust to channel dynamic. The proposed technique achieves near optimal values of fairness while offering primary outage probability of less than 4%. Further, increasing the number of STs results in a linear increase in the computational complexity of the proposed framework. A comparison of several variants of the proposed scheme with the optimal solution is also presented. Finally, we present a novel cumulative reward framework and discuss how the combined-reward approach improves the performance of the communication system.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"580-595"},"PeriodicalIF":0.0,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10504881","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140820363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-17DOI: 10.1109/TMLCN.2024.3388975
Krishnendu S Tharakan;B. N. Bharath;Vimal Bhatia
An efficient caching can be achieved by predicting the popularity of the files accurately. It is well known that the popularity of a file can be nudged by using recommendation, and hence it can be estimated accurately leading to an efficient caching strategy. Motivated by this, in this paper, we consider the problem of joint caching and recommendation in a 5G and beyond heterogeneous network. We model the influence of recommendation on demands by a Probability Transition Matrix (PTM). The proposed framework consists of estimating the PTM and use them to jointly recommend and cache the files. In particular, this paper considers two estimation methods namely a) Bayesian estimation