首页 > 最新文献

IEEE Transactions on Machine Learning in Communications and Networking最新文献

英文 中文
Reinforcement Learning for Robust Header Compression (ROHC) Under Model Uncertainty 模型不确定情况下鲁棒性标题压缩(ROHC)的强化学习
Pub Date : 2024-06-04 DOI: 10.1109/TMLCN.2024.3409200
Shusen Jing;Songyang Zhang;Zhi Ding
Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.
稳健报头压缩(ROHC)位于网络层和 MAC 层之间,在现代无线通信网络中发挥着提高数据效率的重要作用。本研究将双向 ROHC(BD-ROHC)与强化学习(RL)的新架构相结合。我们提出了一个部分可观测马尔可夫决策过程(POMDP),其中压缩器是 POMDP 代理,环境由解压缩器、信道和头源组成。我们的工作采用了著名的深度 Q 网络(DQN),它将行动和观察的历史记录作为输入,并输出相应行动的 Q 值。与现有工作中提出的理想动态编程(DP)相比,新提出的方法可扩展到状态、行动和观察空间。相比之下,当状态数量变多时,由于解压反馈延迟较长和信道模型复杂,DP 通常会产生巨大的计算成本。此外,新方法不需要预先了解过渡动态和模型的精确观测依赖性,而这些在实际应用中往往是不具备的。
{"title":"Reinforcement Learning for Robust Header Compression (ROHC) Under Model Uncertainty","authors":"Shusen Jing;Songyang Zhang;Zhi Ding","doi":"10.1109/TMLCN.2024.3409200","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3409200","url":null,"abstract":"Robust header compression (ROHC), critically positioned between network and MAC layers, plays an important role in modern wireless communication networks for improving data efficiency. This work investigates bi-directional ROHC (BD-ROHC) integrated with a novel architecture of reinforcement learning (RL). We formulate a partially observable Markov decision process (POMDP), where the compressor is the POMDP agent, and the environment consists of the decompressor, channel, and header source. Our work adopts the well-known deep Q-network (DQN), which takes the history of actions and observations as inputs, and outputs the Q-values of corresponding actions. Compared with the ideal dynamic programming (DP) proposed in existing works, the newly proposed method is scalable to the state, action, and observation spaces. In contrast, DP often incurs formidable computation costs when the number of states becomes large due to long decompressor feedback delays and complex channel models. In addition, the new method does not require prior knowledge of the transition dynamics and accurate observation dependency of the model, which are often unavailable in practical applications.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"1033-1044"},"PeriodicalIF":0.0,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547320","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141725562","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Graph Neural Networks Approach for Joint Wireless Power Control and Spectrum Allocation 用于联合无线功率控制和频谱分配的图神经网络方法
Pub Date : 2024-06-03 DOI: 10.1109/TMLCN.2024.3408723
Maher Marwani;Georges Kaddoum
The proliferation of wireless technologies and the escalating performance requirements of wireless applications have led to diverse and dynamic wireless environments, presenting formidable challenges to existing radio resource management (RRM) frameworks. Researchers have proposed utilizing deep learning (DL) models to address these challenges to learn patterns from wireless data and leverage the extracted information to resolve multiple RRM tasks, such as channel allocation and power control. However, it is noteworthy that the majority of existing DL architectures are designed to operate on Euclidean data, thereby disregarding a substantial amount of information about the topological structure of wireless networks. As a result, the performance of DL models may be suboptimal when applied to wireless environments due to the failure to capture the network’s non-Euclidean geometry. This study presents a novel approach to address the challenge of power control and spectrum allocation in an N-link interference environment with shared channels, utilizing a graph neural network (GNN) based framework. In this type of wireless environment, the available bandwidth can be divided into blocks, offering greater flexibility in allocating bandwidth to communication links, but also requiring effective management of interference. One potential solution to mitigate the impact of interference is to control the transmission power of each link while ensuring the network’s data rate performance. Therefore, the power control and spectrum allocation problems are inherently coupled and should be solved jointly. The proposed GNN-based framework presents a promising avenue for tackling this complex challenge. Our experimental results demonstrate that our proposed approach yields significant improvements compared to other existing methods in terms of convergence, generalization, performance, and robustness, particularly in the context of an imperfect channel.
无线技术的普及和无线应用对性能要求的不断提高,导致了无线环境的多样化和动态化,给现有的无线资源管理(RRM)框架带来了严峻的挑战。研究人员提出利用深度学习(DL)模型来应对这些挑战,从无线数据中学习模式,并利用提取的信息来解决多种 RRM 任务,如信道分配和功率控制。然而,值得注意的是,大多数现有的深度学习架构都是针对欧几里得数据设计的,因此忽略了大量有关无线网络拓扑结构的信息。因此,在无线环境中应用 DL 模型时,由于无法捕捉网络的非欧几里得几何结构,其性能可能无法达到最佳。本研究提出了一种新方法,利用基于图神经网络(GNN)的框架来解决共享信道的 N 链路干扰环境中的功率控制和频谱分配难题。在这类无线环境中,可用带宽可被划分为多个区块,从而为通信链路的带宽分配提供了更大的灵活性,但同时也要求对干扰进行有效管理。减轻干扰影响的一个潜在解决方案是控制每个链路的传输功率,同时确保网络的数据速率性能。因此,功率控制和频谱分配问题本质上是耦合的,应联合解决。所提出的基于 GNN 的框架为应对这一复杂挑战提供了一个很有前景的途径。我们的实验结果表明,与其他现有方法相比,我们提出的方法在收敛性、泛化、性能和鲁棒性方面都有显著改进,尤其是在信道不完善的情况下。
{"title":"Graph Neural Networks Approach for Joint Wireless Power Control and Spectrum Allocation","authors":"Maher Marwani;Georges Kaddoum","doi":"10.1109/TMLCN.2024.3408723","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3408723","url":null,"abstract":"The proliferation of wireless technologies and the escalating performance requirements of wireless applications have led to diverse and dynamic wireless environments, presenting formidable challenges to existing radio resource management (RRM) frameworks. Researchers have proposed utilizing deep learning (DL) models to address these challenges to learn patterns from wireless data and leverage the extracted information to resolve multiple RRM tasks, such as channel allocation and power control. However, it is noteworthy that the majority of existing DL architectures are designed to operate on Euclidean data, thereby disregarding a substantial amount of information about the topological structure of wireless networks. As a result, the performance of DL models may be suboptimal when applied to wireless environments due to the failure to capture the network’s non-Euclidean geometry. This study presents a novel approach to address the challenge of power control and spectrum allocation in an N-link interference environment with shared channels, utilizing a graph neural network (GNN) based framework. In this type of wireless environment, the available bandwidth can be divided into blocks, offering greater flexibility in allocating bandwidth to communication links, but also requiring effective management of interference. One potential solution to mitigate the impact of interference is to control the transmission power of each link while ensuring the network’s data rate performance. Therefore, the power control and spectrum allocation problems are inherently coupled and should be solved jointly. The proposed GNN-based framework presents a promising avenue for tackling this complex challenge. Our experimental results demonstrate that our proposed approach yields significant improvements compared to other existing methods in terms of convergence, generalization, performance, and robustness, particularly in the context of an imperfect channel.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"717-732"},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10545547","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141298352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Data-Driven Energy Efficiency Modeling in Large-Scale Networks: An Expert Knowledge and ML-Based Approach 大规模网络中数据驱动的能效建模:基于专家知识和 ML 的方法
Pub Date : 2024-06-03 DOI: 10.1109/TMLCN.2024.3407691
David López-Pérez;Antonio De Domenico;Nicola Piovesan;Mérouane Debbah
The energy consumption of mobile networks poses a critical challenge. Mitigating this concern necessitates the deployment and optimization of network energy-saving solutions, such as carrier shutdown, to dynamically manage network resources. Traditional optimization approaches encounter complexity due to factors like the large number of cells, stochastic traffic, channel variations, and intricate trade-offs. This paper introduces the simulated reality of communication networks (SRCON) framework, a novel, data-driven modeling paradigm that harnesses live network data and employs a blend of machine learning (ML)- and expert-based models. These mix of models accurately characterizes the functioning of network components, and predicts network energy efficiency and user equipment (UE) quality of service for any energy carrier shutdown configuration in a specific network. Distinguishing itself from existing methods, SRCON eliminates the reliance on expensive expert knowledge, drive testing, or incomplete maps for predicting network performance. This paper details the pipeline employed by SRCON to decompose the large network energy efficiency modeling problem into ML- and expert-based submodels. It demonstrates how, by embracing stochasticity, and carefully crafting the relationship between such submodels, the overall computational complexity can be reduced and prediction accuracy enhanced. Results derived from real network data underscore the paradigm shift introduced by SRCON, showcasing significant gains over a state-of-the-art method used by a operator for network energy efficiency modeling. The reliability of this local, data-driven modeling of the network proves to be a key asset for network energy-saving optimization.
移动网络的能耗是一个严峻的挑战。为缓解这一问题,有必要部署和优化网络节能解决方案,如载波关闭,以动态管理网络资源。由于存在大量小区、随机流量、信道变化和复杂的权衡等因素,传统的优化方法非常复杂。本文介绍了通信网络模拟现实(SRCON)框架,这是一种新颖的数据驱动建模范例,它利用实时网络数据,并融合了基于机器学习(ML)和专家的模型。这些混合模型能准确描述网络组件的功能,并预测特定网络中任何能源载体关断配置的网络能效和用户设备(UE)的服务质量。有别于现有方法,SRCON 无需依赖昂贵的专家知识、驱动测试或不完整的地图来预测网络性能。本文详细介绍了 SRCON 将大型网络能效建模问题分解为基于 ML 和专家的子模型的过程。它展示了如何通过采用随机性和精心设计这些子模型之间的关系来降低总体计算复杂度和提高预测准确性。从真实网络数据中得出的结果突显了 SRCON 带来的模式转变,与一家运营商用于网络能效建模的最先进方法相比,SRCON 取得了显著的进步。事实证明,这种由数据驱动的本地网络建模的可靠性是网络节能优化的关键资产。
{"title":"Data-Driven Energy Efficiency Modeling in Large-Scale Networks: An Expert Knowledge and ML-Based Approach","authors":"David López-Pérez;Antonio De Domenico;Nicola Piovesan;Mérouane Debbah","doi":"10.1109/TMLCN.2024.3407691","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3407691","url":null,"abstract":"The energy consumption of mobile networks poses a critical challenge. Mitigating this concern necessitates the deployment and optimization of network energy-saving solutions, such as carrier shutdown, to dynamically manage network resources. Traditional optimization approaches encounter complexity due to factors like the large number of cells, stochastic traffic, channel variations, and intricate trade-offs. This paper introduces the simulated reality of communication networks (SRCON) framework, a novel, data-driven modeling paradigm that harnesses live network data and employs a blend of machine learning (ML)- and expert-based models. These mix of models accurately characterizes the functioning of network components, and predicts network energy efficiency and user equipment (UE) quality of service for any energy carrier shutdown configuration in a specific network. Distinguishing itself from existing methods, SRCON eliminates the reliance on expensive expert knowledge, drive testing, or incomplete maps for predicting network performance. This paper details the pipeline employed by SRCON to decompose the large network energy efficiency modeling problem into ML- and expert-based submodels. It demonstrates how, by embracing stochasticity, and carefully crafting the relationship between such submodels, the overall computational complexity can be reduced and prediction accuracy enhanced. Results derived from real network data underscore the paradigm shift introduced by SRCON, showcasing significant gains over a state-of-the-art method used by a operator for network energy efficiency modeling. The reliability of this local, data-driven modeling of the network proves to be a key asset for network energy-saving optimization.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"780-804"},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10547043","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141429918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Ensemble Learning With Pruning for DDoS Attack Detection in IoT Networks 利用剪枝深度集合学习检测物联网网络中的 DDoS 攻击
Pub Date : 2024-04-30 DOI: 10.1109/TMLCN.2024.3395419
Makhduma F. Saiyedand;Irfan Al-Anbagi
The upsurge of Internet of Things (IoT) devices has increased their vulnerability to Distributed Denial of Service (DDoS) attacks. DDoS attacks have evolved into complex multi-vector threats that high-volume and low-volume attack strategies, posing challenges for detection using traditional methods. These challenges highlight the importance of reliable detection and prevention measures. This paper introduces a novel Deep Ensemble learning with Pruning (DEEPShield) system, to efficiently detect both high- and low-volume DDoS attacks in resource-constrained environments. The DEEPShield system uses ensemble learning by integrating a Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) network with a network traffic analysis system. This system analyzes and preprocesses network traffic while being data-agnostic, resulting in high detection accuracy. In addition, the DEEPShield system applies unit pruning to refine ensemble models, optimizing them for deployment on edge devices while maintaining a balance between accuracy and computational efficiency. To address the lack of a detailed dataset for high- and low-volume DDoS attacks, this paper also introduces a dataset named HL-IoT, which includes both attack types. Furthermore, the testbed evaluation of the DEEPShield system under various load scenarios and network traffic loads showcases its effectiveness and robustness. Compared to the state-of-the-art deep ensembles and deep learning methods across various datasets, including HL-IoT, ToN-IoT, CICIDS-17, and ISCX-12, the DEEPShield system consistently achieves an accuracy over 90% for both DDoS attack types. Furthermore, the DEEPShield system achieves this performance with reduced memory and processing requirements, underscoring its adaptability for edge computing scenarios.
物联网(IoT)设备的激增使其更容易受到分布式拒绝服务(DDoS)攻击。DDoS 攻击已演变成复杂的多载体威胁,采用大流量和小流量攻击策略,给使用传统方法进行检测带来了挑战。这些挑战凸显了可靠的检测和预防措施的重要性。本文介绍了一种新颖的带剪枝功能的深度集合学习(DEEPShield)系统,可在资源受限的环境中高效地检测大流量和小流量 DDoS 攻击。DEEPShield 系统通过将卷积神经网络(CNN)和长短期记忆(LSTM)网络与网络流量分析系统集成在一起,使用了集合学习技术。该系统可分析和预处理网络流量,同时不依赖数据,因此检测准确率很高。此外,DEEPShield 系统还应用单元剪枝来完善集合模型,在保持准确性和计算效率之间平衡的同时,优化这些模型,以便在边缘设备上部署。由于缺乏针对大流量和小流量 DDoS 攻击的详细数据集,本文还引入了一个名为 HL-IoT 的数据集,其中包括这两种攻击类型。此外,在各种负载场景和网络流量负载下对 DEEPShield 系统进行的测试平台评估展示了其有效性和鲁棒性。在各种数据集(包括 HL-IoT、ToN-IoT、CICIDS-17 和 ISCX-12)上,与最先进的深度集合和深度学习方法相比,DEEPShield 系统对两种 DDoS 攻击类型的准确率始终保持在 90% 以上。此外,DEEPShield 系统在降低内存和处理要求的情况下实现了这一性能,凸显了其对边缘计算场景的适应性。
{"title":"Deep Ensemble Learning With Pruning for DDoS Attack Detection in IoT Networks","authors":"Makhduma F. Saiyedand;Irfan Al-Anbagi","doi":"10.1109/TMLCN.2024.3395419","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3395419","url":null,"abstract":"The upsurge of Internet of Things (IoT) devices has increased their vulnerability to Distributed Denial of Service (DDoS) attacks. DDoS attacks have evolved into complex multi-vector threats that high-volume and low-volume attack strategies, posing challenges for detection using traditional methods. These challenges highlight the importance of reliable detection and prevention measures. This paper introduces a novel Deep Ensemble learning with Pruning (DEEPShield) system, to efficiently detect both high- and low-volume DDoS attacks in resource-constrained environments. The DEEPShield system uses ensemble learning by integrating a Convolutional Neural Network (CNN) and a Long Short-Term Memory (LSTM) network with a network traffic analysis system. This system analyzes and preprocesses network traffic while being data-agnostic, resulting in high detection accuracy. In addition, the DEEPShield system applies unit pruning to refine ensemble models, optimizing them for deployment on edge devices while maintaining a balance between accuracy and computational efficiency. To address the lack of a detailed dataset for high- and low-volume DDoS attacks, this paper also introduces a dataset named HL-IoT, which includes both attack types. Furthermore, the testbed evaluation of the DEEPShield system under various load scenarios and network traffic loads showcases its effectiveness and robustness. Compared to the state-of-the-art deep ensembles and deep learning methods across various datasets, including HL-IoT, ToN-IoT, CICIDS-17, and ISCX-12, the DEEPShield system consistently achieves an accuracy over 90% for both DDoS attack types. Furthermore, the DEEPShield system achieves this performance with reduced memory and processing requirements, underscoring its adaptability for edge computing scenarios.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"596-616"},"PeriodicalIF":0.0,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10513369","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140895040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Federated Analytics With Data Augmentation in Domain Generalization Toward Future Networks 面向未来网络的领域泛化中的联合分析与数据增强
Pub Date : 2024-04-25 DOI: 10.1109/TMLCN.2024.3393892
Xunzheng Zhang;Juan Marcelo Parra-Ullauri;Shadi Moazzeni;Xenofon Vasilakos;Reza Nejabati;Dimitra Simeonidou
Federated Domain Generalization (FDG) aims to train a global model that generalizes well to new clients in a privacy-conscious manner, even when domain shifts are encountered. The increasing concerns of knowledge generalization and data privacy also challenge the traditional gather-and-analyze paradigm in networks. Recent investigations mainly focus on aggregation optimization and domain-invariant representations. However, without directly considering the data augmentation and leveraging the knowledge among existing domains, the domain-only data cannot guarantee the generalization ability of the FDG model when testing on the unseen domain. To overcome the problem, this paper proposes a distributed data augmentation method which combines Generative Adversarial Networks (GANs) and Federated Analytics (FA) to enhance the generalization ability of the trained FDG model, called FA-FDG. First, FA-FDG integrates GAN data generators from each Federated Learning (FL) client. Second, an evaluation index called generalization ability of domain (GAD) is proposed in the FA server. Then, the targeted data augmentation is implemented in each FL client with the GAD index and the integrated data generators. Extensive experiments on several data sets have shown the effectiveness of FA-FDG. Specifically, the accuracy of the FDG model improves up to 5.12% in classification problems, and the R-squared index of the FDG model advances up to 0.22 in the regression problem.
联合领域泛化(Federated Domain Generalization,FDG)旨在训练一个全局模型,即使在遇到领域转移时,该模型也能以注重隐私的方式很好地泛化到新客户。人们对知识泛化和数据隐私的关注与日俱增,这也对网络中传统的 "收集-分析 "模式提出了挑战。最近的研究主要集中在聚合优化和领域不变表示法上。然而,如果不直接考虑数据增强和利用现有领域间的知识,纯领域数据就无法保证 FDG 模型在未见领域进行测试时的泛化能力。为了克服这一问题,本文提出了一种结合生成对抗网络(GANs)和联合分析(FA)的分布式数据增强方法,以增强训练好的 FDG 模型的泛化能力,称为 FA-FDG。首先,FA-FDG 整合了每个联邦学习(FL)客户端的 GAN 数据生成器。其次,在 FA 服务器中提出了一个名为领域泛化能力(GAD)的评价指标。然后,利用 GAD 指数和集成的数据生成器,在每个 FL 客户端实施有针对性的数据增强。在多个数据集上进行的大量实验证明了 FA-FDG 的有效性。具体来说,在分类问题上,FDG 模型的准确率提高了 5.12%,在回归问题上,FDG 模型的 R 平方指数提高了 0.22。
{"title":"Federated Analytics With Data Augmentation in Domain Generalization Toward Future Networks","authors":"Xunzheng Zhang;Juan Marcelo Parra-Ullauri;Shadi Moazzeni;Xenofon Vasilakos;Reza Nejabati;Dimitra Simeonidou","doi":"10.1109/TMLCN.2024.3393892","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3393892","url":null,"abstract":"Federated Domain Generalization (FDG) aims to train a global model that generalizes well to new clients in a privacy-conscious manner, even when domain shifts are encountered. The increasing concerns of knowledge generalization and data privacy also challenge the traditional gather-and-analyze paradigm in networks. Recent investigations mainly focus on aggregation optimization and domain-invariant representations. However, without directly considering the data augmentation and leveraging the knowledge among existing domains, the domain-only data cannot guarantee the generalization ability of the FDG model when testing on the unseen domain. To overcome the problem, this paper proposes a distributed data augmentation method which combines Generative Adversarial Networks (GANs) and Federated Analytics (FA) to enhance the generalization ability of the trained FDG model, called FA-FDG. First, FA-FDG integrates GAN data generators from each Federated Learning (FL) client. Second, an evaluation index called generalization ability of domain (GAD) is proposed in the FA server. Then, the targeted data augmentation is implemented in each FL client with the GAD index and the integrated data generators. Extensive experiments on several data sets have shown the effectiveness of FA-FDG. Specifically, the accuracy of the FDG model improves up to 5.12% in classification problems, and the R-squared index of the FDG model advances up to 0.22 in the regression problem.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"560-579"},"PeriodicalIF":0.0,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10508396","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140818800","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Agent Double Deep Q-Learning for Fairness in Multiple-Access Underlay Cognitive Radio Networks 多代理双深度 Q 学习促进多重接入下层认知无线电网络的公平性
Pub Date : 2024-04-18 DOI: 10.1109/TMLCN.2024.3391216
Zain Ali;Zouheir Rezki;Hamid Sadjadpour
Underlay Cognitive Radio (CR) systems were introduced to resolve the issue of spectrum scarcity in wireless communication. In CR systems, an unlicensed Secondary Transmitter (ST) shares the channel with a licensed Primary Transmitter (PT). Spectral efficiency of the CR systems can be further increased if multiple STs share the same channel. In underlay CR systems, the STs are required to keep interference at a low level to avoid outage at the primary system. The restriction on interference in underlay CR prevents some STs from transmitting while other STs may achieve high data rates, thus making the underlay CR network unfair. In this work, we consider the problem of achieving fairness in the rates of the STs. The considered optimization problem is non-convex in nature. The conventional iteration-based optimizers are time-consuming and may not converge when the considered problem is non-convex. To deal with the problem, we propose a deep-Q reinforcement learning (DQ-RL) framework that employs two separate deep neural networks for the computation and estimation of the Q-values which provides a fast solution and is robust to channel dynamic. The proposed technique achieves near optimal values of fairness while offering primary outage probability of less than 4%. Further, increasing the number of STs results in a linear increase in the computational complexity of the proposed framework. A comparison of several variants of the proposed scheme with the optimal solution is also presented. Finally, we present a novel cumulative reward framework and discuss how the combined-reward approach improves the performance of the communication system.
底层认知无线电(CR)系统的出现是为了解决无线通信中频谱稀缺的问题。在认知无线电系统中,未获得许可的二级发射机(ST)与获得许可的一级发射机(PT)共享信道。如果多个 ST 共享同一信道,则可进一步提高 CR 系统的频谱效率。在下层 CR 系统中,ST 必须将干扰控制在较低水平,以避免主系统中断。在底层 CR 中,对干扰的限制使一些 ST 无法进行传输,而其他 ST 则可能实现很高的数据传输速率,从而使底层 CR 网络变得不公平。在这项工作中,我们考虑了如何实现 ST 速率公平性的问题。所考虑的优化问题在本质上是非凸的。传统的基于迭代的优化器非常耗时,而且当所考虑的问题是非凸问题时可能无法收敛。为解决这一问题,我们提出了一种深度 Q 强化学习(DQ-RL)框架,该框架采用两个独立的深度神经网络来计算和估计 Q 值,可提供快速解决方案,并对信道动态具有鲁棒性。所提出的技术可实现接近最优的公平值,同时提供小于 4% 的主中断概率。此外,增加 ST 的数量会导致拟议框架的计算复杂度线性增加。此外,我们还对所提方案的几种变体与最优解决方案进行了比较。最后,我们提出了一个新颖的累积奖励框架,并讨论了综合奖励方法如何提高通信系统的性能。
{"title":"Multi-Agent Double Deep Q-Learning for Fairness in Multiple-Access Underlay Cognitive Radio Networks","authors":"Zain Ali;Zouheir Rezki;Hamid Sadjadpour","doi":"10.1109/TMLCN.2024.3391216","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3391216","url":null,"abstract":"Underlay Cognitive Radio (CR) systems were introduced to resolve the issue of spectrum scarcity in wireless communication. In CR systems, an unlicensed Secondary Transmitter (ST) shares the channel with a licensed Primary Transmitter (PT). Spectral efficiency of the CR systems can be further increased if multiple STs share the same channel. In underlay CR systems, the STs are required to keep interference at a low level to avoid outage at the primary system. The restriction on interference in underlay CR prevents some STs from transmitting while other STs may achieve high data rates, thus making the underlay CR network unfair. In this work, we consider the problem of achieving fairness in the rates of the STs. The considered optimization problem is non-convex in nature. The conventional iteration-based optimizers are time-consuming and may not converge when the considered problem is non-convex. To deal with the problem, we propose a deep-Q reinforcement learning (DQ-RL) framework that employs two separate deep neural networks for the computation and estimation of the Q-values which provides a fast solution and is robust to channel dynamic. The proposed technique achieves near optimal values of fairness while offering primary outage probability of less than 4%. Further, increasing the number of STs results in a linear increase in the computational complexity of the proposed framework. A comparison of several variants of the proposed scheme with the optimal solution is also presented. Finally, we present a novel cumulative reward framework and discuss how the combined-reward approach improves the performance of the communication system.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"580-595"},"PeriodicalIF":0.0,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10504881","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140820363","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Online Learning to Cache and Recommend in the Next Generation Cellular Networks 通过在线学习在下一代蜂窝网络中进行缓存和推荐
Pub Date : 2024-04-17 DOI: 10.1109/TMLCN.2024.3388975
Krishnendu S Tharakan;B. N. Bharath;Vimal Bhatia
An efficient caching can be achieved by predicting the popularity of the files accurately. It is well known that the popularity of a file can be nudged by using recommendation, and hence it can be estimated accurately leading to an efficient caching strategy. Motivated by this, in this paper, we consider the problem of joint caching and recommendation in a 5G and beyond heterogeneous network. We model the influence of recommendation on demands by a Probability Transition Matrix (PTM). The proposed framework consists of estimating the PTM and use them to jointly recommend and cache the files. In particular, this paper considers two estimation methods namely a) Bayesian estimation and b) a genie aided Point estimation. An approximate high probability bound on the regret of both the estimation methods are provided. Using this result, we show that the approximate regret achieved by the genie aided Point estimation approach is $mathcal {O}(T^{2/3} sqrt {log T})$ while the Bayesian estimation method achieves a much better scaling of $mathcal {O}(sqrt {T})$ . These results are extended to a heterogeneous network consisting of M small base stations (SBSs) with a central macro base station. The estimates are available at multiple SBSs, and are combined using appropriate weights. Insights on the choice of these weights are provided by using the derived approximate regret bound in the multiple SBS case. Finally, simulation results confirm the superiority of the proposed algorithms in terms of average cache hit rate, delay and throughput.
准确预测文件的受欢迎程度可以实现高效缓存。众所周知,文件的受欢迎程度可以通过推荐来推测,因此可以准确地估计文件的受欢迎程度,从而制定高效的缓存策略。受此启发,我们在本文中考虑了在 5G 及以上异构网络中联合缓存和推荐的问题。我们通过概率转换矩阵(PTM)来模拟推荐对需求的影响。所提出的框架包括估计 PTM,并利用它们来联合推荐和缓存文件。本文特别考虑了两种估算方法,即 a) 贝叶斯估算和 b) 精灵辅助点估算。本文提供了两种估算方法遗憾值的近似高概率约束。利用这一结果,我们表明精灵辅助点估计方法实现的近似遗憾值为 $mathcal {O}(T^{2/3} sqrt {log T})$,而贝叶斯估计方法实现了更好的缩放,为 $mathcal {O}(sqrt {T})$ 。这些结果被扩展到由 M 个小型基站 (SBS) 和一个中央宏基站组成的异构网络。多个 SBS 均可提供估计值,并使用适当的权重进行组合。在多个 SBS 的情况下,通过使用推导出的近似遗憾约束,可以深入了解这些权重的选择。最后,模拟结果证实了所提算法在平均缓存命中率、延迟和吞吐量方面的优越性。
{"title":"Online Learning to Cache and Recommend in the Next Generation Cellular Networks","authors":"Krishnendu S Tharakan;B. N. Bharath;Vimal Bhatia","doi":"10.1109/TMLCN.2024.3388975","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3388975","url":null,"abstract":"An efficient caching can be achieved by predicting the popularity of the files accurately. It is well known that the popularity of a file can be nudged by using recommendation, and hence it can be estimated accurately leading to an efficient caching strategy. Motivated by this, in this paper, we consider the problem of joint caching and recommendation in a 5G and beyond heterogeneous network. We model the influence of recommendation on demands by a Probability Transition Matrix (PTM). The proposed framework consists of estimating the PTM and use them to jointly recommend and cache the files. In particular, this paper considers two estimation methods namely a) \u0000<monospace>Bayesian estimation</monospace>\u0000 and b) a genie aided \u0000<monospace>Point estimation</monospace>\u0000. An approximate high probability bound on the regret of both the estimation methods are provided. Using this result, we show that the approximate regret achieved by the genie aided \u0000<monospace>Point estimation</monospace>\u0000 approach is \u0000<inline-formula> <tex-math>$mathcal {O}(T^{2/3} sqrt {log T})$ </tex-math></inline-formula>\u0000 while the \u0000<monospace>Bayesian estimation</monospace>\u0000 method achieves a much better scaling of \u0000<inline-formula> <tex-math>$mathcal {O}(sqrt {T})$ </tex-math></inline-formula>\u0000. These results are extended to a heterogeneous network consisting of M small base stations (SBSs) with a central macro base station. The estimates are available at multiple SBSs, and are combined using appropriate weights. Insights on the choice of these weights are provided by using the derived approximate regret bound in the multiple SBS case. Finally, simulation results confirm the superiority of the proposed algorithms in terms of average cache hit rate, delay and throughput.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"511-525"},"PeriodicalIF":0.0,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10504600","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140639381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Link-Quality Anomaly Detection Framework for Software-Defined Wireless Mesh Networks 软件定义无线网格网络的链路质量异常检测框架
Pub Date : 2024-04-15 DOI: 10.1109/TMLCN.2024.3388973
Sotiris Skaperas;Lefteris Mamatas;Vassilis Tsaoussidis
Software-defined wireless mesh networks are being increasingly deployed in diverse settings, such as smart cities and public Wi-Fi access infrastructures. The signal propagation and interference issues that typically characterize these environments can be handled by employing SDN controller mechanisms, effectively monitoring link quality and triggering appropriate mitigation strategies, such as adjusting link and/or routing protocols. In this paper, we propose an unsupervised machine learning (ML) online framework for link quality detection consisting of: 1) improved preprocessing clustering algorithm, based on elastic similarity measures, to efficiently characterize wireless links in terms of reliability, and 2) a novel change point (CP) detector for the real-time identification of anomalies in the quality of selected links, which minimizes the overestimation error through the incorporation of a rank-based test and a recursive max-type procedure. In this sense, considering the communication constraints of such environments, our approach minimizes the detection overhead and the inaccurate decisions caused by overestimation. The proposed detector is validated, both on its individual components and as an overall mechanism, against synthetic but also real data traces; the latter being extracted from real wireless mesh network deployments.
软件定义的无线网格网络正越来越多地部署在各种环境中,如智能城市和公共 Wi-Fi 接入基础设施。信号传播和干扰问题是这些环境的典型特征,可通过采用 SDN 控制器机制来处理,有效监控链路质量并触发适当的缓解策略,如调整链路和/或路由协议。在本文中,我们提出了一种用于链路质量检测的无监督机器学习(ML)在线框架,该框架由以下部分组成:1) 基于弹性相似度量的改进型预处理聚类算法,可有效描述无线链路的可靠性特征;以及 2) 用于实时识别选定链路质量异常的新型变化点(CP)检测器,该检测器通过基于等级的测试和递归最大值类型程序将高估误差降至最低。从这个意义上说,考虑到此类环境的通信限制,我们的方法最大限度地减少了检测开销和高估造成的不准确决策。我们对所提出的检测器进行了验证,无论是对其单个组件还是作为一个整体机制,都进行了合成和真实数据追踪;后者是从真实的无线网状网络部署中提取的。
{"title":"A Link-Quality Anomaly Detection Framework for Software-Defined Wireless Mesh Networks","authors":"Sotiris Skaperas;Lefteris Mamatas;Vassilis Tsaoussidis","doi":"10.1109/TMLCN.2024.3388973","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3388973","url":null,"abstract":"Software-defined wireless mesh networks are being increasingly deployed in diverse settings, such as smart cities and public Wi-Fi access infrastructures. The signal propagation and interference issues that typically characterize these environments can be handled by employing SDN controller mechanisms, effectively monitoring link quality and triggering appropriate mitigation strategies, such as adjusting link and/or routing protocols. In this paper, we propose an unsupervised machine learning (ML) online framework for link quality detection consisting of: 1) improved preprocessing clustering algorithm, based on elastic similarity measures, to efficiently characterize wireless links in terms of reliability, and 2) a novel change point (CP) detector for the real-time identification of anomalies in the quality of selected links, which minimizes the overestimation error through the incorporation of a rank-based test and a recursive max-type procedure. In this sense, considering the communication constraints of such environments, our approach minimizes the detection overhead and the inaccurate decisions caused by overestimation. The proposed detector is validated, both on its individual components and as an overall mechanism, against synthetic but also real data traces; the latter being extracted from real wireless mesh network deployments.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"495-510"},"PeriodicalIF":0.0,"publicationDate":"2024-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10499246","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140639411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
STTMC: A Few-Shot Spatial Temporal Transductive Modulation Classifier STTMC: 少量时空传导调制分类器
Pub Date : 2024-04-11 DOI: 10.1109/TMLCN.2024.3387430
Yunhao Shi;Hua Xu;Zisen Qi;Yue Zhang;Dan Wang;Lei Jiang
The advancement of deep learning (DL) techniques has led to significant progress in Automatic Modulation Classification (AMC). However, most existing DL-based AMC methods require massive training samples, which are difficult to obtain in non-cooperative scenarios. The identification of modulation types under small sample conditions has become an increasingly urgent problem. In this paper, we present a novel few-shot AMC model named the Spatial Temporal Transductive Modulation Classifier (STTMC), which comprises two modules: a feature extraction module and a graph network module. The former is responsible for extracting diverse features through a spatiotemporal parallel network, while the latter facilitates transductive decision-making through a graph network that uses a closed-form solution. Notably, STTMC classifies a group of test signals simultaneously to increase stability of few-shot model with an episode training strategy. Experimental results on the RadioML.2018.01A and RadioML.2016.10A datasets demonstrate that the proposed method perform well in 3way-Kshot, 5way-Kshot and 10way-Kshot configurations. In particular, STTMC outperforms other existing AMC methods by a large margin.
随着深度学习(DL)技术的发展,自动调制分类(AMC)取得了重大进展。然而,大多数现有的基于深度学习的自动调制分类方法都需要大量的训练样本,而这些样本在非合作场景下很难获得。在小样本条件下识别调制类型已成为一个日益紧迫的问题。在本文中,我们提出了一种名为空间时空传导调制分类器(STTMC)的新型少镜头 AMC 模型,它由两个模块组成:特征提取模块和图网络模块。前者负责通过时空并行网络提取各种特征,后者则通过使用闭式解的图网络促进归纳决策。值得注意的是,STTMC 可同时对一组测试信号进行分类,以提高采用插集训练策略的少拍模型的稳定性。在 RadioML.2018.01A 和 RadioML.2016.10A 数据集上的实验结果表明,所提出的方法在 3way-Kshot、5way-Kshot 和 10way-Kshot 配置中表现良好。特别是,STTMC 在很大程度上优于其他现有的 AMC 方法。
{"title":"STTMC: A Few-Shot Spatial Temporal Transductive Modulation Classifier","authors":"Yunhao Shi;Hua Xu;Zisen Qi;Yue Zhang;Dan Wang;Lei Jiang","doi":"10.1109/TMLCN.2024.3387430","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3387430","url":null,"abstract":"The advancement of deep learning (DL) techniques has led to significant progress in Automatic Modulation Classification (AMC). However, most existing DL-based AMC methods require massive training samples, which are difficult to obtain in non-cooperative scenarios. The identification of modulation types under small sample conditions has become an increasingly urgent problem. In this paper, we present a novel few-shot AMC model named the Spatial Temporal Transductive Modulation Classifier (STTMC), which comprises two modules: a feature extraction module and a graph network module. The former is responsible for extracting diverse features through a spatiotemporal parallel network, while the latter facilitates transductive decision-making through a graph network that uses a closed-form solution. Notably, STTMC classifies a group of test signals simultaneously to increase stability of few-shot model with an episode training strategy. Experimental results on the RadioML.2018.01A and RadioML.2016.10A datasets demonstrate that the proposed method perform well in 3way-Kshot, 5way-Kshot and 10way-Kshot configurations. In particular, STTMC outperforms other existing AMC methods by a large margin.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"546-559"},"PeriodicalIF":0.0,"publicationDate":"2024-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10497130","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140648000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Deep Learning Based Induced GNSS Spoof Detection Framework 基于深度学习的诱导式全球导航卫星系统欺骗检测框架
Pub Date : 2024-04-10 DOI: 10.1109/TMLCN.2024.3386649
Asif Iqbal;Muhammad Naveed Aman;Biplab Sikdar
The Global Navigation Satellite System (GNSS) plays a crucial role in critical infrastructure by delivering precise timing and positional data. Nonetheless, the civilian segment of the GNSS remains susceptible to various spoofing attacks, necessitating robust detection mechanisms. The ability to deter such attacks significantly enhances the reliability and security of systems utilizing GNSS technology. Supervised Machine Learning (ML) techniques have shown promise in spoof detection. However, their effectiveness hinges on training data encompassing all possible attack scenarios, rendering them vulnerable to novel attack vectors. To address this limitation, we explore representation learning-based methods. These methods can be trained with a single data class and subsequently applied to classify test samples as either belonging to the training class or not. In this context, we introduce a GNSS spoof detection model comprising a Variational AutoEncoder (VAE) and a Generative Adversarial Network (GAN). The composite model is designed to efficiently learn the class distribution of the training data. The features used for training are extracted from the radio frequency and tracking modules of a standard GNSS receiver. To train our model, we leverage the Texas Spoofing Test Battery (TEXBAT) datasets. Our trained model yields three distinct detectors capable of effectively identifying spoofed signals. The detection performance across simpler to intermediate datasets for these detectors reaches approximately 99%, demonstrating their robustness. In the case of subtle attack scenario represented by DS-7, our approach achieves an approximate detection rate of 95%. In contrast, under supervised learning, the best detection score for DS-7 remains limited to 44.1%.
全球导航卫星系统(GNSS)通过提供精确的定时和定位数据,在关键基础设施中发挥着至关重要的作用。然而,全球导航卫星系统的民用部分仍然容易受到各种欺骗攻击,因此需要强有力的检测机制。遏制此类攻击的能力大大提高了利用 GNSS 技术的系统的可靠性和安全性。有监督的机器学习(ML)技术在欺骗检测方面已显示出前景。然而,这些技术的有效性取决于包含所有可能攻击场景的训练数据,因此容易受到新型攻击载体的攻击。为了解决这一局限性,我们探索了基于表示学习的方法。这些方法可以用单一数据类别进行训练,然后应用于将测试样本分类为属于或不属于训练类别。在此背景下,我们引入了一个由变异自动编码器(VAE)和生成对抗网络(GAN)组成的 GNSS 欺骗检测模型。该复合模型旨在有效学习训练数据的类别分布。用于训练的特征是从标准 GNSS 接收机的射频和跟踪模块中提取的。为了训练我们的模型,我们利用了德克萨斯欺骗测试电池(TEXBAT)数据集。我们训练的模型产生了三种不同的检测器,能够有效识别欺骗信号。这些检测器在简单到中级数据集上的检测性能达到约 99%,证明了它们的鲁棒性。在以 DS-7 为代表的微妙攻击场景中,我们的方法达到了近 95% 的检测率。相比之下,在监督学习下,DS-7 的最佳检测得分仍限制在 44.1%。
{"title":"A Deep Learning Based Induced GNSS Spoof Detection Framework","authors":"Asif Iqbal;Muhammad Naveed Aman;Biplab Sikdar","doi":"10.1109/TMLCN.2024.3386649","DOIUrl":"https://doi.org/10.1109/TMLCN.2024.3386649","url":null,"abstract":"The Global Navigation Satellite System (GNSS) plays a crucial role in critical infrastructure by delivering precise timing and positional data. Nonetheless, the civilian segment of the GNSS remains susceptible to various spoofing attacks, necessitating robust detection mechanisms. The ability to deter such attacks significantly enhances the reliability and security of systems utilizing GNSS technology. Supervised Machine Learning (ML) techniques have shown promise in spoof detection. However, their effectiveness hinges on training data encompassing all possible attack scenarios, rendering them vulnerable to novel attack vectors. To address this limitation, we explore representation learning-based methods. These methods can be trained with a single data class and subsequently applied to classify test samples as either belonging to the training class or not. In this context, we introduce a GNSS spoof detection model comprising a Variational AutoEncoder (VAE) and a Generative Adversarial Network (GAN). The composite model is designed to efficiently learn the class distribution of the training data. The features used for training are extracted from the radio frequency and tracking modules of a standard GNSS receiver. To train our model, we leverage the Texas Spoofing Test Battery (TEXBAT) datasets. Our trained model yields three distinct detectors capable of effectively identifying spoofed signals. The detection performance across simpler to intermediate datasets for these detectors reaches approximately 99%, demonstrating their robustness. In the case of subtle attack scenario represented by DS-7, our approach achieves an approximate detection rate of 95%. In contrast, under supervised learning, the best detection score for DS-7 remains limited to 44.1%.","PeriodicalId":100641,"journal":{"name":"IEEE Transactions on Machine Learning in Communications and Networking","volume":"2 ","pages":"457-478"},"PeriodicalIF":0.0,"publicationDate":"2024-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10495074","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140633565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Transactions on Machine Learning in Communications and Networking
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1