Hongping Gan;Hejie Zheng;Zhangfa Wu;Chunyan Ma;Jie Liu
{"title":"TFD-Net: Transformer Deviation Network for Weakly Supervised Anomaly Detection","authors":"Hongping Gan;Hejie Zheng;Zhangfa Wu;Chunyan Ma;Jie Liu","doi":"10.1109/TNSM.2024.3485545","DOIUrl":null,"url":null,"abstract":"Deep Learning (DL)-based weakly supervised anomaly detection methods enhance the security and performance of communication and networks by promptly identifying and addressing anomalies within imbalanced samples, thus ensuring reliable communication and smooth network operations. However, existing DL-based methods often overly emphasize the local feature representations of samples, thereby neglecting the long-range dependencies and the prior knowledge of the samples, which imposes potential limitations on anomaly detection with a limited number of abnormal samples. To mitigate these challenges, we propose a Transformer deviation network for weakly supervised anomaly detection, called TFD-Net, which can effectively leverage the interdependencies and data priors of samples, yielding enhanced anomaly detection performance. Specifically, we first use a Transformer-based feature extraction module that proficiently captures the dependencies of global features in the samples. Subsequently, TFD-Net employs an anomaly score generation module to obtain corresponding anomaly scores. Finally, we introduce an innovative loss function for TFD-Net, named Transformer Deviation Loss Function (TFD-Loss), which can adequately incorporate prior knowledge of samples into the network training process, addressing the issue of imbalanced samples, and thereby enhancing the detection efficiency. Experimental results on public benchmark datasets demonstrate that TFD-Net substantially outperforms other DL-based methods in weakly supervised anomaly detection task.","PeriodicalId":13423,"journal":{"name":"IEEE Transactions on Network and Service Management","volume":"22 1","pages":"941-954"},"PeriodicalIF":5.4000,"publicationDate":"2024-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Network and Service Management","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10730761/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Deep Learning (DL)-based weakly supervised anomaly detection methods enhance the security and performance of communication and networks by promptly identifying and addressing anomalies within imbalanced samples, thus ensuring reliable communication and smooth network operations. However, existing DL-based methods often overly emphasize the local feature representations of samples, thereby neglecting the long-range dependencies and the prior knowledge of the samples, which imposes potential limitations on anomaly detection with a limited number of abnormal samples. To mitigate these challenges, we propose a Transformer deviation network for weakly supervised anomaly detection, called TFD-Net, which can effectively leverage the interdependencies and data priors of samples, yielding enhanced anomaly detection performance. Specifically, we first use a Transformer-based feature extraction module that proficiently captures the dependencies of global features in the samples. Subsequently, TFD-Net employs an anomaly score generation module to obtain corresponding anomaly scores. Finally, we introduce an innovative loss function for TFD-Net, named Transformer Deviation Loss Function (TFD-Loss), which can adequately incorporate prior knowledge of samples into the network training process, addressing the issue of imbalanced samples, and thereby enhancing the detection efficiency. Experimental results on public benchmark datasets demonstrate that TFD-Net substantially outperforms other DL-based methods in weakly supervised anomaly detection task.
基于深度学习的弱监督异常检测方法通过快速识别和处理不平衡样本中的异常,提高了通信和网络的安全性和性能,从而保证了通信的可靠和网络的平稳运行。然而,现有的基于dl的方法往往过分强调样本的局部特征表示,从而忽略了样本的长期依赖关系和先验知识,这对异常样本数量有限的异常检测造成了潜在的限制。为了缓解这些挑战,我们提出了一个用于弱监督异常检测的Transformer偏差网络,称为TFD-Net,它可以有效地利用样本的相互依赖性和数据先验性,从而提高异常检测性能。具体来说,我们首先使用基于transformer的特征提取模块,该模块熟练地捕获样本中全局特征的依赖关系。随后,TFD-Net利用异常评分生成模块获得相应的异常评分。最后,我们为TFD-Net引入了一种创新的损失函数,即变压器偏差损失函数(Transformer Deviation loss function, TFD-Loss),该函数可以将样本的先验知识充分纳入网络训练过程中,解决样本不平衡的问题,从而提高检测效率。在公共基准数据集上的实验结果表明,TFD-Net在弱监督异常检测任务中显著优于其他基于dl的方法。
期刊介绍:
IEEE Transactions on Network and Service Management will publish (online only) peerreviewed archival quality papers that advance the state-of-the-art and practical applications of network and service management. Theoretical research contributions (presenting new concepts and techniques) and applied contributions (reporting on experiences and experiments with actual systems) will be encouraged. These transactions will focus on the key technical issues related to: Management Models, Architectures and Frameworks; Service Provisioning, Reliability and Quality Assurance; Management Functions; Enabling Technologies; Information and Communication Models; Policies; Applications and Case Studies; Emerging Technologies and Standards.