NNG-Mix: Improving Semi-Supervised Anomaly Detection With Pseudo-Anomaly Generation

IF 8.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE IEEE transactions on neural networks and learning systems Pub Date : 2024-11-20 DOI:10.1109/TNNLS.2024.3497801
Hao Dong;Gaëtan Frusque;Yue Zhao;Eleni Chatzi;Olga Fink
{"title":"NNG-Mix: Improving Semi-Supervised Anomaly Detection With Pseudo-Anomaly Generation","authors":"Hao Dong;Gaëtan Frusque;Yue Zhao;Eleni Chatzi;Olga Fink","doi":"10.1109/TNNLS.2024.3497801","DOIUrl":null,"url":null,"abstract":"Anomaly detection (AD) is essential in identifying rare and often critical events in complex systems, finding applications in fields such as network intrusion detection, financial fraud detection, and fault detection in infrastructure and industrial systems. While AD is typically treated as an unsupervised learning task due to the high cost of label annotation, it is more practical to assume access to a small set of labeled anomaly samples from domain experts, as is the case for semi-supervised AD. Semi-supervised and supervised approaches can leverage such labeled data, resulting in improved performance. In this article, rather than proposing a new semi-supervised or supervised approach for AD, we introduce a novel algorithm for generating additional pseudo-anomalies on the basis of the limited labeled anomalies and a large volume of unlabeled data. This serves as an augmentation to facilitate the detection of new anomalies. Our proposed algorithm, named nearest neighbor Gaussian mix-up (NNG-Mix), efficiently integrates information from both labeled and unlabeled data to generate pseudo-anomalies. We compare the performance of this novel algorithm with commonly applied augmentation techniques, such as Mixup and Cutout. We evaluate NNG-Mix by training various existing semi-supervised and supervised AD algorithms on the original training data along with the generated pseudo-anomalies. Through extensive experiments on 57 benchmark datasets in ADBench, reflecting different data types, we demonstrate that NNG-Mix outperforms other data augmentation methods. It yields significant performance improvements compared to the baselines trained exclusively on the original training data. Notably, NNG-Mix yields up to 16.4%, 8.8%, and 8.0% improvements on Classical, CV, and NLP datasets in ADBench. Our source code is available at <uri>https://github.com/donghao51/NNG-Mix</uri>.","PeriodicalId":13303,"journal":{"name":"IEEE transactions on neural networks and learning systems","volume":"36 6","pages":"10635-10647"},"PeriodicalIF":8.9000,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on neural networks and learning systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10759334/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Anomaly detection (AD) is essential in identifying rare and often critical events in complex systems, finding applications in fields such as network intrusion detection, financial fraud detection, and fault detection in infrastructure and industrial systems. While AD is typically treated as an unsupervised learning task due to the high cost of label annotation, it is more practical to assume access to a small set of labeled anomaly samples from domain experts, as is the case for semi-supervised AD. Semi-supervised and supervised approaches can leverage such labeled data, resulting in improved performance. In this article, rather than proposing a new semi-supervised or supervised approach for AD, we introduce a novel algorithm for generating additional pseudo-anomalies on the basis of the limited labeled anomalies and a large volume of unlabeled data. This serves as an augmentation to facilitate the detection of new anomalies. Our proposed algorithm, named nearest neighbor Gaussian mix-up (NNG-Mix), efficiently integrates information from both labeled and unlabeled data to generate pseudo-anomalies. We compare the performance of this novel algorithm with commonly applied augmentation techniques, such as Mixup and Cutout. We evaluate NNG-Mix by training various existing semi-supervised and supervised AD algorithms on the original training data along with the generated pseudo-anomalies. Through extensive experiments on 57 benchmark datasets in ADBench, reflecting different data types, we demonstrate that NNG-Mix outperforms other data augmentation methods. It yields significant performance improvements compared to the baselines trained exclusively on the original training data. Notably, NNG-Mix yields up to 16.4%, 8.8%, and 8.0% improvements on Classical, CV, and NLP datasets in ADBench. Our source code is available at https://github.com/donghao51/NNG-Mix.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
NNG-Mix:利用伪异常生成改进半监督异常检测
异常检测(AD)是识别复杂系统中罕见和关键事件的关键技术,在网络入侵检测、金融欺诈检测、基础设施和工业系统故障检测等领域都有广泛的应用。由于标签标注的成本很高,人工智能通常被视为一种无监督学习任务,但更实际的做法是假设从领域专家那里获得一小组标记异常样本,就像半监督人工智能一样。半监督和监督方法可以利用这样的标记数据,从而提高性能。在本文中,我们不是提出一种新的半监督或监督的AD方法,而是引入一种基于有限标记异常和大量未标记数据生成额外伪异常的新算法。这有助于发现新的异常情况。我们提出的算法,称为最近邻高斯混合(ngng - mix),有效地整合来自标记和未标记数据的信息来生成伪异常。我们将这种新算法与常用的增强技术(如Mixup和Cutout)的性能进行了比较。我们通过在原始训练数据上训练各种现有的半监督和监督AD算法以及生成的伪异常来评估ng - mix。通过ADBench中57个基准数据集的广泛实验,反映了不同的数据类型,我们证明了ng - mix优于其他数据增强方法。与仅在原始训练数据上训练的基线相比,它产生了显著的性能改进。值得注意的是,在ADBench中,与classic、CV和NLP数据集相比,lng - mix的产量分别提高了16.4%、8.8%和8.0%。我们的源代码可从https://github.com/donghao51/NNG-Mix获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE transactions on neural networks and learning systems
IEEE transactions on neural networks and learning systems COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-COMPUTER SCIENCE, HARDWARE & ARCHITECTURE
CiteScore
23.80
自引率
9.60%
发文量
2102
审稿时长
3-8 weeks
期刊介绍: The focus of IEEE Transactions on Neural Networks and Learning Systems is to present scholarly articles discussing the theory, design, and applications of neural networks as well as other learning systems. The journal primarily highlights technical and scientific research in this domain.
期刊最新文献
Seeing What Few-Shot Learners See: Contrastive Cross-Class Attribution for Explainability. Accelerated Reinforcement Learning With Verifiable Excitation for Cubic Convergence. Learning Optimal Policies With Local Observations for Cooperative Multiagent Reinforcement Learning. Learning From M-Tuple One-vs-All Confidence Comparison Data. Sparse Variational Student-t Processes for Heavy-Tailed Modeling.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1