Impacting Robustness in Deep Learning-Based NIDS through Poisoning Attacks

Algorithms Pub Date : 2024-04-11 DOI:10.3390/a17040155
Shahad Alahmed, Qutaiba Alasad, J. Yuan, Mohammed Alawad
{"title":"Impacting Robustness in Deep Learning-Based NIDS through Poisoning Attacks","authors":"Shahad Alahmed, Qutaiba Alasad, J. Yuan, Mohammed Alawad","doi":"10.3390/a17040155","DOIUrl":null,"url":null,"abstract":"The rapid expansion and pervasive reach of the internet in recent years have raised concerns about evolving and adaptable online threats, particularly with the extensive integration of Machine Learning (ML) systems into our daily routines. These systems are increasingly becoming targets of malicious attacks that seek to distort their functionality through the concept of poisoning. Such attacks aim to warp the intended operations of these services, deviating them from their true purpose. Poisoning renders systems susceptible to unauthorized access, enabling illicit users to masquerade as legitimate ones, compromising the integrity of smart technology-based systems like Network Intrusion Detection Systems (NIDSs). Therefore, it is necessary to continue working on studying the resilience of deep learning network systems while there are poisoning attacks, specifically interfering with the integrity of data conveyed over networks. This paper explores the resilience of deep learning (DL)—based NIDSs against untethered white-box attacks. More specifically, it introduces a designed poisoning attack technique geared especially for deep learning by adding various amounts of altered instances into training datasets at diverse rates and then investigating the attack’s influence on model performance. We observe that increasing injection rates (from 1% to 50%) and random amplified distribution have slightly affected the overall performance of the system, which is represented by accuracy (0.93) at the end of the experiments. However, the rest of the results related to the other measures, such as PPV (0.082), FPR (0.29), and MSE (0.67), indicate that the data manipulation poisoning attacks impact the deep learning model. These findings shed light on the vulnerability of DL-based NIDS under poisoning attacks, emphasizing the significance of securing such systems against these sophisticated threats, for which defense techniques should be considered. Our analysis, supported by experimental results, shows that the generated poisoned data have significantly impacted the model performance and are hard to be detected.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Algorithms","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/a17040155","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

The rapid expansion and pervasive reach of the internet in recent years have raised concerns about evolving and adaptable online threats, particularly with the extensive integration of Machine Learning (ML) systems into our daily routines. These systems are increasingly becoming targets of malicious attacks that seek to distort their functionality through the concept of poisoning. Such attacks aim to warp the intended operations of these services, deviating them from their true purpose. Poisoning renders systems susceptible to unauthorized access, enabling illicit users to masquerade as legitimate ones, compromising the integrity of smart technology-based systems like Network Intrusion Detection Systems (NIDSs). Therefore, it is necessary to continue working on studying the resilience of deep learning network systems while there are poisoning attacks, specifically interfering with the integrity of data conveyed over networks. This paper explores the resilience of deep learning (DL)—based NIDSs against untethered white-box attacks. More specifically, it introduces a designed poisoning attack technique geared especially for deep learning by adding various amounts of altered instances into training datasets at diverse rates and then investigating the attack’s influence on model performance. We observe that increasing injection rates (from 1% to 50%) and random amplified distribution have slightly affected the overall performance of the system, which is represented by accuracy (0.93) at the end of the experiments. However, the rest of the results related to the other measures, such as PPV (0.082), FPR (0.29), and MSE (0.67), indicate that the data manipulation poisoning attacks impact the deep learning model. These findings shed light on the vulnerability of DL-based NIDS under poisoning attacks, emphasizing the significance of securing such systems against these sophisticated threats, for which defense techniques should be considered. Our analysis, supported by experimental results, shows that the generated poisoned data have significantly impacted the model performance and are hard to be detected.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
通过中毒攻击影响基于深度学习的 NIDS 的鲁棒性
近年来,互联网的快速扩张和无处不在的覆盖范围引发了人们对不断发展和适应性强的在线威胁的担忧,特别是随着机器学习(ML)系统广泛融入我们的日常生活。这些系统正日益成为恶意攻击的目标,这些攻击试图通过 "中毒 "概念来扭曲系统的功能。此类攻击旨在扭曲这些服务的预期运行,使其偏离真正的目的。中毒会使系统容易受到未经授权的访问,使非法用户伪装成合法用户,损害网络入侵检测系统(NIDS)等基于智能技术的系统的完整性。因此,有必要继续研究深度学习网络系统在受到中毒攻击时的恢复能力,特别是干扰网络数据传输完整性的攻击。本文探讨了基于深度学习(DL)的 NIDS 对非绑定白盒攻击的恢复能力。更具体地说,本文介绍了一种专为深度学习设计的中毒攻击技术,方法是在训练数据集中以不同的速率添加各种数量的篡改实例,然后研究攻击对模型性能的影响。我们发现,增加注入率(从 1%到 50%)和随机放大分布略微影响了系统的整体性能,这在实验结束时的准确率(0.93)中有所体现。然而,与其他指标相关的其他结果,如 PPV(0.082)、FPR(0.29)和 MSE(0.67),表明数据操纵中毒攻击影响了深度学习模型。这些发现揭示了基于 DL 的 NIDS 在中毒攻击下的脆弱性,强调了确保此类系统免受这些复杂威胁的重要性,并应考虑采用防御技术。在实验结果的支持下,我们的分析表明,生成的中毒数据严重影响了模型的性能,而且很难被检测到。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Label-Setting Algorithm for Multi-Destination K Simple Shortest Paths Problem and Application A Quantum Approach for Exploring the Numerical Results of the Heat Equation Enhancing Indoor Positioning Accuracy with WLAN and WSN: A QPSO Hybrid Algorithm with Surface Tessellation Trajectory Classification and Recognition of Planar Mechanisms Based on ResNet18 Network Computational Test for Conditional Independence
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1