{"title":"A Hybridized Approach for Testing Neural Network Based Intrusion Detection Systems","authors":"Faqeer ur Rehman, C. Izurieta","doi":"10.1109/SmartNets50376.2021.9555416","DOIUrl":null,"url":null,"abstract":"Enhancing the trust of machine learning-based classifiers with large input spaces is a desirable goal; however, due to high labeling costs and limited resources, this is a challenging task. One solution is to use test input prioritization techniques that aim to identify and label only the most effective test instances. These prioritized test inputs can then be used with some popular testing techniques e.g., Metamorphic testing (MT) to test and uncover implementation bugs in computationally complex machine learning classifiers that suffer from the oracle problem. However, there are certain limitations involved with this approach, (i) using a small number of prioritized test inputs may not be enough to check the program correctness over a large variety of input scenarios, and (ii) traditional MT approaches become infeasible when the programs under test exhibit a non-deterministic behavior during training e.g., Neural Network-based classifiers. Therefore, instead of using MT for testing purposes, we propose a metamorphic relation to solve a data generation/labeling problem; that is, enhancing the test inputs effectiveness by extending the prioritized test set with new tests without incurring additional labeling costs. Further, we leverage the prioritized test inputs (both source and follow-up data sets) and propose a statistical hypothesis testing (for detection) and machine learning-based approach (for prediction) of faulty behavior in two other machine learning classifiers (Neural Network-based Intrusion Detection Systems). In our case, the problem is interesting in the sense that injected bugs represent the high accuracy producing mutated program versions that may be difficult to detect by a software developer. The results indicate that (i) the proposed statistical hypothesis testing is able to identify the induced buggy behavior, and (ii) Random Forest outperforms and achieves the best performance over SVM and k-NN algorithms.","PeriodicalId":443191,"journal":{"name":"2021 International Conference on Smart Applications, Communications and Networking (SmartNets)","volume":"123 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Smart Applications, Communications and Networking (SmartNets)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SmartNets50376.2021.9555416","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4
Abstract
Enhancing the trust of machine learning-based classifiers with large input spaces is a desirable goal; however, due to high labeling costs and limited resources, this is a challenging task. One solution is to use test input prioritization techniques that aim to identify and label only the most effective test instances. These prioritized test inputs can then be used with some popular testing techniques e.g., Metamorphic testing (MT) to test and uncover implementation bugs in computationally complex machine learning classifiers that suffer from the oracle problem. However, there are certain limitations involved with this approach, (i) using a small number of prioritized test inputs may not be enough to check the program correctness over a large variety of input scenarios, and (ii) traditional MT approaches become infeasible when the programs under test exhibit a non-deterministic behavior during training e.g., Neural Network-based classifiers. Therefore, instead of using MT for testing purposes, we propose a metamorphic relation to solve a data generation/labeling problem; that is, enhancing the test inputs effectiveness by extending the prioritized test set with new tests without incurring additional labeling costs. Further, we leverage the prioritized test inputs (both source and follow-up data sets) and propose a statistical hypothesis testing (for detection) and machine learning-based approach (for prediction) of faulty behavior in two other machine learning classifiers (Neural Network-based Intrusion Detection Systems). In our case, the problem is interesting in the sense that injected bugs represent the high accuracy producing mutated program versions that may be difficult to detect by a software developer. The results indicate that (i) the proposed statistical hypothesis testing is able to identify the induced buggy behavior, and (ii) Random Forest outperforms and achieves the best performance over SVM and k-NN algorithms.
增强具有大输入空间的基于机器学习的分类器的信任是一个理想的目标;然而,由于高标签成本和有限的资源,这是一项具有挑战性的任务。一种解决方案是使用测试输入优先级技术,旨在识别和标记最有效的测试实例。然后,这些优先级的测试输入可以与一些流行的测试技术一起使用,例如,变形测试(MT),以测试和发现计算复杂的机器学习分类器中遭受oracle问题的实现错误。然而,这种方法有一定的局限性,(i)使用少量的优先测试输入可能不足以检查程序在各种输入场景下的正确性,(ii)当被测试程序在训练期间表现出不确定性行为时,传统的机器翻译方法变得不可行的,例如基于神经网络的分类器。因此,我们提出了一种变质关系来解决数据生成/标记问题,而不是使用MT进行测试;也就是说,通过使用新测试扩展优先测试集来增强测试输入的有效性,而不会产生额外的标记成本。此外,我们利用了优先测试输入(源数据集和后续数据集),并在另外两个机器学习分类器(基于神经网络的入侵检测系统)中提出了错误行为的统计假设检验(用于检测)和基于机器学习的方法(用于预测)。在我们的例子中,这个问题很有趣,因为注入的错误代表了产生突变程序版本的高精度,这可能很难被软件开发人员检测到。结果表明:(i)所提出的统计假设检验能够识别诱导的bug行为,(ii) Random Forest优于SVM和k-NN算法,并取得了最佳性能。