FAST:通过特征选择提升基于不确定性的神经网络测试优先级方法

Jialuo Chen, Jingyi Wang, Xiyue Zhang, Youcheng Sun, Marta Kwiatkowska, Jiming Chen, Peng Cheng
{"title":"FAST:通过特征选择提升基于不确定性的神经网络测试优先级方法","authors":"Jialuo Chen, Jingyi Wang, Xiyue Zhang, Youcheng Sun, Marta Kwiatkowska, Jiming Chen, Peng Cheng","doi":"arxiv-2409.09130","DOIUrl":null,"url":null,"abstract":"Due to the vast testing space, the increasing demand for effective and\nefficient testing of deep neural networks (DNNs) has led to the development of\nvarious DNN test case prioritization techniques. However, the fact that DNNs\ncan deliver high-confidence predictions for incorrectly predicted examples,\nknown as the over-confidence problem, causes these methods to fail to reveal\nhigh-confidence errors. To address this limitation, in this work, we propose\nFAST, a method that boosts existing prioritization methods through guided\nFeAture SelecTion. FAST is based on the insight that certain features may\nintroduce noise that affects the model's output confidence, thereby\ncontributing to high-confidence errors. It quantifies the importance of each\nfeature for the model's correct predictions, and then dynamically prunes the\ninformation from the noisy features during inference to derive a new\nprobability vector for the uncertainty estimation. With the help of FAST, the\nhigh-confidence errors and correctly classified examples become more\ndistinguishable, resulting in higher APFD (Average Percentage of Fault\nDetection) values for test prioritization, and higher generalization ability\nfor model enhancement. We conduct extensive experiments to evaluate FAST across\na diverse set of model structures on multiple benchmark datasets to validate\nthe effectiveness, efficiency, and scalability of FAST compared to the\nstate-of-the-art prioritization techniques.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FAST: Boosting Uncertainty-based Test Prioritization Methods for Neural Networks via Feature Selection\",\"authors\":\"Jialuo Chen, Jingyi Wang, Xiyue Zhang, Youcheng Sun, Marta Kwiatkowska, Jiming Chen, Peng Cheng\",\"doi\":\"arxiv-2409.09130\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Due to the vast testing space, the increasing demand for effective and\\nefficient testing of deep neural networks (DNNs) has led to the development of\\nvarious DNN test case prioritization techniques. However, the fact that DNNs\\ncan deliver high-confidence predictions for incorrectly predicted examples,\\nknown as the over-confidence problem, causes these methods to fail to reveal\\nhigh-confidence errors. To address this limitation, in this work, we propose\\nFAST, a method that boosts existing prioritization methods through guided\\nFeAture SelecTion. FAST is based on the insight that certain features may\\nintroduce noise that affects the model's output confidence, thereby\\ncontributing to high-confidence errors. It quantifies the importance of each\\nfeature for the model's correct predictions, and then dynamically prunes the\\ninformation from the noisy features during inference to derive a new\\nprobability vector for the uncertainty estimation. With the help of FAST, the\\nhigh-confidence errors and correctly classified examples become more\\ndistinguishable, resulting in higher APFD (Average Percentage of Fault\\nDetection) values for test prioritization, and higher generalization ability\\nfor model enhancement. We conduct extensive experiments to evaluate FAST across\\na diverse set of model structures on multiple benchmark datasets to validate\\nthe effectiveness, efficiency, and scalability of FAST compared to the\\nstate-of-the-art prioritization techniques.\",\"PeriodicalId\":501278,\"journal\":{\"name\":\"arXiv - CS - Software Engineering\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Software Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.09130\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09130","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

由于测试空间广阔,对深度神经网络(DNN)进行有效和高效测试的需求与日俱增,因此开发了各种 DNN 测试用例优先级排序技术。然而,由于 DNN 可以为预测错误的示例提供高置信度预测,即所谓的过置信度问题,导致这些方法无法揭示高置信度错误。为了解决这一局限性,我们在这项工作中提出了 FAST 方法,该方法通过引导图像选择来提升现有的优先级排序方法。FAST 基于这样一种认识:某些特征可能会带来噪声,影响模型输出的置信度,从而导致高置信度错误。它量化了每个特征对模型正确预测的重要性,然后在推理过程中动态修剪噪声特征信息,为不确定性估计导出新的概率向量。在 FAST 的帮助下,高置信度错误和正确分类的示例变得更加难以区分,从而为测试优先级的确定带来更高的 APFD(平均故障检测百分比)值,并为模型的增强带来更高的泛化能力。我们进行了广泛的实验,在多个基准数据集上对 FAST 的各种模型结构进行了评估,从而验证了 FAST 与最先进的优先级排序技术相比的有效性、效率和可扩展性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
FAST: Boosting Uncertainty-based Test Prioritization Methods for Neural Networks via Feature Selection
Due to the vast testing space, the increasing demand for effective and efficient testing of deep neural networks (DNNs) has led to the development of various DNN test case prioritization techniques. However, the fact that DNNs can deliver high-confidence predictions for incorrectly predicted examples, known as the over-confidence problem, causes these methods to fail to reveal high-confidence errors. To address this limitation, in this work, we propose FAST, a method that boosts existing prioritization methods through guided FeAture SelecTion. FAST is based on the insight that certain features may introduce noise that affects the model's output confidence, thereby contributing to high-confidence errors. It quantifies the importance of each feature for the model's correct predictions, and then dynamically prunes the information from the noisy features during inference to derive a new probability vector for the uncertainty estimation. With the help of FAST, the high-confidence errors and correctly classified examples become more distinguishable, resulting in higher APFD (Average Percentage of Fault Detection) values for test prioritization, and higher generalization ability for model enhancement. We conduct extensive experiments to evaluate FAST across a diverse set of model structures on multiple benchmark datasets to validate the effectiveness, efficiency, and scalability of FAST compared to the state-of-the-art prioritization techniques.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Promise and Peril of Collaborative Code Generation Models: Balancing Effectiveness and Memorization Shannon Entropy is better Feature than Category and Sentiment in User Feedback Processing Motivations, Challenges, Best Practices, and Benefits for Bots and Conversational Agents in Software Engineering: A Multivocal Literature Review A Taxonomy of Self-Admitted Technical Debt in Deep Learning Systems Investigating team maturity in an agile automotive reorganization
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1