首页 > 最新文献

2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)最新文献

英文 中文
Cache Optimizations for Test Case Reduction 减少测试用例的缓存优化
Dániel Vince, Ákos Kiss
Finding the relevant part of failure-inducing inputs is an important first step on the path of debugging. If much of a test case that triggers a bug does not contribute to the actual failure, then the time required to fix the bug can increase considerably. In this paper, we focus on the memory requirements of automatic test case reduction. During minimization, the same test case might be tested multiple times, and determining the outcome of an input may take time, therefore, different caching solutions were proposed to avoid re-testing previously seen inputs. We investigated the caching solutions of DDMIN and HDD, and found that their scaling is suboptimal. We propose three optimizations for one of the state-of-the-art caching solutions: with the optimizations combined, DDMIN requires 96% and HDD requires 85% less memory compared to the baseline implementation. Furthermore, as a side effect, the reduction becomes faster by 9.9% with DDMIN.
找出诱发故障的输入的相关部分是调试过程中重要的第一步。如果触发错误的大部分测试用例不会导致实际的失败,那么修复错误所需的时间就会大大增加。在本文中,我们主要关注自动测试用例缩减的内存需求。在最小化期间,相同的测试用例可能被测试多次,并且确定输入的结果可能需要时间,因此,提出了不同的缓存解决方案,以避免重新测试以前看到的输入。我们研究了DDMIN和HDD的缓存解决方案,发现它们的可伸缩性不是最优的。我们为最先进的缓存解决方案之一提出了三个优化:与基线实现相比,DDMIN需要96%的内存,HDD需要85%的内存。此外,作为副作用,DDMIN的降低速度更快,达到9.9%。
{"title":"Cache Optimizations for Test Case Reduction","authors":"Dániel Vince, Ákos Kiss","doi":"10.1109/QRS57517.2022.00052","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00052","url":null,"abstract":"Finding the relevant part of failure-inducing inputs is an important first step on the path of debugging. If much of a test case that triggers a bug does not contribute to the actual failure, then the time required to fix the bug can increase considerably. In this paper, we focus on the memory requirements of automatic test case reduction. During minimization, the same test case might be tested multiple times, and determining the outcome of an input may take time, therefore, different caching solutions were proposed to avoid re-testing previously seen inputs. We investigated the caching solutions of DDMIN and HDD, and found that their scaling is suboptimal. We propose three optimizations for one of the state-of-the-art caching solutions: with the optimizations combined, DDMIN requires 96% and HDD requires 85% less memory compared to the baseline implementation. Furthermore, as a side effect, the reduction becomes faster by 9.9% with DDMIN.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122381555","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Formal Verification of Hierarchical Ptolemy II Synchronous-Reactive Models with Bounded Model Checking 具有有界模型检验的分层Ptolemy II同步反应模型的形式化验证
Xiaozhen Zhang, Zhaoming Yang, Hui Kong, W. Kong
Ptolemy II is an open-source modeling and simulation tool for concurrent, real-time and embedded systems, particularly those involving hierarchical heterogeneity. Synchronous- reactive (SR) model of computation which has been implemented in Ptolemy II is commonly used to design safety-critical systems with complicated control logic. Formally verifying the correctness of hierarchical SR models is of great importance and also challenging due to the formalization of a series of specific features including, e.g., instantaneous communication between actors across the level of hierarchy, the combination of SR’s fixed-point semantic with hierarchical structure, and multiple clocks proceeding at different rates in multiclock SR models. In this paper, we tackle such challenges and propose a bounded model checking (BMC) approach to typical actors commonly used in hierarchical SR models. In addition, we implement the proposed BMC approach to hierarchical SR models in a prototype tool called Ptolemy-Z3, which has been integrated into the Ptolemy II environment. Experimental results show that Ptolemy-Z3 outperforms significantly Ptolemy-NuSMV (a verification tool provided by the Ptolemy II environment) in the verification capability of hierarchical SR models.
托勒密II是一个开源的建模和仿真工具,用于并发、实时和嵌入式系统,特别是那些涉及分层异构的系统。在托勒密II中实现的同步-反应(SR)计算模型是设计具有复杂控制逻辑的安全关键系统的常用方法。正式验证分层SR模型的正确性是非常重要的,也是具有挑战性的,因为一系列具体特征的形式化,包括,例如,跨层次行为者之间的瞬时通信,SR的定点语义与分层结构的结合,以及多锁SR模型中以不同速率进行的多个时钟。在本文中,我们解决了这些挑战,并提出了一种有界模型检查(BMC)方法,用于分层SR模型中常用的典型参与者。此外,我们在一个名为托勒密- z3的原型工具中实现了提出的BMC分层SR模型方法,该工具已集成到托勒密II环境中。实验结果表明,托勒密- z3在层次SR模型的验证能力上明显优于托勒密- nusmv(托勒密II环境提供的验证工具)。
{"title":"Formal Verification of Hierarchical Ptolemy II Synchronous-Reactive Models with Bounded Model Checking","authors":"Xiaozhen Zhang, Zhaoming Yang, Hui Kong, W. Kong","doi":"10.1109/QRS57517.2022.00049","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00049","url":null,"abstract":"Ptolemy II is an open-source modeling and simulation tool for concurrent, real-time and embedded systems, particularly those involving hierarchical heterogeneity. Synchronous- reactive (SR) model of computation which has been implemented in Ptolemy II is commonly used to design safety-critical systems with complicated control logic. Formally verifying the correctness of hierarchical SR models is of great importance and also challenging due to the formalization of a series of specific features including, e.g., instantaneous communication between actors across the level of hierarchy, the combination of SR’s fixed-point semantic with hierarchical structure, and multiple clocks proceeding at different rates in multiclock SR models. In this paper, we tackle such challenges and propose a bounded model checking (BMC) approach to typical actors commonly used in hierarchical SR models. In addition, we implement the proposed BMC approach to hierarchical SR models in a prototype tool called Ptolemy-Z3, which has been integrated into the Ptolemy II environment. Experimental results show that Ptolemy-Z3 outperforms significantly Ptolemy-NuSMV (a verification tool provided by the Ptolemy II environment) in the verification capability of hierarchical SR models.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128814697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Crowdsourced Testing Task Assignment based on Knowledge Graphs 基于知识图谱的众包测试任务分配
Peng-Xi Yang, Chao Chang, Yong Tang
The non-professional and uncertain testers in crowdsourced testing could lead to the problems of uneven test report quality, substandard test requirement coverage, a large number of repeated bug reports, and low efficiency of report reviewing. This paper designs a crowdsourced testing task assignment approach based on knowledge graph, trying to make full use of the individual advantages and crowd intelligence of crowdsourced workers in crowdsourced testing through personalized task assignment, with the goal to improve the quality of test reports and test completion efficiency. The approach includes three modules: 1) knowledge graph data acquisition: the concept of collaborative crowdsourced test is introduced, and a complete crowdsourced report submission platform is built to obtain the required data for the knowledge graph. 2) Knowledge graph feature learning: building an internal knowledge graph of the crowdsourced testing field based on the data in the platform and combining the historical task records of crowdsourced workers as input, using the machine learning model to get the crowdsourced workers’ preference for specific tasks, and integrates the three-level page coverage and bug-like status. 3) Knowledge graph task assignment: assign test tasks and audit tasks to crowdsourced workers in order to improve the coverage of test requirements and overall test efficiency. We compare the quantity and quality of bug reports in a crowdsourced test task between the task assignment system based on a knowledge graph and the system based on collaborative filtering, which proves the effectiveness of our task assignment technique.
众包测试中测试人员的非专业和不确定性会导致测试报告质量参差不齐、测试需求覆盖率不达标、大量重复bug报告、报告评审效率低等问题。本文设计了一种基于知识图的众包测试任务分配方法,试图通过个性化的任务分配,充分利用众包工作者在众包测试中的个体优势和群体智能,以提高测试报告的质量和测试完成效率。该方法包括三个模块:1)知识图谱数据获取:引入协同众包测试的概念,构建完整的众包报告提交平台,获取知识图谱所需数据。2)知识图谱特征学习:基于平台内数据,结合众包工作者的历史任务记录作为输入,构建众包测试场内部知识图谱,利用机器学习模型获取众包工作者对特定任务的偏好,并整合三级页面覆盖率和bug样状态。3)知识图任务分配:将测试任务和审核任务分配给众包工作者,以提高测试需求的覆盖率和整体测试效率。比较了基于知识图的任务分配系统和基于协同过滤的任务分配系统在众包测试任务中bug报告的数量和质量,验证了任务分配技术的有效性。
{"title":"Crowdsourced Testing Task Assignment based on Knowledge Graphs","authors":"Peng-Xi Yang, Chao Chang, Yong Tang","doi":"10.1109/QRS57517.2022.00072","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00072","url":null,"abstract":"The non-professional and uncertain testers in crowdsourced testing could lead to the problems of uneven test report quality, substandard test requirement coverage, a large number of repeated bug reports, and low efficiency of report reviewing. This paper designs a crowdsourced testing task assignment approach based on knowledge graph, trying to make full use of the individual advantages and crowd intelligence of crowdsourced workers in crowdsourced testing through personalized task assignment, with the goal to improve the quality of test reports and test completion efficiency. The approach includes three modules: 1) knowledge graph data acquisition: the concept of collaborative crowdsourced test is introduced, and a complete crowdsourced report submission platform is built to obtain the required data for the knowledge graph. 2) Knowledge graph feature learning: building an internal knowledge graph of the crowdsourced testing field based on the data in the platform and combining the historical task records of crowdsourced workers as input, using the machine learning model to get the crowdsourced workers’ preference for specific tasks, and integrates the three-level page coverage and bug-like status. 3) Knowledge graph task assignment: assign test tasks and audit tasks to crowdsourced workers in order to improve the coverage of test requirements and overall test efficiency. We compare the quantity and quality of bug reports in a crowdsourced test task between the task assignment system based on a knowledge graph and the system based on collaborative filtering, which proves the effectiveness of our task assignment technique.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116886264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Empirical Study on Source Code Feature Extraction in Preprocessing of IR-Based Requirements Traceability 基于ir的需求追溯预处理中源代码特征提取的实证研究
Bangchao Wang, Yang Deng, Ruiqi Luo, Huan Jin
In information retrieval-based (IR-based) requirements traceability research, a great deal of researches have focused on establishing trace links between requirements and source code. However, as the description styles of source code and requirements are very different, how to better preprocess the code is crucial for the quality of trace link generation. This paper aims to draw empirical conclusions about code feature extraction, annotation importance assessment, and annotation redundancy removal through comprehensive experiments, which impact the quality of trace links generated by IR-based methods between requirements and source code. The results show that when the average annotaion density is higher than 0.2, feature extraction is recommended. Removing redundancy from code with high annotation redundancy can enhance the quality of trace links. The above experiences can help developers to improve the quality of trace link generation and provide them with advice on writing code.
在基于信息检索(ir)的需求可追溯性研究中,大量的研究集中在建立需求和源代码之间的跟踪链接。然而,由于源代码和需求的描述风格有很大的不同,如何更好地对代码进行预处理对跟踪链接生成的质量至关重要。本文旨在通过综合实验得出影响需求与源代码之间基于ir方法生成的跟踪链接质量的代码特征提取、标注重要性评估和标注冗余去除的经验结论。结果表明,当平均标注密度大于0.2时,建议进行特征提取。从注释冗余度高的代码中去除冗余可以提高跟踪链接的质量。以上经验可以帮助开发人员提高跟踪链接生成的质量,并为他们编写代码提供建议。
{"title":"An Empirical Study on Source Code Feature Extraction in Preprocessing of IR-Based Requirements Traceability","authors":"Bangchao Wang, Yang Deng, Ruiqi Luo, Huan Jin","doi":"10.1109/QRS57517.2022.00110","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00110","url":null,"abstract":"In information retrieval-based (IR-based) requirements traceability research, a great deal of researches have focused on establishing trace links between requirements and source code. However, as the description styles of source code and requirements are very different, how to better preprocess the code is crucial for the quality of trace link generation. This paper aims to draw empirical conclusions about code feature extraction, annotation importance assessment, and annotation redundancy removal through comprehensive experiments, which impact the quality of trace links generated by IR-based methods between requirements and source code. The results show that when the average annotaion density is higher than 0.2, feature extraction is recommended. Removing redundancy from code with high annotation redundancy can enhance the quality of trace links. The above experiences can help developers to improve the quality of trace link generation and provide them with advice on writing code.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114523167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Contextual Operationalization of Metrics as Scores: Is My Metric Value Good? 作为分数的度量的上下文操作化:我的度量值好吗?
Sebastian Hönel, Morgan Ericsson, Welf Löwe, Anna Wingkvist
Software quality models aggregate metrics to indicate quality. Most metrics reflect counts derived from events or attributes that cannot directly be associated with quality. Worse, what constitutes a desirable value for a metric may vary across contexts. We demonstrate an approach to transforming arbitrary metrics into absolute quality scores by leveraging metrics captured from similar contexts. In contrast to metrics, scores represent freestanding quality properties that are also comparable. We provide a web-based tool for obtaining contextualized scores for metrics as obtained from one’s software. Our results indicate that significant differences among various metrics and contexts exist. The suggested approach works with arbitrary contexts. Given sufficient contextual information, it allows for answering the question of whether a metric value is good/bad or common/extreme.
软件质量模型集合度量来指示质量。大多数度量标准反映的计数来源于不能直接与质量相关联的事件或属性。更糟糕的是,在不同的环境中,度量标准的理想值可能会有所不同。我们演示了一种方法,通过利用从类似环境中捕获的度量将任意度量转换为绝对质量分数。与指标相比,分数代表了独立的质量属性,这些属性也具有可比性。我们提供了一个基于网络的工具,用于从一个人的软件中获得指标的情境化分数。我们的研究结果表明,在不同的度量标准和上下文之间存在显著差异。建议的方法适用于任意上下文。给定足够的上下文信息,它允许回答度量值是好/坏或普通/极端的问题。
{"title":"Contextual Operationalization of Metrics as Scores: Is My Metric Value Good?","authors":"Sebastian Hönel, Morgan Ericsson, Welf Löwe, Anna Wingkvist","doi":"10.1109/QRS57517.2022.00042","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00042","url":null,"abstract":"Software quality models aggregate metrics to indicate quality. Most metrics reflect counts derived from events or attributes that cannot directly be associated with quality. Worse, what constitutes a desirable value for a metric may vary across contexts. We demonstrate an approach to transforming arbitrary metrics into absolute quality scores by leveraging metrics captured from similar contexts. In contrast to metrics, scores represent freestanding quality properties that are also comparable. We provide a web-based tool for obtaining contextualized scores for metrics as obtained from one’s software. Our results indicate that significant differences among various metrics and contexts exist. The suggested approach works with arbitrary contexts. Given sufficient contextual information, it allows for answering the question of whether a metric value is good/bad or common/extreme.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125709091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automated Synthesis of Quantum Circuits using Neural Network 基于神经网络的量子电路自动合成
Kentaro Murakami, Jianjun Zhao
While the ability to build quantum computers is improving dramatically, developing quantum algorithms is very limited and relies on human insight and ingenuity. Although several quantum programming languages have been developed, it is challenging for software developers unfamiliar with quantum computing to learn and use these languages. It is, therefore, necessary to develop tools to support developing new quantum algorithms and programs automatically. This paper proposes AutoQC, an approach to automatically synthesizing quantum circuits using the neural network from input and output pairs. We consider a quantum circuit a sequence of quantum gates and synthesize a quantum circuit probabilistically by prioritizing through a neural network at each step. The experimental results highlight the ability of AutoQC to synthesize some essential quantum circuits at a lower cost.
虽然建造量子计算机的能力正在显著提高,但开发量子算法非常有限,并且依赖于人类的洞察力和聪明才智。虽然已经开发了几种量子编程语言,但对于不熟悉量子计算的软件开发人员来说,学习和使用这些语言是具有挑战性的。因此,有必要开发工具来支持自动开发新的量子算法和程序。本文提出了一种利用神经网络从输入和输出对自动合成量子电路的方法——AutoQC。我们认为量子电路是一个量子门序列,并通过神经网络在每一步进行优先级排序,以概率方式合成量子电路。实验结果突出了AutoQC以较低成本合成一些基本量子电路的能力。
{"title":"Automated Synthesis of Quantum Circuits using Neural Network","authors":"Kentaro Murakami, Jianjun Zhao","doi":"10.1109/QRS57517.2022.00075","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00075","url":null,"abstract":"While the ability to build quantum computers is improving dramatically, developing quantum algorithms is very limited and relies on human insight and ingenuity. Although several quantum programming languages have been developed, it is challenging for software developers unfamiliar with quantum computing to learn and use these languages. It is, therefore, necessary to develop tools to support developing new quantum algorithms and programs automatically. This paper proposes AutoQC, an approach to automatically synthesizing quantum circuits using the neural network from input and output pairs. We consider a quantum circuit a sequence of quantum gates and synthesize a quantum circuit probabilistically by prioritizing through a neural network at each step. The experimental results highlight the ability of AutoQC to synthesize some essential quantum circuits at a lower cost.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130655415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ParGCN: Abnormal Transaction Detection based on Graph Neural Networks 基于图神经网络的异常事务检测
Lian Yu, Qi Jing, Ruomiao Li, Zhiya Cheng, Chang Xu
This paper improves GraphSAGE from two aspects: 1) performing a sampling compensation before the training to avoid the possible information losses due to the sampling; and 2) adding a hopping connection with the initial inputs in the aggregating phase to avert the potential loss of the initial features of nodes. The empirical study shows that FastGCN can obtain a relatively higher recall of detection but with a lower precision due to its randomness of Monte-Carlo methods and ignoring the special impacts of neighbors; while the improved GraphSAGE gets a relatively higher precision of detection but with a lower recall due to only focusing on neighbors. This paper proposes a graph-based approach to improve both precision and recall of the abnormal transaction detection by hybridizing the improved GraphSAGE with FastGCN, called ParGCN (Precision and recall), describes the mathematical formulas of the hybrid model, and analyzes the time complexity. A set of experiments on the two data-sets with significant differences of the numbers of features are performed to compare and evaluate the proposed approach to demonstrate the validity in terms of the precision and recall.
本文从两个方面对GraphSAGE进行改进:1)在训练前进行采样补偿,避免采样可能造成的信息损失;2)在聚合阶段的初始输入中加入跳跃连接,避免节点初始特征的潜在损失。实证研究表明,FastGCN由于蒙特卡罗方法的随机性和忽略了邻居的特殊影响,可以获得较高的检测召回率,但精度较低;而改进的GraphSAGE检测精度相对较高,但由于只关注邻居,召回率较低。本文提出了一种基于图的方法,通过将改进的GraphSAGE与FastGCN (precision and recall)相结合,提高异常事务检测的准确率和召回率,称为ParGCN (precision and recall),描述了混合模型的数学公式,并分析了时间复杂度。在特征数量存在显著差异的两组数据集上进行了一组实验,比较和评估了所提出的方法在查准率和查全率方面的有效性。
{"title":"ParGCN: Abnormal Transaction Detection based on Graph Neural Networks","authors":"Lian Yu, Qi Jing, Ruomiao Li, Zhiya Cheng, Chang Xu","doi":"10.1109/QRS57517.2022.00085","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00085","url":null,"abstract":"This paper improves GraphSAGE from two aspects: 1) performing a sampling compensation before the training to avoid the possible information losses due to the sampling; and 2) adding a hopping connection with the initial inputs in the aggregating phase to avert the potential loss of the initial features of nodes. The empirical study shows that FastGCN can obtain a relatively higher recall of detection but with a lower precision due to its randomness of Monte-Carlo methods and ignoring the special impacts of neighbors; while the improved GraphSAGE gets a relatively higher precision of detection but with a lower recall due to only focusing on neighbors. This paper proposes a graph-based approach to improve both precision and recall of the abnormal transaction detection by hybridizing the improved GraphSAGE with FastGCN, called ParGCN (Precision and recall), describes the mathematical formulas of the hybrid model, and analyzes the time complexity. A set of experiments on the two data-sets with significant differences of the numbers of features are performed to compare and evaluate the proposed approach to demonstrate the validity in terms of the precision and recall.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132315209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
IntJect: Vulnerability Intent Bug Seeding IntJect:漏洞意图Bug播种
Benjamin Petit, Ahmed Khanfir, E. Soremekun, Gilles Perrouin, Mike Papadakis
Studying and exposing software vulnerabilities is important to ensure software security, safety, and reliability. Software engineers often inject vulnerabilities into their programs to test the reliability of their test suites, vulnerability detectors, and security measures. However, state-of-the-art vulnerability injection methods only capture code syntax/patterns, they do not learn the intent of the vulnerability and are limited to the syntax of the original dataset. To address this challenge, we propose the first intent-based vulnerability injection method that learns both the program syntax and vulnerability intent. Our approach applies a combination of NLP methods and semantic-preserving program mutations (at the bytecode level) to inject code vulnerabilities. Given a dataset of known vulnerabilities (containing benign and vulnerable code pairs), our approach proceeds by employing semantic-preserving program mutations to transform the existing dataset to semantically similar code. Then, it learns the intent of the vulnerability via neural machine translation (Seq2Seq) models. The key insight is to employ Seq2Seq to learn the intent (context) of the vulnerable code in a manner that is agnostic of the specific program instance. We evaluate the performance of our approach using 1275 vulnerabilities belonging to five (5) CWEs from the Juliet test suite. We examine the effectiveness of our approach in producing compilable and vulnerable code. Our results show that IntJECT is effective, almost all (99%) of the code produced by our approach is vulnerable and compilable. We also demonstrate that the vulnerable programs generated by IntJECT are semantically similar to the withheld original vulnerable code. Finally, we show that our mutation-based data transformation approach outperforms its alternatives, namely data obfuscation and using the original data.
研究和暴露软件漏洞对于确保软件的安全性、安全性和可靠性非常重要。软件工程师经常将漏洞注入到他们的程序中,以测试他们的测试套件、漏洞检测器和安全措施的可靠性。然而,最先进的漏洞注入方法只捕获代码语法/模式,它们不了解漏洞的意图,并且仅限于原始数据集的语法。为了解决这一挑战,我们提出了第一种基于意图的漏洞注入方法,该方法可以同时学习程序语法和漏洞意图。我们的方法结合了NLP方法和保持语义的程序突变(在字节码级别)来注入代码漏洞。给定已知漏洞的数据集(包含良性和易受攻击的代码对),我们的方法通过使用语义保留程序突变将现有数据集转换为语义相似的代码来进行。然后,它通过神经机器翻译(Seq2Seq)模型学习漏洞的意图。关键的洞察力是使用Seq2Seq以一种与特定程序实例无关的方式来了解易受攻击代码的意图(上下文)。我们使用朱丽叶测试套件中的五(5)个CWEs中的1275个漏洞来评估我们方法的性能。我们检查了我们的方法在生成可编译和易受攻击的代码方面的有效性。我们的结果表明,IntJECT是有效的,几乎所有(99%)由我们的方法产生的代码是脆弱的和可编译的。我们还证明了由IntJECT生成的易受攻击的程序在语义上与保留的原始易受攻击代码相似。最后,我们证明了基于突变的数据转换方法优于其替代方法,即数据混淆和使用原始数据。
{"title":"IntJect: Vulnerability Intent Bug Seeding","authors":"Benjamin Petit, Ahmed Khanfir, E. Soremekun, Gilles Perrouin, Mike Papadakis","doi":"10.1109/QRS57517.2022.00013","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00013","url":null,"abstract":"Studying and exposing software vulnerabilities is important to ensure software security, safety, and reliability. Software engineers often inject vulnerabilities into their programs to test the reliability of their test suites, vulnerability detectors, and security measures. However, state-of-the-art vulnerability injection methods only capture code syntax/patterns, they do not learn the intent of the vulnerability and are limited to the syntax of the original dataset. To address this challenge, we propose the first intent-based vulnerability injection method that learns both the program syntax and vulnerability intent. Our approach applies a combination of NLP methods and semantic-preserving program mutations (at the bytecode level) to inject code vulnerabilities. Given a dataset of known vulnerabilities (containing benign and vulnerable code pairs), our approach proceeds by employing semantic-preserving program mutations to transform the existing dataset to semantically similar code. Then, it learns the intent of the vulnerability via neural machine translation (Seq2Seq) models. The key insight is to employ Seq2Seq to learn the intent (context) of the vulnerable code in a manner that is agnostic of the specific program instance. We evaluate the performance of our approach using 1275 vulnerabilities belonging to five (5) CWEs from the Juliet test suite. We examine the effectiveness of our approach in producing compilable and vulnerable code. Our results show that IntJECT is effective, almost all (99%) of the code produced by our approach is vulnerable and compilable. We also demonstrate that the vulnerable programs generated by IntJECT are semantically similar to the withheld original vulnerable code. Finally, we show that our mutation-based data transformation approach outperforms its alternatives, namely data obfuscation and using the original data.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133014206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fine-Tuning Pre-Trained Model to Extract Undesired Behaviors from App Reviews 微调预训练模型,从应用评论中提取不良行为
Wenyu Zhang, Xiaojuan Wang, Shanyan Lai, Chunyang Ye, Hui Zhou
Mobile application markets usually enact policies to describe in detail the minimum requirements that an application should comply with. User comments on mobile applications contain a large amount of information that can be used to find out APP's violations of market policies in a cost-effective way. Existing state-of-the-art methods match user comments with the violations of market policies based on well-designed syntax rules, which however cannot well capture the semantics of user comments and cannot be generalized to the scenarios not covered by the rules. To address this issue, we propose an innovative method, UBC-BERT, to detect undesired behavior from user comments based on their semantics. By incorporating sentence embeddings with attention, we train a classification model for 21 groups of undesirable behaviors based on the fine-tuning of a pre-trained model BERT-BASE. The experimental results show that our solution outperforms the baseline solutions in terms of a higher precision(up to 60.5% more).
移动应用市场通常会制定政策,详细描述应用程序应该遵守的最低要求。用户对移动应用的评论包含了大量的信息,这些信息可以用来以一种经济有效的方式发现APP违反市场政策的行为。现有的最先进的方法基于精心设计的语法规则将用户评论与违反市场政策的行为匹配起来,然而,这些方法不能很好地捕获用户评论的语义,也不能推广到规则未涵盖的场景。为了解决这个问题,我们提出了一种创新的方法,UBC-BERT,根据用户评论的语义来检测用户评论中的不良行为。通过将句子嵌入与注意力相结合,我们在对预训练模型BERT-BASE进行微调的基础上,训练了21组不良行为的分类模型。实验结果表明,我们的解决方案在更高的精度方面优于基线解决方案(高达60.5%以上)。
{"title":"Fine-Tuning Pre-Trained Model to Extract Undesired Behaviors from App Reviews","authors":"Wenyu Zhang, Xiaojuan Wang, Shanyan Lai, Chunyang Ye, Hui Zhou","doi":"10.1109/QRS57517.2022.00115","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00115","url":null,"abstract":"Mobile application markets usually enact policies to describe in detail the minimum requirements that an application should comply with. User comments on mobile applications contain a large amount of information that can be used to find out APP's violations of market policies in a cost-effective way. Existing state-of-the-art methods match user comments with the violations of market policies based on well-designed syntax rules, which however cannot well capture the semantics of user comments and cannot be generalized to the scenarios not covered by the rules. To address this issue, we propose an innovative method, UBC-BERT, to detect undesired behavior from user comments based on their semantics. By incorporating sentence embeddings with attention, we train a classification model for 21 groups of undesirable behaviors based on the fine-tuning of a pre-trained model BERT-BASE. The experimental results show that our solution outperforms the baseline solutions in terms of a higher precision(up to 60.5% more).","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114339266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TokenAuditor: Detecting Manipulation Risk in Token Smart Contract by Fuzzing TokenAuditor:通过模糊测试检测Token智能合约中的操纵风险
Mingpei Cao, Yueze Zhang, Zhenxuan Feng, Jiahao Hu, Yuesheng Zhu
Decentralized cryptocurrencies are influential smart contract applications in the blockchain, drawing interest from industry and academia. The capacity to govern and manage token behavior provided by the token smart contract adds to thriving decentralized applications. However, token smart contracts face security challenges in technology weakness and manipulation risks. In this work, we briefly describe the manipulation risk and propose TokenAuditor, a fuzzing framework detecting those risks in token smart contracts. TokenAuditor constructs basic blocks based on the contract bytecodes and adopts the rarity selection and mutation strategy to generate test cases. The main idea is to select the test cases that have hit rare basic blocks since the fuzzing started as candidates and perform mutation operations on them. In our evaluation, TokenAudiotr discovered 664 manipulation risks of four types in 4021 real-world token contracts.
去中心化加密货币是区块链中有影响力的智能合约应用,引起了工业界和学术界的兴趣。令牌智能合约提供的治理和管理令牌行为的能力增加了蓬勃发展的去中心化应用程序。然而,代币智能合约在技术薄弱和操纵风险方面面临安全挑战。在这项工作中,我们简要描述了操纵风险,并提出了TokenAuditor,这是一个模糊测试框架,可以检测代币智能合约中的这些风险。TokenAuditor基于契约字节码构造基本块,并采用稀有性选择和突变策略生成测试用例。主要思想是选择自模糊测试开始以来已经遇到罕见基本块的测试用例作为候选,并对它们执行突变操作。在我们的评估中,TokenAudiotr在4021份现实世界的代币合约中发现了四种类型的664种操纵风险。
{"title":"TokenAuditor: Detecting Manipulation Risk in Token Smart Contract by Fuzzing","authors":"Mingpei Cao, Yueze Zhang, Zhenxuan Feng, Jiahao Hu, Yuesheng Zhu","doi":"10.1109/QRS57517.2022.00071","DOIUrl":"https://doi.org/10.1109/QRS57517.2022.00071","url":null,"abstract":"Decentralized cryptocurrencies are influential smart contract applications in the blockchain, drawing interest from industry and academia. The capacity to govern and manage token behavior provided by the token smart contract adds to thriving decentralized applications. However, token smart contracts face security challenges in technology weakness and manipulation risks. In this work, we briefly describe the manipulation risk and propose TokenAuditor, a fuzzing framework detecting those risks in token smart contracts. TokenAuditor constructs basic blocks based on the contract bytecodes and adopts the rarity selection and mutation strategy to generate test cases. The main idea is to select the test cases that have hit rare basic blocks since the fuzzing started as candidates and perform mutation operations on them. In our evaluation, TokenAudiotr discovered 664 manipulation risks of four types in 4021 real-world token contracts.","PeriodicalId":143812,"journal":{"name":"2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131869412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2022 IEEE 22nd International Conference on Software Quality, Reliability and Security (QRS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1