首页 > 最新文献

2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)最新文献

英文 中文
Performance Bottleneck Analysis of Drone Computation Offloading to a Shared Fog Node 无人机计算向共享雾节点卸载的性能瓶颈分析
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00070
Qingyang Zhang, F. Machida, E. Andrade
Computing in drones has recently become popular for various real-world applications. To assure the performance and reliability of drone computing, systems can also adopt computation offloading to a nearby fog or edge server through a wireless network. As the offloading performance is significantly affected by the amount of workload, the network stability, and the competing use of a shared resource, performance estimation is essential for such systems. In this paper, we analyze the performance bottleneck of a drone system consisting of multiple drones that offload the tasks to a shared fog node. We investigate how resource conflict due to computation offloading causes the performance bottleneck of the drone computation system. To model the behavior of the system and analyze the performance and availability, we use Stochastic Reward Nets (SRN s). Through the numerical experiments, we confirm that the benefit of computation offloading deteriorates as the number of competing drones increases. To overcome the performance bottleneck, we also discuss potential solutions to mitigate the issue of a shared fog node.
无人机上的计算最近在各种实际应用中变得流行起来。为了确保无人机计算的性能和可靠性,系统还可以通过无线网络将计算卸载到附近的雾服务器或边缘服务器。由于负载量、网络稳定性和共享资源的竞争性使用会显著影响卸载性能,因此对此类系统进行性能评估至关重要。在本文中,我们分析了由多个无人机组成的无人机系统将任务卸载到共享雾节点的性能瓶颈。我们研究了由于计算卸载引起的资源冲突如何导致无人机计算系统的性能瓶颈。为了模拟系统的行为并分析性能和可用性,我们使用了随机奖励网(SRN)。通过数值实验,我们证实了计算卸载的好处随着竞争无人机数量的增加而恶化。为了克服性能瓶颈,我们还讨论了缓解共享雾节点问题的潜在解决方案。
{"title":"Performance Bottleneck Analysis of Drone Computation Offloading to a Shared Fog Node","authors":"Qingyang Zhang, F. Machida, E. Andrade","doi":"10.1109/ISSREW55968.2022.00070","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00070","url":null,"abstract":"Computing in drones has recently become popular for various real-world applications. To assure the performance and reliability of drone computing, systems can also adopt computation offloading to a nearby fog or edge server through a wireless network. As the offloading performance is significantly affected by the amount of workload, the network stability, and the competing use of a shared resource, performance estimation is essential for such systems. In this paper, we analyze the performance bottleneck of a drone system consisting of multiple drones that offload the tasks to a shared fog node. We investigate how resource conflict due to computation offloading causes the performance bottleneck of the drone computation system. To model the behavior of the system and analyze the performance and availability, we use Stochastic Reward Nets (SRN s). Through the numerical experiments, we confirm that the benefit of computation offloading deteriorates as the number of competing drones increases. To overcome the performance bottleneck, we also discuss potential solutions to mitigate the issue of a shared fog node.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114377706","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Machine-Learned Specifications for the Verification and Validation of Autonomous Cyberphysical Systems 自主网络物理系统验证与验证的机器学习规范
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00089
D. Drusinsky, J. Michael, Matthew Litton
Machine learning classifiers can be used as speci-fications for runtime monitoring (RM), which in turn supports evaluating autonomous systems during design-time and detecting/responding to exceptional situations during system operation. In this paper we describe how the use of machine-learned specifications enhances the effectiveness of RM for verification and validation (V & V) of autonomous cyberphysical systems (CPSs). In addition, we show that the development of machine-learned specifications has a predictable cost, at less than $100 per specification, using 2022 cloud computing pricing. Finally, a key benefit of our approach is that developing specifications by training ML models brings the task of developing robust specifications from the realm of doctoral-level experts into the domain of system developers and engineers.
机器学习分类器可以用作运行时监控(RM)的规范,这反过来支持在设计期间评估自主系统,并在系统运行期间检测/响应异常情况。在本文中,我们描述了机器学习规范的使用如何增强RM对自主网络物理系统(cps)的验证和验证(V & V)的有效性。此外,我们表明,机器学习规范的开发具有可预测的成本,使用2022年的云计算定价,每个规范的成本低于100美元。最后,我们的方法的一个关键好处是,通过训练ML模型来开发规范,将开发健壮规范的任务从博士级专家的领域带入了系统开发人员和工程师的领域。
{"title":"Machine-Learned Specifications for the Verification and Validation of Autonomous Cyberphysical Systems","authors":"D. Drusinsky, J. Michael, Matthew Litton","doi":"10.1109/ISSREW55968.2022.00089","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00089","url":null,"abstract":"Machine learning classifiers can be used as speci-fications for runtime monitoring (RM), which in turn supports evaluating autonomous systems during design-time and detecting/responding to exceptional situations during system operation. In this paper we describe how the use of machine-learned specifications enhances the effectiveness of RM for verification and validation (V & V) of autonomous cyberphysical systems (CPSs). In addition, we show that the development of machine-learned specifications has a predictable cost, at less than $100 per specification, using 2022 cloud computing pricing. Finally, a key benefit of our approach is that developing specifications by training ML models brings the task of developing robust specifications from the realm of doctoral-level experts into the domain of system developers and engineers.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121904095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automated Dependability Assessment in DevOps Environments DevOps环境中的自动化可靠性评估
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00046
James J. Cusick, Alberto Avritzer, Allen Tse, Andrea Janes
We present an industrial experience report of the application of automated dependability assessment to a major Digital Transformation Initiative. This effort involved significant investment in the development, automation and visualization of the Non-Functional Requirements (NFRs). We present the details around the objectives of the NFR effort, challenges, the technical approach adopted, summary of results, and lessons learned. A clear description of steps which worked well are provided as well as the challenges which were found in meeting a wide range of technical and organizational goals in this process. A focus on the methods and results of developing NFRs within a DevOps environment and across a large heterogeneous computing platform are emphasized.
我们提出了一份自动化可靠性评估应用于主要数字化转型计划的行业经验报告。这项工作涉及对非功能需求(nfr)的开发、自动化和可视化的重大投资。我们将详细介绍NFR工作的目标、挑战、采用的技术方法、结果总结和经验教训。书中清晰地描述了运作良好的步骤,以及在实现这一过程中广泛的技术和组织目标时所遇到的挑战。重点是在DevOps环境中和跨大型异构计算平台开发nfr的方法和结果。
{"title":"Automated Dependability Assessment in DevOps Environments","authors":"James J. Cusick, Alberto Avritzer, Allen Tse, Andrea Janes","doi":"10.1109/ISSREW55968.2022.00046","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00046","url":null,"abstract":"We present an industrial experience report of the application of automated dependability assessment to a major Digital Transformation Initiative. This effort involved significant investment in the development, automation and visualization of the Non-Functional Requirements (NFRs). We present the details around the objectives of the NFR effort, challenges, the technical approach adopted, summary of results, and lessons learned. A clear description of steps which worked well are provided as well as the challenges which were found in meeting a wide range of technical and organizational goals in this process. A focus on the methods and results of developing NFRs within a DevOps environment and across a large heterogeneous computing platform are emphasized.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126982001","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
XAI for Communication Networks XAI代表通信网络
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00093
Sayan Mukherjee, J. Rupe, J. Zhu
Explainable AI (XAI) is a topic of intense activity in the research community today. However, for AI models deployed in the critical infrastructure of communications networks, explainability alone is not enough to earn the trust of network operations teams comprising human experts with many decades of collective experience. In the present work we discuss some use cases in communications networks and state some of the additional properties, including accountability, that XAI models would have to satisfy before they can be widely deployed. In particular, we advocate for a human-in-the-Ioop approach to train and validate XAI models. Additionally, we discuss the use cases of XAI models around improving data preprocessing and data augmentation techniques, and refining data labeling rules for producing consistently labeled network datasets.
可解释人工智能(XAI)是当今研究界的一个热门话题。然而,对于部署在通信网络关键基础设施中的人工智能模型,仅凭可解释性不足以赢得由具有数十年集体经验的人类专家组成的网络运营团队的信任。在目前的工作中,我们讨论了通信网络中的一些用例,并说明了XAI模型在被广泛部署之前必须满足的一些附加属性,包括责任。特别地,我们提倡使用human- In -the- loop方法来训练和验证XAI模型。此外,我们还讨论了XAI模型的用例,围绕改进数据预处理和数据增强技术,以及改进数据标记规则以生成一致标记的网络数据集。
{"title":"XAI for Communication Networks","authors":"Sayan Mukherjee, J. Rupe, J. Zhu","doi":"10.1109/ISSREW55968.2022.00093","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00093","url":null,"abstract":"Explainable AI (XAI) is a topic of intense activity in the research community today. However, for AI models deployed in the critical infrastructure of communications networks, explainability alone is not enough to earn the trust of network operations teams comprising human experts with many decades of collective experience. In the present work we discuss some use cases in communications networks and state some of the additional properties, including accountability, that XAI models would have to satisfy before they can be widely deployed. In particular, we advocate for a human-in-the-Ioop approach to train and validate XAI models. Additionally, we discuss the use cases of XAI models around improving data preprocessing and data augmentation techniques, and refining data labeling rules for producing consistently labeled network datasets.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124512584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Effective Performance Fuzzing 迈向有效的性能模糊测试
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00055
Yiqun Chen, M. Bradbury, N. Suri
Fuzzing is an automated testing technique that utilizes injection of random inputs in a target program to help uncover vulnerabilities. Performance fuzzing extends the classic fuzzing approach and generates inputs that trigger poor performance. During our evaluation of performance fuzzing tools, we have identified certain conventionally used assumptions that do not always hold true. Our research (re)evaluates PERFFUZZ [1] in order to identify the limitations of current techniques, and guide the direction of future work for improvements to performance fuzzing. Our experimental results highlight two specific limitations. Firstly, we identify the assumption that the length of execution paths correlate to program performance is not always the case, and thus cannot reflect the quality of test cases generated by performance fuzzing. Secondly, the default testing parameters by the fuzzing process (timeouts and size limits) overly confine the input search space. Based on these observations, we suggest further investigation on performance fuzzing guidance, as well as controlled fuzzing and testing parameters.
模糊测试是一种自动化测试技术,它利用在目标程序中注入随机输入来帮助发现漏洞。性能模糊测试扩展了经典的模糊测试方法,并生成触发性能差的输入。在我们对性能模糊测试工具的评估过程中,我们已经确定了某些传统使用的假设并不总是正确的。我们的研究(重新)评估了PERFFUZZ[1],以确定当前技术的局限性,并指导未来改进性能模糊测试的工作方向。我们的实验结果突出了两个特定的局限性。首先,我们确定执行路径的长度与程序性能相关的假设并不总是如此,因此不能反映性能模糊测试生成的测试用例的质量。其次,模糊处理的默认测试参数(超时和大小限制)过度限制了输入搜索空间。基于这些观察结果,我们建议进一步研究性能模糊指导,以及控制模糊和测试参数。
{"title":"Towards Effective Performance Fuzzing","authors":"Yiqun Chen, M. Bradbury, N. Suri","doi":"10.1109/ISSREW55968.2022.00055","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00055","url":null,"abstract":"Fuzzing is an automated testing technique that utilizes injection of random inputs in a target program to help uncover vulnerabilities. Performance fuzzing extends the classic fuzzing approach and generates inputs that trigger poor performance. During our evaluation of performance fuzzing tools, we have identified certain conventionally used assumptions that do not always hold true. Our research (re)evaluates PERFFUZZ [1] in order to identify the limitations of current techniques, and guide the direction of future work for improvements to performance fuzzing. Our experimental results highlight two specific limitations. Firstly, we identify the assumption that the length of execution paths correlate to program performance is not always the case, and thus cannot reflect the quality of test cases generated by performance fuzzing. Secondly, the default testing parameters by the fuzzing process (timeouts and size limits) overly confine the input search space. Based on these observations, we suggest further investigation on performance fuzzing guidance, as well as controlled fuzzing and testing parameters.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127719026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
WoSAR 2022 Workshop Keynotes WoSAR 2022研讨会主题演讲
Pub Date : 2022-10-01 DOI: 10.1109/issrew55968.2022.00014
{"title":"WoSAR 2022 Workshop Keynotes","authors":"","doi":"10.1109/issrew55968.2022.00014","DOIUrl":"https://doi.org/10.1109/issrew55968.2022.00014","url":null,"abstract":"","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127756999","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards the Quantitative Verification of Deep Learning for Safe Perception 面向安全感知的深度学习定量验证
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00069
Philipp Schleiss, Yuki Hagiwara, Iwo Kurzidem, Francesco Carella
Deep learning (DL) is seen as an inevitable building block for perceiving the environment with sufficient detail and accuracy as required by automated driving functions. Despite this, its black-box nature and the therewith intertwined unpredictability still hinders its use in safety-critical systems. As such, this work addresses the problem of making this seemingly unpredictable nature measurable by providing a risk-based verification strategy, such as required by ISO 21448. In detail, a method is developed to break down acceptable risk into quantitative performance targets of individual DL-based components along the perception architecture. To verify these targets, the DL input space is split into areas according to the dimensions of a fine-grained operational design domain $(mu mathbf{ODD})$. As it is not feasible to reach full test coverage, the strategy suggests to distribute test efforts across these areas according to the associated risk. Moreover, the testing approach provides answers with respect to how much test coverage and confidence in the test result is required and how these figures relate to safety integrity levels (SILs).
深度学习(DL)被视为自动驾驶功能所要求的以足够的细节和准确性感知环境的不可避免的组成部分。尽管如此,它的黑盒子性质和由此交织在一起的不可预测性仍然阻碍了它在安全关键系统中的使用。因此,这项工作通过提供基于风险的验证策略(如ISO 21448所要求的)来解决使这种看似不可预测的性质可测量的问题。详细地说,开发了一种方法,将可接受的风险分解为沿着感知架构的单个基于dl的组件的定量性能目标。为了验证这些目标,DL输入空间根据细粒度操作设计域$(mu mathbf{ODD})$的维度划分为多个区域。由于达到完全的测试覆盖是不可行的,该策略建议根据相关的风险在这些区域之间分配测试工作。此外,测试方法提供了关于测试覆盖率和测试结果置信度的答案,以及这些数字如何与安全完整性水平(SILs)相关。
{"title":"Towards the Quantitative Verification of Deep Learning for Safe Perception","authors":"Philipp Schleiss, Yuki Hagiwara, Iwo Kurzidem, Francesco Carella","doi":"10.1109/ISSREW55968.2022.00069","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00069","url":null,"abstract":"Deep learning (DL) is seen as an inevitable building block for perceiving the environment with sufficient detail and accuracy as required by automated driving functions. Despite this, its black-box nature and the therewith intertwined unpredictability still hinders its use in safety-critical systems. As such, this work addresses the problem of making this seemingly unpredictable nature measurable by providing a risk-based verification strategy, such as required by ISO 21448. In detail, a method is developed to break down acceptable risk into quantitative performance targets of individual DL-based components along the perception architecture. To verify these targets, the DL input space is split into areas according to the dimensions of a fine-grained operational design domain $(mu mathbf{ODD})$. As it is not feasible to reach full test coverage, the strategy suggests to distribute test efforts across these areas according to the associated risk. Moreover, the testing approach provides answers with respect to how much test coverage and confidence in the test result is required and how these figures relate to safety integrity levels (SILs).","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125968972","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
SSSML 2022 Workshop Committee: ISSREW 2022 SSSML 2022研讨会委员会:ISSREW 2022
Pub Date : 2022-10-01 DOI: 10.1109/issrew55968.2022.00025
{"title":"SSSML 2022 Workshop Committee: ISSREW 2022","authors":"","doi":"10.1109/issrew55968.2022.00025","DOIUrl":"https://doi.org/10.1109/issrew55968.2022.00025","url":null,"abstract":"","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131044873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Method for Component Evaluation for Live Testing of Cloud Systems 一种面向云系统实时测试的组件评估方法
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00045
Oussama Jebbar, F. Khendek, M. Toeroe
Live testing is about testing a subsystem in production without causing any unacceptable disturbance to the production traffic. A subsystem is tested in production for multiple purposes such as deployment verification, fault prediction, fault localization, etc. The main challenge of live testing is alleviating the risk of test interferences as it may lead to a violation of a system's functional or non-functional requirements. To properly handle this risk, one needs to know which components present a risk of test interferences and what is the cost of the countermeasures to handle that risk. Existing literature relies heavily on human judgement, which can be time consuming, not always feasible, may provide misleading insight. In this paper we go through the challenges of automating this evaluation process and propose a solution to overcome them. Our solution consists of a method for components evaluation which goes through three steps, evaluation of test interferences that may manifest in external behaviour, evaluation of test interferences that may manifest in resource consumption, and finally the evaluation of the cost of implementing the countermeasures to overcome the risk of test interferences.
实时测试是指在不给生产流量造成任何不可接受的干扰的情况下测试生产中的子系统。子系统在生产环境中进行测试,用于多种目的,如部署验证、故障预测、故障定位等。实时测试的主要挑战是减轻测试干扰的风险,因为它可能导致违反系统的功能或非功能需求。要正确处理此风险,需要知道哪些组件存在测试干扰的风险,以及处理该风险的对策的成本是多少。现有文献在很大程度上依赖于人类的判断,这可能是耗时的,并不总是可行的,可能会提供误导性的见解。在本文中,我们讨论了自动化评估过程的挑战,并提出了克服这些挑战的解决方案。我们的解决方案由组件评估方法组成,该方法经过三个步骤,评估可能在外部行为中表现出来的测试干扰,评估可能在资源消耗中表现出来的测试干扰,最后评估实现克服测试干扰风险的对策的成本。
{"title":"A Method for Component Evaluation for Live Testing of Cloud Systems","authors":"Oussama Jebbar, F. Khendek, M. Toeroe","doi":"10.1109/ISSREW55968.2022.00045","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00045","url":null,"abstract":"Live testing is about testing a subsystem in production without causing any unacceptable disturbance to the production traffic. A subsystem is tested in production for multiple purposes such as deployment verification, fault prediction, fault localization, etc. The main challenge of live testing is alleviating the risk of test interferences as it may lead to a violation of a system's functional or non-functional requirements. To properly handle this risk, one needs to know which components present a risk of test interferences and what is the cost of the countermeasures to handle that risk. Existing literature relies heavily on human judgement, which can be time consuming, not always feasible, may provide misleading insight. In this paper we go through the challenges of automating this evaluation process and propose a solution to overcome them. Our solution consists of a method for components evaluation which goes through three steps, evaluation of test interferences that may manifest in external behaviour, evaluation of test interferences that may manifest in resource consumption, and finally the evaluation of the cost of implementing the countermeasures to overcome the risk of test interferences.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130949299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correlating Test Events With Monitoring Logs For Test Log Reduction And Anomaly Prediction 将测试事件与监控日志相关联,以减少测试日志并预测异常
Pub Date : 2022-10-01 DOI: 10.1109/ISSREW55968.2022.00079
Bahareh Afshinpour, Roland Groz, Massih-Reza Amini
Automated fault identification in long test logs is a tough problem, mainly because of their sequential character and the impossibility of constructing training sets for zero-day faults. To reduce software testers' workload, rule-based approaches have been extensively investigated as solutions for efficiently finding and predicting the fault. Based on software system status monitoring log analysis, we propose a new learning-based technique to automate anomaly detection, correlate test events to anomalies and predict system failures. Since the meaning of fault is not established in system status monitoring-based fault detection, the suggested technique first detects periods of time when a software system status encounters aberrant situations (Bug-Zones). The suggested technique is then tested in a real-time system for anomaly prediction of new tests. The model may be used in two ways. It can assist testers to focus on faulty-like time intervals by reducing the number of test logs. It may also be used to forecast a Bug-Zone in an online system, allowing system administrators to anticipate or even prevent a system failure. An extensive study on a real-world database acquired by a telecommunication operator demonstrates that our approach achieves 71 % accuracy as a Bug-Zones predictor.
长测试日志中的故障自动识别是一个棘手的问题,主要是因为长测试日志具有序列性,且无法构建零日故障的训练集。为了减少软件测试人员的工作量,基于规则的方法作为有效发现和预测故障的解决方案得到了广泛的研究。在软件系统状态监测日志分析的基础上,提出了一种新的基于学习的异常自动检测技术,将测试事件与异常关联起来,预测系统故障。由于在基于系统状态监视的故障检测中不能确定故障的含义,因此建议的技术首先检测软件系统状态遇到异常情况的时间段(Bug-Zones)。然后在新测试的异常预测实时系统中对所建议的技术进行了测试。该模型可用于两种方式。它可以帮助测试人员通过减少测试日志的数量来关注类似错误的时间间隔。它还可以用于预测在线系统中的Bug-Zone,允许系统管理员预测甚至防止系统故障。对一家电信运营商获得的真实世界数据库的广泛研究表明,我们的方法作为bug区域预测器达到了71%的准确率。
{"title":"Correlating Test Events With Monitoring Logs For Test Log Reduction And Anomaly Prediction","authors":"Bahareh Afshinpour, Roland Groz, Massih-Reza Amini","doi":"10.1109/ISSREW55968.2022.00079","DOIUrl":"https://doi.org/10.1109/ISSREW55968.2022.00079","url":null,"abstract":"Automated fault identification in long test logs is a tough problem, mainly because of their sequential character and the impossibility of constructing training sets for zero-day faults. To reduce software testers' workload, rule-based approaches have been extensively investigated as solutions for efficiently finding and predicting the fault. Based on software system status monitoring log analysis, we propose a new learning-based technique to automate anomaly detection, correlate test events to anomalies and predict system failures. Since the meaning of fault is not established in system status monitoring-based fault detection, the suggested technique first detects periods of time when a software system status encounters aberrant situations (Bug-Zones). The suggested technique is then tested in a real-time system for anomaly prediction of new tests. The model may be used in two ways. It can assist testers to focus on faulty-like time intervals by reducing the number of test logs. It may also be used to forecast a Bug-Zone in an online system, allowing system administrators to anticipate or even prevent a system failure. An extensive study on a real-world database acquired by a telecommunication operator demonstrates that our approach achieves 71 % accuracy as a Bug-Zones predictor.","PeriodicalId":178302,"journal":{"name":"2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128762729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2022 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1