首页 > 最新文献

Journal of Systems and Software最新文献

英文 中文
Automatic generation of efficient oracles: The less-than case 自动生成高效报文小于的情况
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-09-04 DOI: 10.1016/j.jss.2024.112203

Grover’s algorithm is a well-known contribution to quantum computing. It searches one value within an unordered sequence faster than any classical algorithm. A fundamental part of this algorithm is the so-called oracle, a quantum circuit that marks the quantum state corresponding to the desired value. A generalisation of it is the oracle for Amplitude Amplification, that marks multiple desired states. In this work we present a classical algorithm that builds a phase-marking oracle for Amplitude Amplification. This oracle performs a less-than operation, marking states representing natural numbers smaller than a given one. Results of both simulations and experiments are shown to prove its functionality. This less-than oracle implementation works on any number of qubits and does not require any ancilla qubits. Regarding depth, the proposed implementation is compared with the one generated by Qiskit automatic method, Diagonal. We show that the depth of our less-than oracle implementation is always lower. In addition, a comparison with another method for oracle generation in terms of gate count is also conducted to prove the efficiency of our method. The result presented here is part of a research work that aims to achieve reusable quantum operations that can be composed to perform more complex ones. The final aim is to provide Quantum Developers with tools that can be easily integrated in their programs/circuits.

格罗弗算法是量子计算的一项著名贡献。它在无序序列中搜索一个值的速度比任何经典算法都快。该算法的一个基本部分是所谓的 "神谕",它是一个量子电路,用于标记与所需值相对应的量子态。它的一个概括是振幅放大算法,可以标记多个所需的状态。在这项工作中,我们提出了一种经典算法,它能为振幅放大建立相位标记神谕。该算法执行 "小于 "操作,标记代表小于给定自然数的状态。模拟和实验结果都证明了它的功能。这种 "小于 "神谕的实现适用于任何数量的量子比特,并且不需要任何辅助量子比特。在深度方面,我们将提议的实现与 Qiskit 自动方法 Diagonal 生成的实现进行了比较。结果表明,我们的小于甲骨文实现的深度总是更低。此外,我们还在门数方面与另一种甲骨文生成方法进行了比较,以证明我们方法的效率。本文介绍的成果是一项研究工作的一部分,该研究旨在实现可重复使用的量子运算,这些运算可以组成更复杂的运算。最终目的是为量子开发人员提供可轻松集成到他们的程序/电路中的工具。
{"title":"Automatic generation of efficient oracles: The less-than case","authors":"","doi":"10.1016/j.jss.2024.112203","DOIUrl":"10.1016/j.jss.2024.112203","url":null,"abstract":"<div><p>Grover’s algorithm is a well-known contribution to quantum computing. It searches one value within an unordered sequence faster than any classical algorithm. A fundamental part of this algorithm is the so-called oracle, a quantum circuit that marks the quantum state corresponding to the desired value. A generalisation of it is the oracle for Amplitude Amplification, that marks multiple desired states. In this work we present a classical algorithm that builds a phase-marking oracle for Amplitude Amplification. This oracle performs a less-than operation, marking states representing natural numbers smaller than a given one. Results of both simulations and experiments are shown to prove its functionality. This less-than oracle implementation works on any number of qubits and does not require any ancilla qubits. Regarding depth, the proposed implementation is compared with the one generated by Qiskit automatic method, <em>Diagonal</em>. We show that the depth of our less-than oracle implementation is always lower. In addition, a comparison with another method for oracle generation in terms of gate count is also conducted to prove the efficiency of our method. The result presented here is part of a research work that aims to achieve reusable quantum operations that can be composed to perform more complex ones. The final aim is to provide Quantum Developers with tools that can be easily integrated in their programs/circuits.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0164121224002474/pdfft?md5=045c01aa1d649973c047ead28c3395fc&pid=1-s2.0-S0164121224002474-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142161758","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring the problems, their causes and solutions of AI pair programming: A study on GitHub and Stack Overflow 探索人工智能结对编程的问题、原因和解决方案:对 GitHub 和 Stack Overflow 的研究
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-09-02 DOI: 10.1016/j.jss.2024.112204

With the recent advancement of Artificial Intelligence (AI) and Large Language Models (LLMs), AI-based code generation tools become a practical solution for software development. GitHub Copilot, the AI pair programmer, utilizes machine learning models trained on a large corpus of code snippets to generate code suggestions using natural language processing. Despite its popularity in software development, there is limited empirical evidence on the actual experiences of practitioners who work with Copilot. To this end, we conducted an empirical study to understand the problems that practitioners face when using Copilot, as well as their underlying causes and potential solutions. We collected data from 473 GitHub issues, 706 GitHub discussions, and 142 Stack Overflow posts. Our results reveal that (1) Operation Issue and Compatibility Issue are the most common problems faced by Copilot users, (2) Copilot Internal Error, Network Connection Error, and Editor/IDE Compatibility Issue are identified as the most frequent causes, and (3) Bug Fixed by Copilot, Modify Configuration/Setting, and Use Suitable Version are the predominant solutions. Based on the results, we discuss the potential areas of Copilot for enhancement, and provide the implications for the Copilot users, the Copilot team, and researchers.

随着近年来人工智能(AI)和大型语言模型(LLM)的发展,基于 AI 的代码生成工具已成为软件开发的实用解决方案。GitHub Copilot(人工智能配对程序员)利用在大量代码片段语料库中训练的机器学习模型,通过自然语言处理生成代码建议。尽管 Copilot 在软件开发领域很受欢迎,但有关从业人员使用 Copilot 的实际经验的实证证据却很有限。为此,我们开展了一项实证研究,以了解从业人员在使用 Copilot 时所面临的问题及其根本原因和潜在解决方案。我们从 473 个 GitHub 问题、706 个 GitHub 讨论和 142 个 Stack Overflow 帖子中收集了数据。结果显示:(1) 操作问题和兼容性问题是 Copilot 用户面临的最常见问题;(2) Copilot 内部错误、网络连接错误和编辑器/IDE 兼容性问题是最常见的原因;(3) Copilot 修正的错误、修改配置/设置和使用合适的版本是最主要的解决方案。基于这些结果,我们讨论了 Copilot 的潜在改进领域,并提供了对 Copilot 用户、Copilot 团队和研究人员的启示。
{"title":"Exploring the problems, their causes and solutions of AI pair programming: A study on GitHub and Stack Overflow","authors":"","doi":"10.1016/j.jss.2024.112204","DOIUrl":"10.1016/j.jss.2024.112204","url":null,"abstract":"<div><p>With the recent advancement of Artificial Intelligence (AI) and Large Language Models (LLMs), AI-based code generation tools become a practical solution for software development. GitHub Copilot, the AI pair programmer, utilizes machine learning models trained on a large corpus of code snippets to generate code suggestions using natural language processing. Despite its popularity in software development, there is limited empirical evidence on the actual experiences of practitioners who work with Copilot. To this end, we conducted an empirical study to understand the problems that practitioners face when using Copilot, as well as their underlying causes and potential solutions. We collected data from 473 GitHub issues, 706 GitHub discussions, and 142 Stack Overflow posts. Our results reveal that (1) <em>Operation Issue</em> and <em>Compatibility Issue</em> are the most common problems faced by Copilot users, (2) <em>Copilot Internal Error</em>, <em>Network Connection Error</em>, and <em>Editor/IDE Compatibility Issue</em> are identified as the most frequent causes, and (3) <em>Bug Fixed by Copilot</em>, <em>Modify Configuration/Setting</em>, and <em>Use Suitable Version</em> are the predominant solutions. Based on the results, we discuss the potential areas of Copilot for enhancement, and provide the implications for the Copilot users, the Copilot team, and researchers.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142150765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DeepFeature: Guiding adversarial testing for deep neural network systems using robust features DeepFeature:利用鲁棒特征指导深度神经网络系统的对抗测试
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-31 DOI: 10.1016/j.jss.2024.112201

With the deployment of Deep Neural Network (DNN) systems in security-critical fields, more and more researchers are concerned about DNN robustness. Unfortunately, DNNs are vulnerable to adversarial attacks and produce completely wrong outputs. This inspired numerous testing works devoted to improving the adversarial robustness of DNNs. Coverage and uncertainty criteria were proposed to guide sample selections for DNN retraining. However, they are greatly limited to evaluating DNN abnormal behaviors rather than locating the root cause of adversarial vulnerability. This work aims to bridge this gap. We propose an adversarial testing framework, DeepFeature, using robust features. DeepFeature generates robust features related to the model decision-making. It locates the weak features within these features that fail to be transformed by the DNN. They are the main culprits of vulnerability. DeepFeature selects diverse samples containing weak features for adversarial retraining. Our evaluation shows that DeepFeature significantly improves overall robustness, average improved by 77.83%, and individual robustness, average improved by 42.81‰, of the models in adversarial testing. Compared with coverage and uncertainty criteria, these two performances are improved by 3.93% and 15.00% in DeepFeature, respectively. The positive correlation coefficient between DeepFeature and improved robustness can achieve 0.858, and the p-value is 0.001.

随着深度神经网络(DNN)系统在安全关键领域的部署,越来越多的研究人员开始关注 DNN 的鲁棒性。不幸的是,DNN 容易受到对抗性攻击,并产生完全错误的输出。这激发了众多致力于提高 DNN 抗对抗鲁棒性的测试工作。人们提出了覆盖率和不确定性标准来指导 DNN 再训练的样本选择。然而,它们在很大程度上仅限于评估 DNN 的异常行为,而不是找出对抗性漏洞的根本原因。本研究旨在弥补这一不足。我们提出了一种使用鲁棒特征的对抗测试框架 DeepFeature。DeepFeature 生成与模型决策相关的鲁棒特征。它能找出这些特征中无法被 DNN 转换的弱特征。它们是造成漏洞的罪魁祸首。DeepFeature 会选择包含弱特征的各种样本进行对抗性再训练。我们的评估表明,DeepFeature 显著提高了模型在对抗测试中的整体鲁棒性(平均提高 77.83%)和个体鲁棒性(平均提高 42.81‰)。与覆盖率和不确定性标准相比,DeepFeature 的这两项性能分别提高了 3.93% 和 15.00%。DeepFeature 与鲁棒性提高之间的正相关系数为 0.858,P 值为 0.001。
{"title":"DeepFeature: Guiding adversarial testing for deep neural network systems using robust features","authors":"","doi":"10.1016/j.jss.2024.112201","DOIUrl":"10.1016/j.jss.2024.112201","url":null,"abstract":"<div><p>With the deployment of Deep Neural Network (DNN) systems in security-critical fields, more and more researchers are concerned about DNN robustness. Unfortunately, DNNs are vulnerable to adversarial attacks and produce completely wrong outputs. This inspired numerous testing works devoted to improving the adversarial robustness of DNNs. Coverage and uncertainty criteria were proposed to guide sample selections for DNN retraining. However, they are greatly limited to evaluating DNN abnormal behaviors rather than locating the root cause of adversarial vulnerability. This work aims to bridge this gap. We propose an adversarial testing framework, DeepFeature, using robust features. DeepFeature generates robust features related to the model decision-making. It locates the weak features within these features that fail to be transformed by the DNN. They are the main culprits of vulnerability. DeepFeature selects diverse samples containing weak features for adversarial retraining. Our evaluation shows that DeepFeature significantly improves overall robustness, average improved by 77.83%, and individual robustness, average improved by 42.81‰, of the models in adversarial testing. Compared with coverage and uncertainty criteria, these two performances are improved by 3.93% and 15.00% in DeepFeature, respectively. The positive correlation coefficient between DeepFeature and improved robustness can achieve 0.858, and the <span><math><mi>p</mi></math></span>-value is 0.001.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142167885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
How are discussions linked? A link analysis study on GitHub Discussions 讨论是如何链接的?GitHub 讨论的链接分析研究
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-30 DOI: 10.1016/j.jss.2024.112196

Software development requires collaborative efforts and consensus among developers, emphasizing the need for effective communication and knowledge sharing within the teams. In line with this, GitHub introduced GitHub Discussions, a collaborative communication feature for the community around an open-source or internal project. The user-friendly interface of Discussions facilitates easy sharing of links to different resources. Maintainers highlight this feature’s significance, enabling users to maintain a well-informed project environment. We hypothesize that link-sharing activities in Discussions contribute to disseminating project knowledge. To investigate this hypothesis, we conducted a mixed-method study combining qualitative and quantitative analysis based on a convenience sample of ten open-source projects. We aimed to gain insight into the scope and intentions behind these shared links. Our findings indicate that link-sharing activities are common in the Discussions. Users share links to resources directly or indirectly related to the project/repository. Discussions users share links to project documentation, source code, and issues, aiming to clarify concepts, provide supplementary information, offer context to questions, and suggest new features. These findings offer insights for project maintainers to understand Discussions usage better and enable the GitHub Engineering team to promote the feature adoption.

软件开发需要开发人员之间的协同努力和共识,这就强调了团队内部有效沟通和知识共享的必要性。为此,GitHub 推出了 GitHub Discussions,这是一项围绕开源项目或内部项目的社区协作交流功能。讨论 "界面友好,便于分享不同资源的链接。维护者强调了这一功能的重要性,它使用户能够维护一个消息灵通的项目环境。我们假设,讨论区中的链接共享活动有助于传播项目知识。为了研究这一假设,我们以十个开源项目为样本,进行了定性和定量分析相结合的混合方法研究。我们旨在深入了解这些共享链接背后的范围和意图。我们的研究结果表明,链接共享活动在讨论中很常见。用户共享与项目/资源库直接或间接相关的资源链接。讨论用户共享项目文档、源代码和问题的链接,目的是澄清概念、提供补充信息、提供问题的来龙去脉以及建议新功能。这些发现为项目维护者更好地了解讨论区的使用情况提供了见解,并使 GitHub 工程团队能够促进功能的采用。
{"title":"How are discussions linked? A link analysis study on GitHub Discussions","authors":"","doi":"10.1016/j.jss.2024.112196","DOIUrl":"10.1016/j.jss.2024.112196","url":null,"abstract":"<div><p>Software development requires collaborative efforts and consensus among developers, emphasizing the need for effective communication and knowledge sharing within the teams. In line with this, GitHub introduced GitHub Discussions, a collaborative communication feature for the community around an open-source or internal project. The user-friendly interface of Discussions facilitates easy sharing of links to different resources. Maintainers highlight this feature’s significance, enabling users to maintain a well-informed project environment. We hypothesize that link-sharing activities in Discussions contribute to disseminating project knowledge. To investigate this hypothesis, we conducted a mixed-method study combining qualitative and quantitative analysis based on a convenience sample of ten open-source projects. We aimed to gain insight into the scope and intentions behind these shared links. Our findings indicate that link-sharing activities are common in the Discussions. Users share links to resources directly or indirectly related to the project/repository. Discussions users share links to project documentation, source code, and issues, aiming to clarify concepts, provide supplementary information, offer context to questions, and suggest new features. These findings offer insights for project maintainers to understand Discussions usage better and enable the GitHub Engineering team to promote the feature adoption.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142150764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Software modernization powered by dynamic language product lines 动态语言产品线推动软件现代化
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-28 DOI: 10.1016/j.jss.2024.112188

Legacy software poses a critical challenge for organizations due to the costs of maintaining and modernizing outdated systems, as well as the scarcity of experts in aging programming languages. The issue extends beyond commercial applications, affecting public administration, as exemplified by the urgent need for COBOL programmers during the COVID-19 pandemic. In response, this work introduces a modernization approach based on dynamic language product lines, a subset of dynamic software product lines. This approach leverages open language implementations and dynamically generated micro-languages for the incremental migration of legacy systems to modern technologies. The language can be reconfigured at runtime to adapt to the execution of either legacy or modern code, and to generate a compatibility layer between the data types handled by the two languages. Through this process, the costs of modernizing legacy systems can be spread across several iterations, as developers can replace legacy code incrementally, with legacy and modern code coexisting until a complete refactoring is possible. By moving the overhead of making legacy and modern features work together in a hybrid system from the system implementation to the language implementation, the quality of the system itself does not degrade due to the introduction of glue code. To demonstrate the practical applicability of this approach, we present a case study on a COBOL system migration to Java. Using the Neverlang language workbench to create modular and reconfigurable language implementations, both the COBOL interpreter and the application evolve to spread the development effort across several iterations. Through this study, this work presents a viable solution for organizations dealing with the complexity of modernizing legacy software to contemporary technologies. The contributions of this work are (i) a language-oriented, incremental refactoring process for legacy systems, (ii) a concrete application of open language implementations, and (iii) a general template for the implementation of interoperability between languages in hybrid systems.

由于维护和更新过时系统的成本高昂,以及老旧编程语言专家的稀缺,遗留软件给企业带来了严峻的挑战。这一问题已超出商业应用范围,影响到了公共管理,在 COVID-19 大流行期间对 COBOL 程序员的迫切需求就是例证。为此,这项工作引入了一种基于动态语言产品线(动态软件产品线的一个子集)的现代化方法。这种方法利用开放式语言实现和动态生成的微型语言,将传统系统逐步迁移到现代技术。语言可在运行时重新配置,以适应传统代码或现代代码的执行,并在两种语言处理的数据类型之间生成一个兼容层。通过这一过程,传统系统现代化的成本可以分摊到若干次迭代中,因为开发人员可以逐步替换传统代码,让传统代码和现代代码共存,直到完全重构成为可能。在混合系统中,将传统功能和现代功能协同工作的开销从系统实现转移到语言实现,系统本身的质量就不会因为引入胶合代码而下降。为了证明这种方法的实际适用性,我们介绍了一个将 COBOL 系统迁移到 Java 的案例研究。通过使用 Neverlang 语言工作台来创建模块化和可重新配置的语言实现,COBOL 解释器和应用程序都得到了发展,从而将开发工作分散到多个迭代中。通过这项研究,本作品为企业提供了一个可行的解决方案,以应对将传统软件现代化为现代技术的复杂性。这项工作的贡献在于:(i) 为遗留系统提供了一个面向语言的增量重构过程;(ii) 开放语言实现的具体应用;(iii) 在混合系统中实现语言间互操作性的通用模板。
{"title":"Software modernization powered by dynamic language product lines","authors":"","doi":"10.1016/j.jss.2024.112188","DOIUrl":"10.1016/j.jss.2024.112188","url":null,"abstract":"<div><p>Legacy software poses a critical challenge for organizations due to the costs of maintaining and modernizing outdated systems, as well as the scarcity of experts in aging programming languages. The issue extends beyond commercial applications, affecting public administration, as exemplified by the urgent need for <span>COBOL</span> programmers during the COVID-19 pandemic. In response, this work introduces a modernization approach based on dynamic language product lines, a subset of dynamic software product lines. This approach leverages open language implementations and dynamically generated micro-languages for the incremental migration of legacy systems to modern technologies. The language can be reconfigured at runtime to adapt to the execution of either legacy or modern code, and to generate a compatibility layer between the data types handled by the two languages. Through this process, the costs of modernizing legacy systems can be spread across several iterations, as developers can replace legacy code incrementally, with legacy and modern code coexisting until a complete refactoring is possible. By moving the overhead of making legacy and modern features work together in a hybrid system from the system implementation to the language implementation, the quality of the system itself does not degrade due to the introduction of glue code. To demonstrate the practical applicability of this approach, we present a case study on a <span>COBOL</span> system migration to <span>Java</span>. Using the <span>Neverlang</span> language workbench to create modular and reconfigurable language implementations, both the <span>COBOL</span> interpreter and the application evolve to spread the development effort across several iterations. Through this study, this work presents a viable solution for organizations dealing with the complexity of modernizing legacy software to contemporary technologies. The contributions of this work are (i) a language-oriented, incremental refactoring process for legacy systems, (ii) a concrete application of open language implementations, and (iii) a general template for the implementation of interoperability between languages in hybrid systems.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0164121224002322/pdfft?md5=087cacb89bcf9dc4a2cefae984eef08a&pid=1-s2.0-S0164121224002322-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142137250","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research artifacts for human-oriented experiments in software engineering: An ACM badges-driven structure proposal 软件工程中以人为导向的实验研究工件:ACM 徽章驱动结构提案
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-27 DOI: 10.1016/j.jss.2024.112187

Context:

The Open Science (OS) movement promotes the value of making public the research artifacts (datasets, analysis scripts, guidelines, etc.) used during empirical studies. OS is widely known in areas such as Medicine or Biology, where the process of sharing research artifacts is subject to strict protocols. Unfortunately, in Software Engineering (SE), this process is carried out in a non-systematic way, resulting in incomplete or inaccurate material shared by researchers, which hinders the reproducibility and replicability of empirical studies. Nevertheless, in recent years, it seems that the Empirical Software Engineering (ESE) community is embracing some of the proposed OS initiatives, such as the one proposed by the Association for Computing Machinery (ACM), which provides a badge system to evaluate the quality of a research artifact. This badge system has been adopted by several SE conferences as a method of assessing research artifacts.

Aims:

Focusing on human-oriented experiments (HOEs) in SE, whose research artifacts are more complex than those for computational experiments, this work applies Design Science Research (DSR) with a twofold purpose: (i) review the current status of HOEs research artifacts publication through evaluation of this practice in the most relevant ESE journals , and (ii) propose a structured outline for HOEs research artifacts driven by the aforementioned ACM badging policy.

Method:

Regarding the first purpose, we carried out a survey to analyze the current status of the publication of research artifacts considering relevant peer review journals and the quality of 106 research artifacts published in these journals with respect to the ACM badging policy. For the second purpose, an iterative process was carried out to review the content of 106 research artifacts research and their concordance with ACM badges, obtaining a structured scheme for HOEs research artifacts that has been validated through a detailed review of 12 research artifacts obtained from some of those of ACM badges in relevant SE conferences. In addition, we validated the proposal in the research artifacts of 2 of our own experiments.

Results:

Our survey reveals issues such as the 39,70% of journal studies making completely accessible their research artifacts; most of the analyzed research artifacts are incomplete; the most common repositories used in the ESE community to share the research artifacts are GitHub, institutional repositories, and Zenodo. On the other hand, the validated and structured research artifact outline consists of a list of ordered sections containing a set of artifacts, which can be mandatory or not to achieve a particular ACM badge. For its internal validation, several improvement iterations on the first release of the outline have been carried out based on the conference guidelines, the ACM badging policy, and other relevant proposals.

背景:开放科学(Open Science,OS)运动提倡将实证研究中使用的研究人工制品(数据集、分析脚本、指南等)公之于众的价值。OS 在医学或生物学等领域广为人知,在这些领域,共享研究成果的过程需要遵守严格的协议。遗憾的是,在软件工程(SE)领域,这一过程是以非系统化的方式进行的,导致研究人员共享的材料不完整或不准确,从而阻碍了实证研究的再现性和可复制性。尽管如此,近年来,实证软件工程(ESE)界似乎正在接受一些建议的操作系统倡议,例如由美国计算机协会(ACM)提出的倡议,该倡议提供了一个徽章系统来评估研究成果的质量。该徽章系统已被多个 SE 会议采用,作为评估研究成果的一种方法。目的:设计科学研究(DSR)的目的有两个:(i) 通过评估在最相关的ESE期刊上发表HOEs研究成果的做法,回顾HOEs研究成果发表的现状;(ii) 根据上述ACM徽章政策,提出HOEs研究成果的结构化纲要。方法:关于第一个目的,我们进行了一项调查,分析了在相关同行评审期刊上发表研究成果的现状,以及根据 ACM 徽章政策在这些期刊上发表的 106 篇研究成果的质量。第二个目的是对 106 项研究成果的研究内容及其与 ACM 徽章的一致性进行反复审查,通过详细审查从相关 SE 会议上的一些 ACM 徽章研究成果中获得的 12 项研究成果,验证了 HOEs 研究成果的结构化方案。结果:我们的调查发现了一些问题,如39.70%的期刊研究可完全访问其研究成果;大多数分析的研究成果不完整;ESE社区中最常用的研究成果共享库是GitHub、机构资源库和Zenodo。另一方面,经过验证的结构化研究工件大纲由一系列有序的部分组成,其中包含一组工件,这些工件可以是获得特定 ACM 徽章的必备工件,也可以是非必备工件。为了对其进行内部验证,我们根据会议指南、ACM 徽章政策和其他相关建议,对第一版大纲进行了多次改进迭代。结论:尽管 ESE 社区在与操作系统相关的标准化、审查和数字出版方面做出了巨大努力,但研究工件的可用性和完整性仍有待提高。我们关于制定结构化研究成果大纲的建议符合 SE 中 HOE 的要求。不过,我们还需要进一步研究,不仅要对其进行改进和外部验证,还要在研究界推广使用。
{"title":"Research artifacts for human-oriented experiments in software engineering: An ACM badges-driven structure proposal","authors":"","doi":"10.1016/j.jss.2024.112187","DOIUrl":"10.1016/j.jss.2024.112187","url":null,"abstract":"<div><h3>Context:</h3><p>The Open Science (OS) movement promotes the value of making public the research artifacts (datasets, analysis scripts, guidelines, etc.) used during empirical studies. OS is widely known in areas such as Medicine or Biology, where the process of sharing research artifacts is subject to strict protocols. Unfortunately, in Software Engineering (SE), this process is carried out in a non-systematic way, resulting in incomplete or inaccurate material shared by researchers, which hinders the reproducibility and replicability of empirical studies. Nevertheless, in recent years, it seems that the Empirical Software Engineering (ESE) community is embracing some of the proposed OS initiatives, such as the one proposed by the Association for Computing Machinery (ACM), which provides a badge system to evaluate the quality of a research artifact. This badge system has been adopted by several SE conferences as a method of assessing research artifacts.</p></div><div><h3>Aims:</h3><p>Focusing on human-oriented experiments (HOEs) in SE, whose research artifacts are more complex than those for computational experiments, this work applies Design Science Research (DSR) with a twofold purpose: (i) review the current status of HOEs research artifacts publication through evaluation of this practice in the most relevant ESE journals , and (ii) propose a structured outline for HOEs research artifacts driven by the aforementioned ACM badging policy.</p></div><div><h3>Method:</h3><p>Regarding the first purpose, we carried out a survey to analyze the current status of the publication of research artifacts considering relevant peer review journals and the quality of 106 research artifacts published in these journals with respect to the ACM badging policy. For the second purpose, an iterative process was carried out to review the content of 106 research artifacts research and their concordance with ACM badges, obtaining a structured scheme for HOEs research artifacts that has been validated through a detailed review of 12 research artifacts obtained from some of those of ACM badges in relevant SE conferences. In addition, we validated the proposal in the research artifacts of 2 of our own experiments.</p></div><div><h3>Results:</h3><p>Our survey reveals issues such as the 39,70% of journal studies making completely accessible their research artifacts; most of the analyzed research artifacts are incomplete; the most common repositories used in the ESE community to share the research artifacts are GitHub, institutional repositories, and Zenodo. On the other hand, the validated and structured research artifact outline consists of a list of ordered sections containing a set of artifacts, which can be mandatory or not to achieve a particular ACM badge. For its internal validation, several improvement iterations on the first release of the outline have been carried out based on the conference guidelines, the ACM badging policy, and other relevant proposals.","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0164121224002310/pdfft?md5=fb4a53f470e5e9d69349ac3f01883dd0&pid=1-s2.0-S0164121224002310-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GBSR: Graph-based suspiciousness refinement for improving fault localization GBSR:基于图的可疑性细化,用于改进故障定位
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-26 DOI: 10.1016/j.jss.2024.112189

Fault Localization (FL) is an important and time-consuming phase of software debugging. The essence of FL lies in the process of calculating the suspiciousness of different program entities (e.g., statements) and generating a ranking list to guide developers in their code inspection. Nonetheless, a prevalent challenge within existing FL methodologies is the propensity for program entities with analogous execution information to receive a similar suspiciousness. This phenomenon can lead to confusion among developers, thereby reducing the effectiveness of debugging significantly. To alleviate this issue, we introduce fine-grained contextual information (such as partial code structural, coverage, and features from mutation analysis) to enrich the characteristics of program entities. Graphical structures are proposed to organize such information, where the passed and failed tests are constructed separately with the consideration of their differential impacts. In order to support the analysis of multidimensional features and the representation of large-scale programs, the PageRank algorithm is adopted to compute each program entity’s weight. Rather than altering the fundamental FL process, we leverage these computed weights to refine the suspiciousness produced by various FL techniques, thereby providing developers with a more precise and actionable ranking of potential fault locations. The proposed strategy Graph-Based Suspiciousness Refinement (GBSR) is evaluated on 243 real-world faulty programs from the Defects4J. The results demonstrate that GBSR can improve the accuracy of various FL techniques. Specifically, for the refinement with traditional SBFL and MBFL techniques, the number of faults localized by the first position of the ranking list (Top-1) is increased by 189% and 68%, respectively. Furthermore, GBSR can also boost the state-of-the-art learning-based FL technique Grace by achieving a 2.8% performance improvement in Top-1.

故障定位(FL)是软件调试中一个重要而耗时的阶段。故障定位的精髓在于计算不同程序实体(如语句)的可疑程度,并生成一个排名列表,以指导开发人员进行代码检查。然而,现有 FL 方法面临的一个普遍挑战是,具有类似执行信息的程序实体往往具有相似的可疑度。这种现象会导致开发人员产生混淆,从而大大降低调试的有效性。为了缓解这一问题,我们引入了细粒度的上下文信息(如部分代码结构、覆盖率和突变分析的特征)来丰富程序实体的特征。我们提出了图形结构来组织这些信息,其中通过的测试和失败的测试分别构建,并考虑了它们的不同影响。为了支持多维特征的分析和大规模程序的表示,我们采用了 PageRank 算法来计算每个程序实体的权重。我们并不改变基本的 FL 流程,而是利用这些计算出的权重来完善各种 FL 技术产生的可疑度,从而为开发人员提供更精确、更可操作的潜在故障位置排名。我们在 Defects4J 中的 243 个真实世界故障程序上评估了所提出的基于图形的可疑度细化(GBSR)策略。结果表明,GBSR 可以提高各种 FL 技术的准确性。具体来说,在使用传统的 SBFL 和 MBFL 技术进行细化时,由排名表第一位(Top-1)定位的故障数量分别增加了 189% 和 68%。此外,GBSR 还能提高最先进的基于学习的 FL 技术 Grace 的性能,在 Top-1 中提高了 2.8%。
{"title":"GBSR: Graph-based suspiciousness refinement for improving fault localization","authors":"","doi":"10.1016/j.jss.2024.112189","DOIUrl":"10.1016/j.jss.2024.112189","url":null,"abstract":"<div><p>Fault Localization (FL) is an important and time-consuming phase of software debugging. The essence of FL lies in the process of calculating the suspiciousness of different program entities (e.g., statements) and generating a ranking list to guide developers in their code inspection. Nonetheless, a prevalent challenge within existing FL methodologies is the propensity for program entities with analogous execution information to receive a similar suspiciousness. This phenomenon can lead to confusion among developers, thereby reducing the effectiveness of debugging significantly. To alleviate this issue, we introduce fine-grained contextual information (such as partial code structural, coverage, and features from mutation analysis) to enrich the characteristics of program entities. Graphical structures are proposed to organize such information, where the passed and failed tests are constructed separately with the consideration of their differential impacts. In order to support the analysis of multidimensional features and the representation of large-scale programs, the PageRank algorithm is adopted to compute each program entity’s weight. Rather than altering the fundamental FL process, we leverage these computed weights to refine the suspiciousness produced by various FL techniques, thereby providing developers with a more precise and actionable ranking of potential fault locations. The proposed strategy Graph-Based Suspiciousness Refinement (GBSR) is evaluated on 243 real-world faulty programs from the Defects4J. The results demonstrate that GBSR can improve the accuracy of various FL techniques. Specifically, for the refinement with traditional SBFL and MBFL techniques, the number of faults localized by the first position of the ranking list (<span><math><mrow><mi>T</mi><mi>o</mi><mi>p</mi></mrow></math></span>-1) is increased by 189% and 68%, respectively. Furthermore, GBSR can also boost the state-of-the-art learning-based FL technique Grace by achieving a 2.8% performance improvement in <span><math><mrow><mi>T</mi><mi>o</mi><mi>p</mi></mrow></math></span>-1.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142098169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Web application testing—Challenges and opportunities 网络应用程序测试--挑战与机遇
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-17 DOI: 10.1016/j.jss.2024.112186

Context:

A large part of the software produced by many companies and organizations today are web applications. Testing web applications is vital to ensure and maintain the quality of these systems. They play an important role in promoting brands and enabling better communication with customers.

Objective:

There is a gap in existing literature research in recent years regarding testing of web applications, although the landscape of web applications techniques has changed. New methods, frameworks, environments and techniques have recently been used both for developing and testing these applications. This paper presents an overview of the research directions, problems and challenges in the field of Web application testing in the last decade. Our paper investigates current implementation and validation techniques, the quality of existing approaches and reveals areas of incomplete or superficial research.

Methods:

In this paper, a systematic literature review about the state of web application testing has been conducted. Based on initially about 6000 papers, that were extracted from Science Direct, Springer Link, WebOfScience, IEEE Explore and ACM, we used a final number of 72 papers after a filtration process for this literature review. The extracted papers were examined for demographics, problems, techniques and tools. We looked at the quality, the empirical evidence and the test application used for validating the different methods in the extracted papers.

Results:

The most important journals, authors, tools and research directions in this field are discovered, a deep analysis of quality, rigor and empirical evidence is given, and the most important validation applications are described. We found that three groups of authors contributed to more than 25% of the papers, the three most important journals published more than 50% of the papers. 30% of the developed tools were open accessible. Most papers had a good description of study design and threats for validity, but have little industrial relevance. Only 6 papers validated on industrial applications. Only 40% of these papers compared their technique with other existing techniques, and applications used for validation are usually outdated.

Conclusions:

We discuss trends and challenges in research in web application testing. We also show gaps in research and areas that need more attention from the research community. Research in web application testing needs more focus on industrial relevance and scalability to analyze the usability for industry. New techniques should be validated on modern test application frameworks to get comparable results. The results can help researchers to get an overview of publication venues, active researchers, current research gaps and problems in the field.

背景:如今,许多公司和组织生产的软件中,有很大一部分是网络应用程序。测试网络应用程序对于确保和维护这些系统的质量至关重要。目标:尽管网络应用技术的格局已发生变化,但近年来关于网络应用程序测试的现有文献研究仍存在空白。新的方法、框架、环境和技术最近已被用于开发和测试这些应用程序。本文概述了近十年来网络应用程序测试领域的研究方向、问题和挑战。本文研究了当前的实施和验证技术、现有方法的质量,并揭示了研究不完整或肤浅的领域。最初,我们从 Science Direct、Springer Link、WebOfScience、IEEE Explore 和 ACM 中提取了约 6000 篇论文,经过筛选后,我们最终使用了 72 篇论文进行文献综述。我们对提取的论文进行了人口统计学、问题、技术和工具方面的审查。结果:我们发现了该领域最重要的期刊、作者、工具和研究方向,对质量、严谨性和经验证据进行了深入分析,并介绍了最重要的验证应用。我们发现,三组作者发表了 25% 以上的论文,三份最重要的期刊发表了 50% 以上的论文。30%的已开发工具可公开获取。大多数论文对研究设计和有效性威胁进行了很好的描述,但与工业相关性不强。只有 6 篇论文对工业应用进行了验证。结论:我们讨论了网络应用程序测试研究的趋势和挑战。结论:我们讨论了网络应用程序测试研究的趋势和挑战,同时也指出了研究中的不足以及需要研究界给予更多关注的领域。网络应用程序测试研究需要更加关注行业相关性和可扩展性,以便为行业分析可用性。新技术应在现代测试应用框架上进行验证,以获得可比较的结果。研究结果可以帮助研究人员全面了解该领域的发表地点、活跃的研究人员、当前的研究差距和问题。
{"title":"Web application testing—Challenges and opportunities","authors":"","doi":"10.1016/j.jss.2024.112186","DOIUrl":"10.1016/j.jss.2024.112186","url":null,"abstract":"<div><h3>Context:</h3><p>A large part of the software produced by many companies and organizations today are web applications. Testing web applications is vital to ensure and maintain the quality of these systems. They play an important role in promoting brands and enabling better communication with customers.</p></div><div><h3>Objective:</h3><p>There is a gap in existing literature research in recent years regarding testing of web applications, although the landscape of web applications techniques has changed. New methods, frameworks, environments and techniques have recently been used both for developing and testing these applications. This paper presents an overview of the research directions, problems and challenges in the field of Web application testing in the last decade. Our paper investigates current implementation and validation techniques, the quality of existing approaches and reveals areas of incomplete or superficial research.</p></div><div><h3>Methods:</h3><p>In this paper, a systematic literature review about the state of web application testing has been conducted. Based on initially about 6000 papers, that were extracted from Science Direct, Springer Link, WebOfScience, IEEE Explore and ACM, we used a final number of 72 papers after a filtration process for this literature review. The extracted papers were examined for demographics, problems, techniques and tools. We looked at the quality, the empirical evidence and the test application used for validating the different methods in the extracted papers.</p></div><div><h3>Results:</h3><p>The most important journals, authors, tools and research directions in this field are discovered, a deep analysis of quality, rigor and empirical evidence is given, and the most important validation applications are described. We found that three groups of authors contributed to more than 25% of the papers, the three most important journals published more than 50% of the papers. 30% of the developed tools were open accessible. Most papers had a good description of study design and threats for validity, but have little industrial relevance. Only 6 papers validated on industrial applications. Only 40% of these papers compared their technique with other existing techniques, and applications used for validation are usually outdated.</p></div><div><h3>Conclusions:</h3><p>We discuss trends and challenges in research in web application testing. We also show gaps in research and areas that need more attention from the research community. Research in web application testing needs more focus on industrial relevance and scalability to analyze the usability for industry. New techniques should be validated on modern test application frameworks to get comparable results. The results can help researchers to get an overview of publication venues, active researchers, current research gaps and problems in the field.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0164121224002309/pdfft?md5=3ef92544adb97aa90f903887254cfc82&pid=1-s2.0-S0164121224002309-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142167884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards a common data-driven culture: A longitudinal study of the tensions and emerging solutions involved in becoming data-driven in a large public sector organization 建立共同的数据驱动文化:对大型公共部门机构在数据驱动过程中出现的紧张局势和新解决方案的纵向研究
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-17 DOI: 10.1016/j.jss.2024.112185

In recent years, the push to make organizations data-driven has led to data-focused software projects, both in the private and public sectors. The strive for increasing data-driven initiatives introduces a range of new socio-technical challenges, yet there are to date few empirical studies in terms of how data-focused initiatives affect large organizations with significant variations in terms of data needs and usage. This study presents a longitudinal descriptive case study of how data-driven initiatives in the Norwegian public sector cause organizational tensions in a very large, complex organization. We conducted 32 semi-structured interviews over a period of 18 months representing two different data-intensive parts of the organization that had developed incompatible data cultures. Our study shows that these cultural differences create organizational conflicts that hinder data-driven initiatives. The findings also suggest, however, that overcoming these is possible through the strategic, top-down facilitation of a common data-driven culture built on uniting data principles, in turn potentially leading to improved decision-making and enhanced innovation.

近年来,为推动组织数据化,私营和公共部门都开展了以数据为重点的软件项目。数据驱动项目的不断增加带来了一系列新的社会技术挑战,但迄今为止,关于数据驱动项目如何影响在数据需求和使用方面存在显著差异的大型组织的实证研究却寥寥无几。本研究以纵向描述性案例研究的形式,介绍了挪威公共部门的数据驱动计划如何在一个非常庞大、复杂的组织中造成组织紧张局势。在18个月的时间里,我们进行了32次半结构式访谈,这些访谈代表了该组织中两个不同的数据密集型部门,它们已经形成了互不兼容的数据文化。我们的研究表明,这些文化差异造成了组织冲突,阻碍了数据驱动计划的实施。不过,研究结果也表明,通过自上而下的战略推动,在统一数据原则的基础上建立共同的数据驱动文化,是有可能克服这些冲突的,进而有可能改进决策和加强创新。
{"title":"Towards a common data-driven culture: A longitudinal study of the tensions and emerging solutions involved in becoming data-driven in a large public sector organization","authors":"","doi":"10.1016/j.jss.2024.112185","DOIUrl":"10.1016/j.jss.2024.112185","url":null,"abstract":"<div><p>In recent years, the push to make organizations data-driven has led to data-focused software projects, both in the private and public sectors. The strive for increasing data-driven initiatives introduces a range of new socio-technical challenges, yet there are to date few empirical studies in terms of how data-focused initiatives affect large organizations with significant variations in terms of data needs and usage. This study presents a longitudinal descriptive case study of how data-driven initiatives in the Norwegian public sector cause organizational tensions in a very large, complex organization. We conducted 32 semi-structured interviews over a period of 18 months representing two different data-intensive parts of the organization that had developed incompatible data cultures. Our study shows that these cultural differences create organizational conflicts that hinder data-driven initiatives. The findings also suggest, however, that overcoming these is possible through the strategic, top-down facilitation of a common data-driven culture built on uniting data principles, in turn potentially leading to improved decision-making and enhanced innovation.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142098170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive data quality scoring operations framework using drift-aware mechanism for industrial applications 利用漂移感知机制为工业应用提供自适应数据质量评分操作框架
IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-08-14 DOI: 10.1016/j.jss.2024.112184

Within data-driven artificial intelligence (AI) systems for industrial applications, ensuring the reliability of the incoming data streams is an integral part of trustworthy decision-making. An approach to assess data validity is data quality scoring, which assigns a score to each data point or stream based on various quality dimensions. However, certain dimensions exhibit dynamic qualities, which require adaptation on the basis of the system’s current conditions. Existing methods often overlook this aspect, making them inefficient in dynamic production environments. In this paper, we introduce the Adaptive Data Quality Scoring Operations Framework, a novel framework developed to address the challenges posed by dynamic quality dimensions in industrial data streams. The framework introduces an innovative approach by integrating a dynamic change detector mechanism that actively monitors and adapts to changes in data quality, ensuring the relevance of quality scores. We evaluate the proposed framework performance in a real-world industrial use case. The experimental results reveal high predictive performance and efficient processing time, highlighting its effectiveness in practical quality-driven AI applications.

在用于工业应用的数据驱动型人工智能(AI)系统中,确保输入数据流的可靠性是可靠决策不可或缺的一部分。评估数据有效性的一种方法是数据质量评分,它根据不同的质量维度为每个数据点或数据流分配分数。然而,某些维度表现出动态特性,需要根据系统的当前条件进行调整。现有的方法往往忽略了这一点,导致它们在动态生产环境中效率低下。在本文中,我们介绍了自适应数据质量评分操作框架,这是一个为应对工业数据流中动态质量维度带来的挑战而开发的新型框架。该框架通过集成动态变化检测器机制引入了一种创新方法,该机制可主动监测和适应数据质量的变化,确保质量评分的相关性。我们在一个真实的工业应用案例中对所提出的框架性能进行了评估。实验结果表明,该框架具有较高的预测性能和高效的处理时间,突出了其在实际质量驱动型人工智能应用中的有效性。
{"title":"Adaptive data quality scoring operations framework using drift-aware mechanism for industrial applications","authors":"","doi":"10.1016/j.jss.2024.112184","DOIUrl":"10.1016/j.jss.2024.112184","url":null,"abstract":"<div><p>Within data-driven artificial intelligence (AI) systems for industrial applications, ensuring the reliability of the incoming data streams is an integral part of trustworthy decision-making. An approach to assess data validity is data quality scoring, which assigns a score to each data point or stream based on various quality dimensions. However, certain dimensions exhibit dynamic qualities, which require adaptation on the basis of the system’s current conditions. Existing methods often overlook this aspect, making them inefficient in dynamic production environments. In this paper, we introduce the Adaptive Data Quality Scoring Operations Framework, a novel framework developed to address the challenges posed by dynamic quality dimensions in industrial data streams. The framework introduces an innovative approach by integrating a dynamic change detector mechanism that actively monitors and adapts to changes in data quality, ensuring the relevance of quality scores. We evaluate the proposed framework performance in a real-world industrial use case. The experimental results reveal high predictive performance and efficient processing time, highlighting its effectiveness in practical quality-driven AI applications.</p></div>","PeriodicalId":51099,"journal":{"name":"Journal of Systems and Software","volume":null,"pages":null},"PeriodicalIF":3.7,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0164121224002280/pdfft?md5=11e8fc908484f0a491cea864fab6396e&pid=1-s2.0-S0164121224002280-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142002324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Systems and Software
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1