首页 > 最新文献

Automated Software Engineering最新文献

英文 中文
Bug Classification in quantum software: a rule-based framework and its evaluation 量子软件中的Bug分类:基于规则的框架及其评价
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-18 DOI: 10.1007/s10515-025-00585-7
Mir Mohammad Yousuf, Shabir Ahmad Sofi

Accurate bug classification is essential for improving software quality, particularly in the emerging and complex domain of quantum computing. This paper introduces a rule-based framework for automated classification of quantum software issues across five dimensions: bug type, bug category, severity, quality attribute, and quantum-specific subtype. The framework integrates weighted keyword heuristics, TF–IDF similarity, and domain-specific rules to capture both general software defects and quantum-domain failure modes. The proposed approach was applied to 12,910 issues from 36 Qiskit repositories and validated on a stratified subset of 4,984 manually annotated issues. On this manually labeled subset, the framework achieved accuracies between 0.82 and 0.85 and macro-F1 scores ranging from 0.68 to 0.77, demonstrating strong agreement with human annotations without requiring supervised training. When compared with standard machine-learning baselines (Logistic Regression, Decision Tree, Random Forest, Gradient Boosting), the rule-based approach consistently outperformed all models across tasks, showing particularly large improvements in fine-grained and low-frequency categories such as Category (macro-F1: 0.26 vs. 0.69) and Quantum-Specific Subtype (0.15 vs. 0.77). Beyond predictive accuracy, the framework was applied to real-world 12,910 issues from 36 Qiskit repositories for large-scale distributional analysis. The results revealed that approximately 67% of issues were classical and 27% quantum-specific, with interoperability, usability, and maintainability identified as the most frequently affected quality attributes. Low-severity issues dominated (68.8%), while critical bugs accounted for around 21%. Quantum-specific defects were most prevalent at the circuit and gate abstraction levels, reflecting the hybrid and hardware-constrained nature of current quantum software development. Overall, the proposed rule-based framework offers a transparent, interpretable, and empirically validated approach for automated bug triaging in quantum software. Beyond its immediate practical utility, it provides a reproducible methodological framework that can support future hybrid and learning-based advances in quantum software engineering.

准确的错误分类对于提高软件质量至关重要,特别是在新兴和复杂的量子计算领域。本文介绍了一个基于规则的量子软件问题自动分类框架,该框架跨越五个维度:缺陷类型、缺陷类别、严重性、质量属性和量子特定子类型。该框架集成了加权关键字启发式、TF-IDF相似性和领域特定规则,以捕获一般软件缺陷和量子域故障模式。提出的方法应用于来自36个Qiskit存储库的12,910个问题,并在4,984个手动注释问题的分层子集上进行验证。在这个手动标记的子集上,该框架的准确率在0.82到0.85之间,macro-F1得分在0.68到0.77之间,显示出与人类注释的高度一致,而不需要监督训练。与标准机器学习基线(逻辑回归、决策树、随机森林、梯度增强)相比,基于规则的方法在任务中始终优于所有模型,在细粒度和低频类别(如类别(macro-F1: 0.26 vs. 0.69)和量子特定亚型(0.15 vs. 0.77)中表现出特别大的改进。除了预测准确性之外,该框架还应用于来自36个Qiskit存储库的12,910个实际问题,以进行大规模分布分析。结果显示,大约67%的问题是经典问题,27%是量子问题,互操作性、可用性和可维护性被确定为最常受影响的质量属性。低严重性问题占主导地位(68.8%),而严重错误约占21%。量子特定缺陷在电路和门抽象层次上最为普遍,反映了当前量子软件开发的混合和硬件约束性质。总的来说,提出的基于规则的框架为量子软件中的自动错误分类提供了透明、可解释和经验验证的方法。除了它的直接实用之外,它还提供了一个可重复的方法框架,可以支持未来量子软件工程中基于混合和学习的进步。
{"title":"Bug Classification in quantum software: a rule-based framework and its evaluation","authors":"Mir Mohammad Yousuf,&nbsp;Shabir Ahmad Sofi","doi":"10.1007/s10515-025-00585-7","DOIUrl":"10.1007/s10515-025-00585-7","url":null,"abstract":"<div><p>Accurate bug classification is essential for improving software quality, particularly in the emerging and complex domain of quantum computing. This paper introduces a rule-based framework for automated classification of quantum software issues across five dimensions: bug type, bug category, severity, quality attribute, and quantum-specific subtype. The framework integrates weighted keyword heuristics, TF–IDF similarity, and domain-specific rules to capture both general software defects and quantum-domain failure modes. The proposed approach was applied to 12,910 issues from 36 Qiskit repositories and validated on a stratified subset of 4,984 manually annotated issues. On this manually labeled subset, the framework achieved accuracies between 0.82 and 0.85 and macro-F1 scores ranging from 0.68 to 0.77, demonstrating strong agreement with human annotations without requiring supervised training. When compared with standard machine-learning baselines (Logistic Regression, Decision Tree, Random Forest, Gradient Boosting), the rule-based approach consistently outperformed all models across tasks, showing particularly large improvements in fine-grained and low-frequency categories such as Category (macro-F1: 0.26 vs. 0.69) and Quantum-Specific Subtype (0.15 vs. 0.77). Beyond predictive accuracy, the framework was applied to real-world 12,910 issues from 36 Qiskit repositories for large-scale distributional analysis. The results revealed that approximately 67% of issues were classical and 27% quantum-specific, with interoperability, usability, and maintainability identified as the most frequently affected quality attributes. Low-severity issues dominated (68.8%), while critical bugs accounted for around 21%. Quantum-specific defects were most prevalent at the circuit and gate abstraction levels, reflecting the hybrid and hardware-constrained nature of current quantum software development. Overall, the proposed rule-based framework offers a transparent, interpretable, and empirically validated approach for automated bug triaging in quantum software. Beyond its immediate practical utility, it provides a reproducible methodological framework that can support future hybrid and learning-based advances in quantum software engineering.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778678","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic assistance to mitigate rollback inconsistencies in collaborative edits 自动协助减轻协作编辑中的回滚不一致
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-13 DOI: 10.1007/s10515-025-00583-9
Saikat Mondal, Gias Uddin, Chanchal K. Roy

The success of technical Q&A sites such as Stack Overflow depends on two key factors: (a) active user participation and (b) the quality of the shared knowledge. Stack Overflow introduced an edit system that allows users to suggest improvements to posts (i.e., questions and answers) to enhance the quality of the content. However, users, such as post owners or site moderators, can reject these suggested edits by rollbacks due to unsatisfactory or low-quality edits, or because they violate edit guidelines. Unfortunately, subjectivity bias in determining whether an edit is satisfactory or unsatisfactory can lead to inconsistencies in the rollback decisions. For example, one user might accept the formatting of a method name (e.g., getActivity()) as a code term, while another might reject it. Such inconsistencies can demotivate and frustrate users whose edits are rejected. Furthermore, several post owners prefer to keep their content unchanged and even resist necessary edits. As a result, they sometimes roll back necessary edits and revert posts to a flawed version, which violates editing guidelines. The problems mentioned above are further compounded by the lack of specific guidelines and tools to assist users in ensuring consistency in rollback actions. In this study, we investigate the types, prevalence, and impact of rollback edit inconsistencies and propose a solution to address them. The outcomes of this research are fivefold. First, we manually investigated 764 rollback edits (382 questions + 382 answers) and identified eight types of inconsistent rollback. Second, we surveyed 44 practitioners to assess the impact of rollback inconsistencies. More than 80% of the participants found our identified inconsistency types detrimental to post quality. Third, we developed rule-based algorithms and machine learning (ML) models to detect the eight types of rollback inconsistencies. Both approaches achieve over 90% accuracy. Fourth, we introduced a tool, iEdit, which integrates these algorithms into a browser extension and assists Stack Overflow users during their edits. Fifth, we surveyed 16 Stack Overflow users to evaluate the effectiveness of iEdit. The participants found the tool’s suggestions helpful in avoiding inconsistent rollback edits.

像Stack Overflow这样的技术问答网站的成功取决于两个关键因素:(A)用户的积极参与和(b)共享知识的质量。Stack Overflow引入了一个编辑系统,允许用户对帖子提出改进建议(即问题和答案),以提高内容的质量。但是,用户(如帖子所有者或网站版主)可以由于不满意或低质量的编辑或违反编辑准则而通过回滚拒绝这些建议的编辑。不幸的是,在确定编辑是否令人满意时的主观性偏差可能导致回滚决策中的不一致。例如,一个用户可能接受方法名的格式(例如,getActivity())作为代码项,而另一个用户可能拒绝它。这种不一致会使编辑被拒绝的用户失去动力和沮丧。此外,一些帖子所有者更喜欢保持他们的内容不变,甚至拒绝必要的编辑。因此,他们有时回滚必要的编辑,并将帖子恢复到有缺陷的版本,这违反了编辑准则。由于缺乏具体的指导方针和工具来帮助用户确保回滚操作的一致性,上述问题进一步复杂化。在本研究中,我们调查回滚编辑不一致的类型、流行程度和影响,并提出解决这些问题的解决方案。这项研究的结果是五倍的。首先,我们手动调查了764个回滚编辑(382个问题+ 382个答案),并确定了8种不一致的回滚类型。其次,我们调查了44位从业者,以评估回滚不一致性的影响。超过80%的参与者认为我们确定的不一致类型不利于帖子质量。第三,我们开发了基于规则的算法和机器学习(ML)模型来检测八种类型的回滚不一致。两种方法的准确率都超过90%。第四,我们引入了一个工具,iEdit,它将这些算法集成到浏览器扩展中,并在Stack Overflow用户编辑时提供帮助。第五,我们调查了16个Stack Overflow用户来评估iEdit的有效性。参与者发现该工具的建议有助于避免不一致的回滚编辑。
{"title":"Automatic assistance to mitigate rollback inconsistencies in collaborative edits","authors":"Saikat Mondal,&nbsp;Gias Uddin,&nbsp;Chanchal K. Roy","doi":"10.1007/s10515-025-00583-9","DOIUrl":"10.1007/s10515-025-00583-9","url":null,"abstract":"<div><p>The success of technical Q&amp;A sites such as Stack Overflow depends on two key factors: (a) active user participation and (b) the quality of the shared knowledge. Stack Overflow introduced an edit system that allows users to suggest improvements to posts (i.e., questions and answers) to enhance the quality of the content. However, users, such as post owners or site moderators, can reject these suggested edits by rollbacks due to unsatisfactory or low-quality edits, or because they violate edit guidelines. Unfortunately, subjectivity bias in determining whether an edit is satisfactory or unsatisfactory can lead to inconsistencies in the rollback decisions. For example, one user might accept the formatting of a method name (e.g., <span>getActivity()</span>) as a code term, while another might reject it. Such inconsistencies can demotivate and frustrate users whose edits are rejected. Furthermore, several post owners prefer to keep their content unchanged and even resist necessary edits. As a result, they sometimes roll back necessary edits and revert posts to a flawed version, which violates editing guidelines. The problems mentioned above are further compounded by the lack of specific guidelines and tools to assist users in ensuring consistency in rollback actions. In this study, we investigate the types, prevalence, and impact of rollback edit inconsistencies and propose a solution to address them. The outcomes of this research are fivefold. First, we manually investigated 764 rollback edits (382 questions + 382 answers) and identified eight types of inconsistent rollback. Second, we surveyed 44 practitioners to assess the impact of rollback inconsistencies. More than 80% of the participants found our identified inconsistency types detrimental to post quality. Third, we developed rule-based algorithms and machine learning (ML) models to detect the eight types of rollback inconsistencies. Both approaches achieve over 90% accuracy. Fourth, we introduced a tool, <span>iEdit</span>, which integrates these algorithms into a browser extension and assists Stack Overflow users during their edits. Fifth, we surveyed 16 Stack Overflow users to evaluate the effectiveness of <span>iEdit</span>. The participants found the tool’s suggestions helpful in avoiding inconsistent rollback edits.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145778774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing mutation testing for deep neural networks: a novel approach to generating high-quality mutants 增强深度神经网络的突变检测:一种生成高质量突变体的新方法
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-08 DOI: 10.1007/s10515-025-00581-x
Yu Xie, Zhiyi Zhang, Yongming Yao, Chen Liu, Wenting Chen, Zhiqiu Huang

In deep neural network (DNN) testing, high-quality test cases are essential to ensure the effectiveness and completeness of the testing. Mutation testing has been used in DNN testing to evaluate the quality of test cases. However, existing methods don’t consider the importance of different neurons; they randomly select neurons for mutation, resulting in low-quality mutants. To address this issue, we propose a novel method for generating high-quality mutants tailored for DNNs. Our approach first analyzes the output distribution of neurons and calculates their activation thresholds. Depending on these thresholds, we identify the behavior of neurons and select the most frequently activated neurons for each category as category-critical neurons. Next, we mutate these category-critical neurons to produce a candidate set of high-quality mutants. Although mutating category-critical neurons could reveal more errors, some mutants can cause significant changes in DNNs, making them easily killed and reducing the effectiveness of mutation testing. Thus, we further optimize the candidate set to get high-quality mutants. We calculate each mutant’s Category Above Ratio (CAR) score according to its execution behavior. Lastly, mutants whose CAR score is greater than or equal to the CAR threshold will be selected as higher-quality mutants. We also validate our method on several popular datasets and models. Experimental results demonstrate that our approach generates higher-quality mutants than random methods, reducing the number of mutants by 79.7% to 96.3%, while maintaining the effectiveness of mutation testing. Furthermore, retraining models with test cases generated by high-quality mutants could improve the robustness of the models.

在深度神经网络(DNN)测试中,高质量的测试用例是保证测试有效性和完整性的关键。突变测试已被用于DNN测试,以评估测试用例的质量。然而,现有的方法没有考虑到不同神经元的重要性;他们随机选择神经元进行突变,从而产生低质量的突变体。为了解决这个问题,我们提出了一种新的方法来生成高质量的dnn突变体。我们的方法首先分析神经元的输出分布并计算它们的激活阈值。根据这些阈值,我们识别神经元的行为,并为每个类别选择最频繁激活的神经元作为类别关键神经元。接下来,我们对这些类别关键神经元进行突变,以产生一组候选的高质量突变体。虽然突变类别关键神经元可以揭示更多的错误,但一些突变可以引起dnn的显著变化,使它们容易被杀死,降低突变测试的有效性。因此,我们进一步优化候选集以获得高质量的突变体。我们根据每个突变体的执行行为计算其类别高于比率(CAR)得分。最后,CAR分数大于或等于CAR阈值的突变体将被选择为高质量突变体。我们还在几个流行的数据集和模型上验证了我们的方法。实验结果表明,与随机方法相比,该方法产生的突变体质量更高,在保持突变检测有效性的同时,突变体数量减少了79.7% ~ 96.3%。此外,使用由高质量突变体生成的测试用例对模型进行再训练可以提高模型的鲁棒性。
{"title":"Enhancing mutation testing for deep neural networks: a novel approach to generating high-quality mutants","authors":"Yu Xie,&nbsp;Zhiyi Zhang,&nbsp;Yongming Yao,&nbsp;Chen Liu,&nbsp;Wenting Chen,&nbsp;Zhiqiu Huang","doi":"10.1007/s10515-025-00581-x","DOIUrl":"10.1007/s10515-025-00581-x","url":null,"abstract":"<div><p>In deep neural network (DNN) testing, high-quality test cases are essential to ensure the effectiveness and completeness of the testing. Mutation testing has been used in DNN testing to evaluate the quality of test cases. However, existing methods don’t consider the importance of different neurons; they randomly select neurons for mutation, resulting in low-quality mutants. To address this issue, we propose a novel method for generating high-quality mutants tailored for DNNs. Our approach first analyzes the output distribution of neurons and calculates their activation thresholds. Depending on these thresholds, we identify the behavior of neurons and select the most frequently activated neurons for each category as category-critical neurons. Next, we mutate these category-critical neurons to produce a candidate set of high-quality mutants. Although mutating category-critical neurons could reveal more errors, some mutants can cause significant changes in DNNs, making them easily killed and reducing the effectiveness of mutation testing. Thus, we further optimize the candidate set to get high-quality mutants. We calculate each mutant’s Category Above Ratio (CAR) score according to its execution behavior. Lastly, mutants whose CAR score is greater than or equal to the CAR threshold will be selected as higher-quality mutants. We also validate our method on several popular datasets and models. Experimental results demonstrate that our approach generates higher-quality mutants than random methods, reducing the number of mutants by 79.7% to 96.3%, while maintaining the effectiveness of mutation testing. Furthermore, retraining models with test cases generated by high-quality mutants could improve the robustness of the models.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145729588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Domain-constrained synthesis of inconsistent key aspects in textual vulnerability descriptions 文本漏洞描述中不一致关键方面的领域约束综合
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-27 DOI: 10.1007/s10515-025-00582-w
Linyi Han, Shidong Pan, Zhenchang Xing, Sofonias Yitagesu, Xiaowang Zhang, Zhiyong Feng, Jiamou Sun, Qing Huang

Textual Vulnerability Descriptions (TVDs) are crucial for security analysts to understand and address software vulnerabilities. However, the key aspect inconsistencies in TVDs from different repositories pose challenges for achieving a comprehensive understanding of vulnerabilities. Existing approaches aim to mitigate inconsistencies by aligning TVDs with external knowledge bases, but they often discard valuable information and fail to synthesize comprehensive representations. In this paper, we propose a domain-constrained LLM-based synthesis framework for unifying key aspects of TVDs. Our framework consists of three stages: 1) Extraction, guided by rule-based templates to ensure all critical details are captured; 2) Self-evaluation, using domain-specific anchor words to assess semantic variability across sources; and 3) Fusion, leveraging information entropy to reconcile inconsistencies and prioritize relevant details. This framework improves synthesis performance, increasing the F1 score for key aspect augmentation from 0.82 to 0.87, while enhancing comprehension and efficiency by over 30%. We further develop Digest Labels, a practical tool for visualizing TVDs, which human evaluations show significantly boosts usability.

文本漏洞描述(tvd)对于安全分析人员理解和处理软件漏洞至关重要。然而,来自不同存储库的tvd的关键方面的不一致性给实现对漏洞的全面理解带来了挑战。现有的方法旨在通过将tvd与外部知识库对齐来减轻不一致性,但它们经常丢弃有价值的信息,并且无法合成全面的表示。在本文中,我们提出了一个基于领域约束的基于llm的综合框架来统一tvd的关键方面。我们的框架包括三个阶段:1)提取,以基于规则的模板为指导,以确保捕获所有关键细节;2)自我评价,使用特定领域的锚词评估不同来源的语义可变性;3)融合,利用信息熵来调和不一致并优先考虑相关细节。该框架提高了合成性能,将关键方面增强的F1分数从0.82提高到0.87,同时提高了30%以上的理解和效率。我们进一步开发了Digest Labels,这是一个可视化tvd的实用工具,人类评估表明它显著提高了可用性。
{"title":"Domain-constrained synthesis of inconsistent key aspects in textual vulnerability descriptions","authors":"Linyi Han,&nbsp;Shidong Pan,&nbsp;Zhenchang Xing,&nbsp;Sofonias Yitagesu,&nbsp;Xiaowang Zhang,&nbsp;Zhiyong Feng,&nbsp;Jiamou Sun,&nbsp;Qing Huang","doi":"10.1007/s10515-025-00582-w","DOIUrl":"10.1007/s10515-025-00582-w","url":null,"abstract":"<div><p>Textual Vulnerability Descriptions (TVDs) are crucial for security analysts to understand and address software vulnerabilities. However, the key aspect inconsistencies in TVDs from different repositories pose challenges for achieving a comprehensive understanding of vulnerabilities. Existing approaches aim to mitigate inconsistencies by aligning TVDs with external knowledge bases, but they often discard valuable information and fail to synthesize comprehensive representations. In this paper, we propose a domain-constrained LLM-based synthesis framework for unifying key aspects of TVDs. Our framework consists of three stages: 1) Extraction, guided by rule-based templates to ensure all critical details are captured; 2) Self-evaluation, using domain-specific anchor words to assess semantic variability across sources; and 3) Fusion, leveraging information entropy to reconcile inconsistencies and prioritize relevant details. This framework improves synthesis performance, increasing the F1 score for key aspect augmentation from 0.82 to 0.87, while enhancing comprehension and efficiency by over 30%. We further develop Digest Labels, a practical tool for visualizing TVDs, which human evaluations show significantly boosts usability.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145612511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Correction: Automating software size measurement from python code using language models 更正:使用语言模型从python代码自动化软件大小测量
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-21 DOI: 10.1007/s10515-025-00576-8
Samet Tenekeci, Hüseyin Ünlü, Bedir Arda Gül, Damla Keleş, Murat Küçük, Onur Demirörs
{"title":"Correction: Automating software size measurement from python code using language models","authors":"Samet Tenekeci,&nbsp;Hüseyin Ünlü,&nbsp;Bedir Arda Gül,&nbsp;Damla Keleş,&nbsp;Murat Küçük,&nbsp;Onur Demirörs","doi":"10.1007/s10515-025-00576-8","DOIUrl":"10.1007/s10515-025-00576-8","url":null,"abstract":"","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145561610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automating transparent learner profiling through explainable AI 通过可解释的人工智能自动化透明的学习者分析
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-19 DOI: 10.1007/s10515-025-00574-w
Abdelkader Ouared, Madeth May, Claudine Piau-Toffolon, Nicolas Dugué

With the continuous development of digital learning environments, the demand for providing transparency and explanations about learner profiles (LPs) in the digital era is constantly increasing. Digital learning traces play a fundamental role in understanding LP categories and justifying the reasoning behind their behavior, such as learners in difficulty, active/inactive learners, those in progress, and success-oriented vs. at-risk learners. However, effectively identifying LPs from the massive data generated by these environments is challenging due to its vulnerability to misunderstandings and biases. Moreover, stakeholder may not know what queries to ask or what patterns to look for in the data, while traditional Learning Analytics tools struggle to bridge the gap between granular digital traces and human-interpretable conceptual insights. To address this gap, we propose XAI-Profile, a framework that leverages Explainable AI (XAI) to transform raw learner data into interpretable, actionable profiles. XAI-Profile combines explainable machine learning models with interactive visual workflows to bridge between low-level data traces and high-level pedagogical insights. The framework is built on three pillars: (1) Goal-Oriented Requirements for capturing stakeholders’ needs by refining a problem into sub-problems using goal-oriented AND/OR refinement, and mapping them to the target LP with their data indicators. (2) Visual Analytics Design for mapping LP categories by automatically correlating them with trust objects (e.g., rule-based models, frequent sequences in the data), visualization goals, and analysis types. (3) Interaction and Actionability, a guided process enabling stakeholders to discover interesting and unexpected patterns, validate hypotheses, and drill into context-specific explanations. A case study using data from the écri+ project demonstrates how stakeholders can quickly identify high-level behavioral patterns and drill down into specific profiles of interest. The results highlight how XAI-Profile advances learning analytics practices by making AI-driven insights both trustworthy and actionable fostering a human-centered approach to learner analytics.

随着数字化学习环境的不断发展,数字化时代对学习者档案(lp)提供透明度和解释的需求不断增加。数字学习痕迹在理解LP类别和证明其行为背后的原因方面发挥着基本作用,例如困难学习者、活跃/不活跃学习者、正在进步的学习者、以成功为导向的学习者和有风险的学习者。然而,由于容易受到误解和偏见的影响,从这些环境产生的大量数据中有效识别有限合伙人是一项挑战。此外,利益相关者可能不知道该问什么问题或在数据中寻找什么模式,而传统的学习分析工具难以弥合颗粒数字痕迹和人类可解释的概念见解之间的差距。为了解决这一差距,我们提出了XAI- profile,这是一个利用可解释人工智能(XAI)将原始学习者数据转换为可解释、可操作的配置文件的框架。XAI-Profile将可解释的机器学习模型与交互式可视化工作流相结合,在低级数据跟踪和高级教学见解之间架起了桥梁。该框架建立在三个支柱上:(1)面向目标的需求,通过使用面向目标的AND/OR优化将问题细化为子问题,并将其映射到目标LP及其数据指标,从而捕获利益相关者的需求。(2)可视化分析设计通过将LP类别与信任对象(例如,基于规则的模型,数据中的频繁序列),可视化目标和分析类型自动关联来映射它们。(3)交互和可操作性,一个引导过程,使利益相关者能够发现有趣的和意想不到的模式,验证假设,并深入到特定于上下文的解释。使用来自于该项目数据的案例研究演示了涉众如何快速识别高级行为模式并深入到感兴趣的特定概要文件。结果突出了XAI-Profile如何通过使人工智能驱动的见解既可信又可操作,从而促进以人为中心的学习者分析方法,从而推进学习分析实践。
{"title":"Automating transparent learner profiling through explainable AI","authors":"Abdelkader Ouared,&nbsp;Madeth May,&nbsp;Claudine Piau-Toffolon,&nbsp;Nicolas Dugué","doi":"10.1007/s10515-025-00574-w","DOIUrl":"10.1007/s10515-025-00574-w","url":null,"abstract":"<div><p>With the continuous development of digital learning environments, the demand for providing transparency and explanations about learner profiles (<i>LP</i>s) in the digital era is constantly increasing. Digital learning traces play a fundamental role in understanding <i>LP</i> categories and justifying the reasoning behind their behavior, such as learners <i>in difficulty</i>, <i>active/inactive</i> learners, <i>those in progress</i>, and <i>success-oriented vs. at-risk learners</i>. However, effectively identifying <i>LP</i>s from the massive data generated by these environments is challenging due to its vulnerability to misunderstandings and biases. Moreover, stakeholder may not know what queries to ask or what patterns to look for in the data, while traditional Learning Analytics tools struggle to bridge the gap between granular digital traces and human-interpretable conceptual insights. To address this gap, we propose XAI-Profile, a framework that leverages Explainable AI (XAI) to transform raw learner data into interpretable, actionable profiles. XAI-Profile combines explainable machine learning models with interactive visual workflows to bridge between low-level data traces and high-level pedagogical insights. The framework is built on three pillars: <b>(1)</b> <i>Goal-Oriented Requirements</i> for capturing stakeholders’ needs by refining a problem into sub-problems using goal-oriented AND/OR refinement, and mapping them to the target <i>LP</i> with their data indicators. <b>(2)</b> <i>Visual Analytics Design</i> for mapping <i>LP</i> categories by automatically correlating them with trust objects (e.g., rule-based models, frequent sequences in the data), visualization goals, and analysis types. <b>(3)</b> <i>Interaction and Actionability</i>, a guided process enabling stakeholders to discover interesting and unexpected patterns, validate hypotheses, and drill into context-specific explanations. A case study using data from the <span>écri+</span> project demonstrates how stakeholders can quickly identify high-level behavioral patterns and drill down into specific profiles of interest. The results highlight how XAI-Profile advances learning analytics practices by making AI-driven insights both trustworthy and actionable fostering a human-centered approach to learner analytics.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145561178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting software refactoring opportunities using two-level TB-Stacking ensemble models 使用两级TB-Stacking集成模型预测软件重构机会
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-17 DOI: 10.1007/s10515-025-00577-7
Abdulmajeed Alameer, Amal Alazba

Software refactoring is essential for maintaining the quality and extendibility of source code. Although prior research achieved high accuracy in predicting software refactoring opportunities using traditional machine learning methods, further refinements can still boost predictive reliability across multiple refactoring operations. The main objective of this study is to develop a more accurate model for software refactoring predictions. The main objective of this study is to develop a more accurate model for software refactoring predictions. We propose a two-level Tree-based Stacking (TB-Stacking) Ensemble designed to predict refactoring opportunities at the class, method, and variable levels. Various stacking ensembles were built to identify the optimal configurations of base and meta-models. Additionally, two feature selection techniques were employed to identify the most relevant features for accurate predictions. We evaluated our model using a publicly available dataset comprising over two million refactoring instances. The performance of nine traditional machine learning models and seven tree-based ensembles was also assessed, with statistical comparisons drawn against our proposed TB-Stacking model. Results indicate that the TB-Stacking Ensemble consistently outperformed traditional models in all refactoring operations, and demonstrated competitive or superior performance compared to tree-based ensembles. It demonstrated significant improvements in prediction capabilities highlighted by statistical analysis. The model’s robust performance across various refactoring tasks establishes a new benchmark for future refactoring tools.

软件重构对于维护源代码的质量和可扩展性至关重要。尽管先前的研究在使用传统机器学习方法预测软件重构机会方面取得了很高的准确性,但进一步的改进仍然可以提高跨多个重构操作的预测可靠性。本研究的主要目的是为软件重构预测开发一个更准确的模型。本研究的主要目的是为软件重构预测开发一个更准确的模型。我们提出了一个两层基于树的堆栈(TB-Stacking)集成,旨在预测类、方法和变量级别的重构机会。建立了不同的叠加集成,以确定基本模型和元模型的最优配置。此外,采用两种特征选择技术来识别最相关的特征,以进行准确的预测。我们使用包含超过200万个重构实例的公开数据集来评估我们的模型。我们还评估了9种传统机器学习模型和7种基于树的集成模型的性能,并与我们提出的TB-Stacking模型进行了统计比较。结果表明,TB-Stacking集成在所有重构操作中始终优于传统模型,并且与基于树的集成相比表现出竞争力或更高的性能。通过统计分析,它显示了预测能力的显著提高。该模型在各种重构任务中的健壮性能为未来的重构工具建立了一个新的基准。
{"title":"Predicting software refactoring opportunities using two-level TB-Stacking ensemble models","authors":"Abdulmajeed Alameer,&nbsp;Amal Alazba","doi":"10.1007/s10515-025-00577-7","DOIUrl":"10.1007/s10515-025-00577-7","url":null,"abstract":"<div><p>Software refactoring is essential for maintaining the quality and extendibility of source code. Although prior research achieved high accuracy in predicting software refactoring opportunities using traditional machine learning methods, further refinements can still boost predictive reliability across multiple refactoring operations. The main objective of this study is to develop a more accurate model for software refactoring predictions. The main objective of this study is to develop a more accurate model for software refactoring predictions. We propose a two-level Tree-based Stacking (TB-Stacking) Ensemble designed to predict refactoring opportunities at the class, method, and variable levels. Various stacking ensembles were built to identify the optimal configurations of base and meta-models. Additionally, two feature selection techniques were employed to identify the most relevant features for accurate predictions. We evaluated our model using a publicly available dataset comprising over two million refactoring instances. The performance of nine traditional machine learning models and seven tree-based ensembles was also assessed, with statistical comparisons drawn against our proposed TB-Stacking model. Results indicate that the TB-Stacking Ensemble consistently outperformed traditional models in all refactoring operations, and demonstrated competitive or superior performance compared to tree-based ensembles. It demonstrated significant improvements in prediction capabilities highlighted by statistical analysis. The model’s robust performance across various refactoring tasks establishes a new benchmark for future refactoring tools.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145561089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial intelligence and machine learning in enhancing software project management processes: A systematic literature review 人工智能和机器学习在增强软件项目管理过程中的应用:系统的文献综述
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-11 DOI: 10.1007/s10515-025-00578-6
Usama Ali, Mehwish Naseer

The growing complexity of software systems and the increasing demand for global software development necessitate innovative approaches to enhance project management, quality assurance, and risk mitigation. In this study, Artificial Intelligence (AI) and Machine Learning (ML) techniques are examined for overcoming problems in software project management, notably effort estimation, scheduling, resource allocation, risk management, and defect prediction. By systematically reviewing the literature, we show that AI/ML models like Support Vector Machines, neural networks, and ensemble learning can enhance estimation accuracy, maximize resource utilization, and reduce risks. Furthermore, the practical benefits and challenges of implementing an AI/ML system into a real-world system are discussed using real-world case studies, which include data quality and integration issues, and the interpretability of the model. In addition, advanced models, such as graph convolutional networks and deep neural networks, hold great promise as a defect prediction and bug severity classifier. The focus of this research is to leverage the transformative capabilities of AI/ML toward defect-free, efficient, and customer-centric software development. Finally, it suggests future research interests, including integrating explanation model AI, managing data in a better way, and implementing the scalable hybrid approach to meet the newer needs of the industry.

软件系统的日益复杂和对全球软件开发的日益增长的需求需要创新的方法来加强项目管理、质量保证和风险降低。在这项研究中,人工智能(AI)和机器学习(ML)技术被用于克服软件项目管理中的问题,特别是工作量估计、调度、资源分配、风险管理和缺陷预测。通过系统地回顾文献,我们发现支持向量机、神经网络和集成学习等AI/ML模型可以提高估计精度、最大化资源利用率并降低风险。此外,使用现实世界的案例研究讨论了将AI/ML系统实现到现实世界系统中的实际好处和挑战,其中包括数据质量和集成问题,以及模型的可解释性。此外,高级模型,如图卷积网络和深度神经网络,作为缺陷预测和错误严重性分类器具有很大的前景。这项研究的重点是利用AI/ML的变革能力来实现无缺陷、高效和以客户为中心的软件开发。最后,提出了未来的研究方向,包括整合解释模型AI,更好地管理数据,以及实现可扩展的混合方法以满足行业的新需求。
{"title":"Artificial intelligence and machine learning in enhancing software project management processes: A systematic literature review","authors":"Usama Ali,&nbsp;Mehwish Naseer","doi":"10.1007/s10515-025-00578-6","DOIUrl":"10.1007/s10515-025-00578-6","url":null,"abstract":"<div><p>The growing complexity of software systems and the increasing demand for global software development necessitate innovative approaches to enhance project management, quality assurance, and risk mitigation. In this study, Artificial Intelligence (AI) and Machine Learning (ML) techniques are examined for overcoming problems in software project management, notably effort estimation, scheduling, resource allocation, risk management, and defect prediction. By systematically reviewing the literature, we show that AI/ML models like Support Vector Machines, neural networks, and ensemble learning can enhance estimation accuracy, maximize resource utilization, and reduce risks. Furthermore, the practical benefits and challenges of implementing an AI/ML system into a real-world system are discussed using real-world case studies, which include data quality and integration issues, and the interpretability of the model. In addition, advanced models, such as graph convolutional networks and deep neural networks, hold great promise as a defect prediction and bug severity classifier. The focus of this research is to leverage the transformative capabilities of AI/ML toward defect-free, efficient, and customer-centric software development. Finally, it suggests future research interests, including integrating explanation model AI, managing data in a better way, and implementing the scalable hybrid approach to meet the newer needs of the industry.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145510717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The arts and crafts of android adware across a decade 十年来安卓广告软件的艺术和工艺
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-11 DOI: 10.1007/s10515-025-00575-9
Chao Wang, Tianming Liu, Yanjie Zhao, Lin Zhang, Xiaoning Du, Li Li, Haoyu Wang

Adware represents a pervasive threat in the mobile ecosystem, yet its inherent characteristics have been largely overlooked by previous research. This work takes a crucial step towards demystifying Android adware. We present AdwareZoo, a comprehensive dataset comprising 15,996 adware samples across 118 distinct families collected from security reports and app repositories. We identify adware family payloads by isolating packages from samples for VirusTotal rescanning, unveiling distinctive patterns in family naming conventions, and exposing the misclassification of legitimate ad networks as adware. Our analysis of payload location strategies reveals that over 30% of adware families employ payloads beyond conventional Java/Kotlin code. Based on our dataset analysis, we conducted a comprehensive Adware Characterization of 92 distinct adware families, revealing diverse implementation patterns and evolving techniques across the mobile ecosystem. To facilitate this analysis, we developed an Adware Characterization Schema that provided a structured taxonomy for systematically classifying the observed behaviors. Our investigation uncovered multiple categories of fraudulent activities, including aggressive ad display techniques, sophisticated click fraud implementations, privacy information leakage, malicious promotion mechanisms, and various persistence and evasion mechanisms employed to avoid detection while maximizing illicit revenue. This research establishes foundations for comprehending the fraudulent and adversarial techniques within the mobile adware landscape and facilitates the development of more robust detection mechanisms against these evolving threats.

广告软件是移动生态系统中普遍存在的威胁,但其固有特征在很大程度上被之前的研究所忽视。这项工作为揭开Android广告软件的神秘面纱迈出了关键一步。我们展示了AdwareZoo,这是一个综合数据集,包括从安全报告和应用程序存储库中收集的118个不同家族的15,996个广告软件样本。我们通过从VirusTotal重新扫描的样本中分离软件包来识别广告软件家族的有效载荷,揭示家族命名惯例中的独特模式,并揭露将合法广告网络错误分类为广告软件。我们对有效载荷定位策略的分析显示,超过30%的广告软件家族使用的有效载荷超出了传统的Java/Kotlin代码。基于我们的数据集分析,我们对92个不同的广告软件家族进行了全面的广告软件特征分析,揭示了移动生态系统中不同的实施模式和不断发展的技术。为了便于分析,我们开发了一个广告软件表征模式,为系统地分类观察到的行为提供了一个结构化的分类。我们的调查发现了多种类型的欺诈活动,包括激进的广告展示技术、复杂的点击欺诈实施、隐私信息泄露、恶意推广机制,以及各种用于避免检测的持久性和规避机制,同时最大限度地提高非法收入。本研究为理解移动广告软件领域中的欺诈和对抗技术奠定了基础,并促进了针对这些不断发展的威胁的更强大的检测机制的发展。
{"title":"The arts and crafts of android adware across a decade","authors":"Chao Wang,&nbsp;Tianming Liu,&nbsp;Yanjie Zhao,&nbsp;Lin Zhang,&nbsp;Xiaoning Du,&nbsp;Li Li,&nbsp;Haoyu Wang","doi":"10.1007/s10515-025-00575-9","DOIUrl":"10.1007/s10515-025-00575-9","url":null,"abstract":"<div><p>Adware represents a pervasive threat in the mobile ecosystem, yet its inherent characteristics have been largely overlooked by previous research. This work takes a crucial step towards demystifying Android adware. We present <b>AdwareZoo</b>, a comprehensive dataset comprising <b>15,996</b> adware samples across <b>118</b> distinct families collected from security reports and app repositories. We identify adware family payloads by isolating packages from samples for VirusTotal rescanning, unveiling distinctive patterns in family naming conventions, and exposing the misclassification of legitimate ad networks as adware. Our analysis of payload location strategies reveals that over 30% of adware families employ payloads beyond conventional Java/Kotlin code. Based on our dataset analysis, we conducted a comprehensive Adware Characterization of <b>92</b> distinct adware families, revealing diverse implementation patterns and evolving techniques across the mobile ecosystem. To facilitate this analysis, we developed an Adware Characterization Schema that provided a structured taxonomy for systematically classifying the observed behaviors. Our investigation uncovered multiple categories of fraudulent activities, including aggressive ad display techniques, sophisticated click fraud implementations, privacy information leakage, malicious promotion mechanisms, and various persistence and evasion mechanisms employed to avoid detection while maximizing illicit revenue. This research establishes foundations for comprehending the fraudulent and adversarial techniques within the mobile adware landscape and facilitates the development of more robust detection mechanisms against these evolving threats.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145510718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-view adaptive contrastive learning for information retrieval based fault localization 基于信息检索的多视图自适应对比学习故障定位
IF 3.1 2区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-11-06 DOI: 10.1007/s10515-025-00573-x
Chunying Zhou, Xiaoyuan Xie, Gong Chen, Peng He, Bing Li

Most studies focused on information retrieval-based techniques for fault localization, which built representations for bug reports and source code files and matched their semantic vectors through similarity measurement. However, such approaches often ignore some useful information that might help improve localization performance, such as 1) the interaction relationship between bug reports and source code files; 2) the similarity relationship between bug reports; and 3) the co-citation relationship between source code files. In this paper, we propose a novel approach named Multi-View Adaptive Contrastive Learning for Information Retrieval Fault Localization (MACL-IRFL) to learn the above-mentioned relationships for software fault localization. Specifically, we first generate data augmentations from report-code interaction view, report-report similarity view and code-code co-citation view separately, and adopt graph neural network to aggregate the information of bug reports or source code files from the three views in the embedding process. Moreover, we perform contrastive learning across these views. Our design of contrastive learning task will force the bug report representations to encode information shared by report-report and report-code views, and the source code file representations shared by code-code and report-code views, thereby alleviating the noise from auxiliary information. Finally, to evaluate the performance of our approach, we conduct extensive experiments on five open-source Java projects. The results show that our model can improve over the best baseline up to 28.93%, 25.57% and 20.35% on Accuracy@1, MAP and MRR, respectively.

大多数研究集中在基于信息检索的故障定位技术上,该技术为bug报告和源代码文件建立表示,并通过相似度度量匹配它们的语义向量。然而,这种方法往往忽略了一些可能有助于提高本地化性能的有用信息,例如1)bug报告和源代码文件之间的交互关系;2) bug报告之间的相似关系;3)源代码文件之间的共引关系。本文提出了一种基于多视图自适应对比学习的信息检索故障定位方法(MACL-IRFL)来学习上述关系进行软件故障定位。具体而言,我们首先分别从报告-代码交互视图、报告-报告相似视图和代码-代码共引视图生成数据增强,并在嵌入过程中采用图神经网络对三个视图中的bug报告或源代码文件信息进行聚合。此外,我们在这些视图中执行对比学习。我们设计的对比学习任务将迫使bug报告表示编码report-report视图和report-code视图共享的信息,以及code-code视图和report-code视图共享的源代码文件表示,从而减轻辅助信息的噪声。最后,为了评估我们的方法的性能,我们在五个开源Java项目上进行了广泛的实验。结果表明,该模型在Accuracy@1、MAP和MRR上分别比最佳基线提高28.93%、25.57%和20.35%。
{"title":"Multi-view adaptive contrastive learning for information retrieval based fault localization","authors":"Chunying Zhou,&nbsp;Xiaoyuan Xie,&nbsp;Gong Chen,&nbsp;Peng He,&nbsp;Bing Li","doi":"10.1007/s10515-025-00573-x","DOIUrl":"10.1007/s10515-025-00573-x","url":null,"abstract":"<div><p>Most studies focused on information retrieval-based techniques for fault localization, which built representations for bug reports and source code files and matched their semantic vectors through similarity measurement. However, such approaches often ignore some useful information that might help improve localization performance, such as 1) the interaction relationship between bug reports and source code files; 2) the similarity relationship between bug reports; and 3) the co-citation relationship between source code files. In this paper, we propose a novel approach named <b>M</b>ulti-View <b>A</b>daptive <b>C</b>ontrastive <b>L</b>earning for <b>I</b>nformation <b>R</b>etrieval <b>F</b>ault <b>L</b>ocalization (MACL-IRFL) to learn the above-mentioned relationships for software fault localization. Specifically, we first generate data augmentations from report-code interaction view, report-report similarity view and code-code co-citation view separately, and adopt graph neural network to aggregate the information of bug reports or source code files from the three views in the embedding process. Moreover, we perform contrastive learning across these views. Our design of contrastive learning task will force the bug report representations to encode information shared by report-report and report-code views, and the source code file representations shared by code-code and report-code views, thereby alleviating the noise from auxiliary information. Finally, to evaluate the performance of our approach, we conduct extensive experiments on five open-source Java projects. The results show that our model can improve over the best baseline up to 28.93%, 25.57% and 20.35% on Accuracy@1, MAP and MRR, respectively.</p></div>","PeriodicalId":55414,"journal":{"name":"Automated Software Engineering","volume":"33 1","pages":""},"PeriodicalIF":3.1,"publicationDate":"2025-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145456693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Automated Software Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1