首页 > 最新文献

Information and Software Technology最新文献

英文 中文
Required knowledge, skills and transversal competences for a career in software engineering 软件工程职业所需的知识、技能和横向能力
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-10-25 DOI: 10.1016/j.infsof.2025.107949
Nikolaos Mittas , Dimitrios Trygoniaris , Apostolos Ampatzoglou , Elvira-Maria Arvanitou , Christina Volioti , Alexander Chatzigeorgiou , Lefteris Angelis

Context

Possessing up-to-date knowledge, skills and transversal competencies (KSTs) is essential for both the successful delivery of software projects and a career in software engineering (SE). However, the technological landscape is changing rapidly, posing continuous challenges: for professionals entering the market or pivoting careers, for organizations hiring and monitoring workforce expertise and for educational institutes designing or updating their curricula.

Objectives

We study job requirements within and across SE occupations (Applications Programmers, Software Developers, Systems Analysts, Web and Multimedia Developers) to assist software organizations to better face skill mismatch and skills’ gap problems, software engineers in upskilling and reskilling endeavors and software education institutes in providing more industrially relevant curricula.

Method

In this study, we leverage a large corpus of online job advertisements, which are jointly collected by CEDEFOP and Eurostat. The dataset is analyzed through the lens of concepts and techniques from the study of biodiversity of species to assess the variation of expertise and identify skills that are transferable or unique in these occupations. Specifically, we adopt established diversity indices, such as alpha diversity, beta diversity, ordination methods, and indicator species analysis, aiming to quantify both the variety of skills within occupations and the differences across them. This approach highlights both the breadth and distinctiveness of expertise across occupations, rendering the biodiversity perspective a central and practical part of our methodology.

Results

The results reveal that the complete list of KSTs that is used to characterize the profiles of OJAs for SE-related occupations is very broad and that skillset required for each occupation is quite distinct, since there are statistically significant differences in the composition of the skillsets. Transversal Skills and Competences (T) appear to be the most transferable qualification; or “adapt to change” and “work in teams” are the KSTs that appears more uniformly to all studied software occupations, and “computer programming” is the top hard-skill that appears more uniformly to all occupations. However, each occupation shows some specific qualifications.

Conclusion

The results are contrasted against the literature, are interpreted, various implications to researchers and practitioners are provided, and a retrospective analysis of the tailoring of the biodiversity approach to SE labor landscape is provided. Overall, the proposed biodiversity analysis adds value by providing a novel, theory-driven methodology to assess skill variation, identifying both common and occupation-specific KSTs, and supporting evidence-based workforce and curriculum design.
拥有最新的知识、技能和横向能力(KSTs)对于软件项目的成功交付和软件工程(SE)的职业生涯都是必不可少的。然而,技术领域正在迅速变化,不断提出挑战:对于进入市场或转向职业的专业人员,对于雇用和监控劳动力专业知识的组织,以及设计或更新课程的教育机构。我们研究SE职业(应用程序程序员、软件开发人员、系统分析师、网络和多媒体开发人员)内部和跨职业的工作需求,以帮助软件组织更好地面对技能不匹配和技能差距问题,帮助软件工程师提高技能和再培训的努力,帮助软件教育机构提供更多与行业相关的课程。在这项研究中,我们利用了大量的在线招聘广告,这些广告是由CEDEFOP和欧盟统计局联合收集的。通过物种生物多样性研究的概念和技术来分析数据集,以评估专业知识的变化,并确定这些职业中可转移或独特的技能。具体而言,我们采用已建立的多样性指标,如α多样性、β多样性、协调方法和指标物种分析,旨在量化职业内部技能的多样性和不同职业之间的差异。这种方法强调了跨职业专业知识的广度和独特性,使生物多样性视角成为我们方法的核心和实用部分。结果表明,用于表征se相关职业oja概况的完整kst列表非常广泛,并且每个职业所需的技能组合非常不同,因为技能组合的组成在统计上存在显著差异。横向技能和能力(T)似乎是最可转移的资格;或者“适应变化”和“团队合作”是在所有被研究的软件职业中表现得更为一致的kst,而“计算机编程”是在所有职业中表现得更为一致的顶级硬技能。然而,每个职业都有一些特定的要求。结论将研究结果与文献进行了对比,并对研究结果进行了解释,提出了对研究者和实践者的启示,并对生物多样性方法在东南劳动力景观中的裁剪进行了回顾性分析。总体而言,拟议的生物多样性分析通过提供一种新颖的、理论驱动的方法来评估技能差异,识别常见的和特定职业的kst,并支持基于证据的劳动力和课程设计,从而增加了价值。
{"title":"Required knowledge, skills and transversal competences for a career in software engineering","authors":"Nikolaos Mittas ,&nbsp;Dimitrios Trygoniaris ,&nbsp;Apostolos Ampatzoglou ,&nbsp;Elvira-Maria Arvanitou ,&nbsp;Christina Volioti ,&nbsp;Alexander Chatzigeorgiou ,&nbsp;Lefteris Angelis","doi":"10.1016/j.infsof.2025.107949","DOIUrl":"10.1016/j.infsof.2025.107949","url":null,"abstract":"<div><h3>Context</h3><div>Possessing up-to-date knowledge, skills and transversal competencies (KSTs) is essential for both the successful delivery of software projects and a career in software engineering (SE). However, the technological landscape is changing rapidly, posing continuous challenges: for professionals entering the market or pivoting careers, for organizations hiring and monitoring workforce expertise and for educational institutes designing or updating their curricula.</div></div><div><h3>Objectives</h3><div>We study job requirements within and across SE occupations (Applications Programmers, Software Developers, Systems Analysts, Web and Multimedia Developers) to assist software organizations to better face skill mismatch and skills’ gap problems, software engineers in upskilling and reskilling endeavors and software education institutes in providing more industrially relevant curricula.</div></div><div><h3>Method</h3><div>In this study, we leverage a large corpus of online job advertisements, which are jointly collected by CEDEFOP and Eurostat. The dataset is analyzed through the lens of concepts and techniques from the study of biodiversity of species to assess the variation of expertise and identify skills that are transferable or unique in these occupations. Specifically, we adopt established diversity indices, such as alpha diversity, beta diversity, ordination methods, and indicator species analysis, aiming to quantify both the variety of skills within occupations and the differences across them. This approach highlights both the breadth and distinctiveness of expertise across occupations, rendering the biodiversity perspective a central and practical part of our methodology.</div></div><div><h3>Results</h3><div>The results reveal that the complete list of KSTs that is used to characterize the profiles of OJAs for SE-related occupations is very broad and that skillset required for each occupation is quite distinct, since there are statistically significant differences in the composition of the skillsets. Transversal Skills and Competences (T) appear to be the most transferable qualification; or “adapt to change” and “work in teams” are the KSTs that appears more uniformly to all studied software occupations, and “computer programming” is the top hard-skill that appears more uniformly to all occupations. However, each occupation shows some specific qualifications.</div></div><div><h3>Conclusion</h3><div>The results are contrasted against the literature, are interpreted, various implications to researchers and practitioners are provided, and a retrospective analysis of the tailoring of the biodiversity approach to SE labor landscape is provided. Overall, the proposed biodiversity analysis adds value by providing a novel, theory-driven methodology to assess skill variation, identifying both common and occupation-specific KSTs, and supporting evidence-based workforce and curriculum design.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107949"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145420378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SELink: A semantic-enhanced modular framework for issue–commit link recovery SELink:一个语义增强的模块框架,用于问题-提交链接恢复
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-14 DOI: 10.1016/j.infsof.2025.107973
Tao Zheng , Yangyang Zhao , Jiamin Guo , Liming Nie , Mingyue Jiang , Yuming Zhou

Context:

Issue–commit link recovery (ICLR) is a fundamental task in software traceability that aims to reconstruct missing or incomplete links between issue reports and code commits. Accurate link recovery supports downstream tasks such as bug localization, defect prediction, and change impact analysis. However, existing approaches often suffer from input truncation and limited semantic discrimination, which reduce their effectiveness in real-world settings.

Objective:

This study aims to improve the robustness and precision of issue–commit link recovery by addressing two key challenges: (1) the loss of semantic information caused by token truncation in long textual inputs, and (2) the difficulty of accurately ranking semantically similar candidate commits.

Methods:

We propose SELink, a Semantic-Enhanced framework that integrates large language models into the ICLR pipeline. SELink comprises three key components: a semantic summarization module that compresses verbose texts while preserving core semantics; a prediction module based on a pre-trained encoder, with flexibility to adopt various backbone models; and a semantic re-ranking module that refines the top-K candidate links through decoder-only LLM inference. In our implementation, the prediction module is instantiated using the EALink framework, though it can be substituted with other compatible encoders.

Results:

Experimental results show that SELink significantly outperforms state-of-the-art baseline models across most evaluation metrics and projects. In particular, it achieves notable improvements in Hit@1, NDCG@1, Recall@1 and LAG, highlighting its effectiveness in top-rank prediction accuracy. Ablation studies confirm the contribution of each module, and additional experiments across three different LLMs demonstrate the framework’s adaptability.

Conclusion:

SELink effectively addresses two long-standing limitations in issue–commit link recovery by leveraging LLM-based semantic compression and re-ranking. Its modular design and superior performance suggest that semantic-enhanced frameworks hold significant potential for advancing traceability tasks in software engineering.
上下文:问题-提交链接恢复(ICLR)是软件可追溯性中的一项基本任务,旨在重建问题报告和代码提交之间缺失或不完整的链接。准确的链接恢复支持下游任务,如bug定位、缺陷预测和变更影响分析。然而,现有的方法经常受到输入截断和有限的语义识别的影响,这降低了它们在现实环境中的有效性。目的:本研究旨在通过解决两个关键挑战来提高问题-提交链接恢复的鲁棒性和准确性:(1)长文本输入中令牌截断导致的语义信息丢失;(2)准确排序语义相似的候选提交的困难。方法:我们提出SELink,这是一个语义增强框架,将大型语言模型集成到ICLR管道中。SELink包括三个关键组件:一个语义摘要模块,在保留核心语义的同时压缩冗长的文本;基于预训练编码器的预测模块,可灵活采用各种骨干模型;以及一个语义重新排序模块,该模块通过仅解码器的LLM推理来提炼前k个候选链接。在我们的实现中,预测模块是使用EALink框架实例化的,尽管它可以用其他兼容的编码器替代。结果:实验结果表明,SELink在大多数评估指标和项目中显著优于最先进的基线模型。特别是在Hit@1, NDCG@1, Recall@1和LAG上取得了显著的改进,突出了其在top-rank预测精度上的有效性。消融研究证实了每个模块的贡献,另外在三个不同的llm上进行的实验证明了该框架的适应性。结论:SELink通过利用基于llm的语义压缩和重新排序,有效地解决了问题提交链接恢复中两个长期存在的限制。它的模块化设计和优越的性能表明,语义增强的框架在推进软件工程中的可跟踪性任务方面具有重要的潜力。
{"title":"SELink: A semantic-enhanced modular framework for issue–commit link recovery","authors":"Tao Zheng ,&nbsp;Yangyang Zhao ,&nbsp;Jiamin Guo ,&nbsp;Liming Nie ,&nbsp;Mingyue Jiang ,&nbsp;Yuming Zhou","doi":"10.1016/j.infsof.2025.107973","DOIUrl":"10.1016/j.infsof.2025.107973","url":null,"abstract":"<div><h3>Context:</h3><div>Issue–commit link recovery (ICLR) is a fundamental task in software traceability that aims to reconstruct missing or incomplete links between issue reports and code commits. Accurate link recovery supports downstream tasks such as bug localization, defect prediction, and change impact analysis. However, existing approaches often suffer from input truncation and limited semantic discrimination, which reduce their effectiveness in real-world settings.</div></div><div><h3>Objective:</h3><div>This study aims to improve the robustness and precision of issue–commit link recovery by addressing two key challenges: (1) the loss of semantic information caused by token truncation in long textual inputs, and (2) the difficulty of accurately ranking semantically similar candidate commits.</div></div><div><h3>Methods:</h3><div>We propose SELink, a Semantic-Enhanced framework that integrates large language models into the ICLR pipeline. SELink comprises three key components: a semantic summarization module that compresses verbose texts while preserving core semantics; a prediction module based on a pre-trained encoder, with flexibility to adopt various backbone models; and a semantic re-ranking module that refines the top-K candidate links through decoder-only LLM inference. In our implementation, the prediction module is instantiated using the EALink framework, though it can be substituted with other compatible encoders.</div></div><div><h3>Results:</h3><div>Experimental results show that SELink significantly outperforms state-of-the-art baseline models across most evaluation metrics and projects. In particular, it achieves notable improvements in Hit@1, NDCG@1, Recall@1 and LAG, highlighting its effectiveness in top-rank prediction accuracy. Ablation studies confirm the contribution of each module, and additional experiments across three different LLMs demonstrate the framework’s adaptability.</div></div><div><h3>Conclusion:</h3><div>SELink effectively addresses two long-standing limitations in issue–commit link recovery by leveraging LLM-based semantic compression and re-ranking. Its modular design and superior performance suggest that semantic-enhanced frameworks hold significant potential for advancing traceability tasks in software engineering.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107973"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145571288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fairness set and forgotten: Mining fairness toolkit usage in open-source machine learning projects 公平性设置和遗忘:在开源机器学习项目中挖掘公平性工具包的使用
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-04 DOI: 10.1016/j.infsof.2025.107957
Alfonso Cannavale, Gianmario Voria, Antonio Scognamiglio, Giammaria Giordano, Gemma Catolino, Fabio Palomba

Context:

The development of machine learning (ML) systems in high-stakes domains has amplified concerns about fairness, prompting the creation of fairness toolkits offering metrics and mitigation techniques. Open-source software (OSS) ecosystems, a critical driver of AI innovation, present a unique opportunity to study the practical adoption of these toolkits.

Objective:

This paper aims to empirically characterize the adoption of fairness toolkits in OSS ML projects by investigating for what purposes they are used and how their usage evolves over time.

Methods:

We conducted a mining study on GitHub repositories related to real-world ML projects that integrate fairness toolkits such as AIF360 and Fairlearn. Starting from 1,096 candidate repositories, we applied systematic filtering to identify a final dataset of 20 relevant ML projects (comprising 5,777 total commits). We analyzed toolkit usage by examining invoked APIs and commit histories to uncover patterns of adoption and evolution.

Results:

Our findings reveal that fairness toolkits are predominantly used for diagnostic purposes, with analytic components integrated early in the project lifecycle and rarely modified thereafter. In contrast, mitigation techniques are infrequently adopted, tend to appear later, and exhibit short, unstable lifespans.

Conclusion:

Our results show that the adoption of fairness toolkits in OSS ML projects is limited and often restricted to initial diagnostic phases, with active mitigation practices remaining rare. These findings highlight the need for improved support to foster more sustained and effective integration of fairness practices within open-source development.
背景:高风险领域机器学习(ML)系统的发展加剧了人们对公平性的担忧,促使人们创建了提供指标和缓解技术的公平性工具包。开源软件(OSS)生态系统是人工智能创新的关键驱动力,为研究这些工具包的实际应用提供了独特的机会。目的:本文旨在通过调查它们的使用目的以及它们的使用如何随着时间的推移而演变,以经验的方式描述开源软件ML项目中公平性工具包的采用。方法:我们对与现实世界ML项目相关的GitHub存储库进行了挖掘研究,这些项目集成了公平工具包,如AIF360和Fairlearn。从1096个候选存储库开始,我们应用系统过滤来确定20个相关ML项目的最终数据集(包括5777个总提交)。我们通过检查调用的api和提交历史来分析工具包的使用情况,以揭示采用和发展的模式。结果:我们的发现表明公平性工具包主要用于诊断目的,在项目生命周期的早期集成了分析组件,此后很少修改。相比之下,缓解技术很少被采用,往往出现得较晚,寿命较短,不稳定。结论:我们的结果表明,在开源软件ML项目中,公平性工具包的采用是有限的,而且通常仅限于初始诊断阶段,积极的缓解实践仍然很少。这些发现强调了在开源开发中促进更持久和有效的公平实践整合的改进支持的必要性。
{"title":"Fairness set and forgotten: Mining fairness toolkit usage in open-source machine learning projects","authors":"Alfonso Cannavale,&nbsp;Gianmario Voria,&nbsp;Antonio Scognamiglio,&nbsp;Giammaria Giordano,&nbsp;Gemma Catolino,&nbsp;Fabio Palomba","doi":"10.1016/j.infsof.2025.107957","DOIUrl":"10.1016/j.infsof.2025.107957","url":null,"abstract":"<div><h3>Context:</h3><div>The development of machine learning (ML) systems in high-stakes domains has amplified concerns about fairness, prompting the creation of fairness toolkits offering metrics and mitigation techniques. Open-source software (OSS) ecosystems, a critical driver of AI innovation, present a unique opportunity to study the practical adoption of these toolkits.</div></div><div><h3>Objective:</h3><div>This paper aims to empirically characterize the adoption of fairness toolkits in OSS ML projects by investigating <em>for what purposes</em> they are used and <em>how</em> their usage evolves over time.</div></div><div><h3>Methods:</h3><div>We conducted a mining study on <span>GitHub</span> repositories related to real-world ML projects that integrate fairness toolkits such as <span>AIF360</span> and <span>Fairlearn</span>. Starting from 1,096 candidate repositories, we applied systematic filtering to identify a final dataset of 20 relevant ML projects (comprising 5,777 total commits). We analyzed toolkit usage by examining invoked APIs and commit histories to uncover patterns of adoption and evolution.</div></div><div><h3>Results:</h3><div>Our findings reveal that fairness toolkits are predominantly used for diagnostic purposes, with analytic components integrated early in the project lifecycle and rarely modified thereafter. In contrast, mitigation techniques are infrequently adopted, tend to appear later, and exhibit short, unstable lifespans.</div></div><div><h3>Conclusion:</h3><div>Our results show that the adoption of fairness toolkits in OSS ML projects is limited and often restricted to initial diagnostic phases, with active mitigation practices remaining rare. These findings highlight the need for improved support to foster more sustained and effective integration of fairness practices within open-source development.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107957"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145468020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using LLMs to enhance code quality: A systematic literature review 使用llm提高代码质量:系统的文献回顾
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-19 DOI: 10.1016/j.infsof.2025.107960
Nawaf Alomari , Moussa Redah , Ahmad Ashraf , Mohammad Alshayeb

Context:

Large Language Models (LLMs) are increasingly used in software engineering to enhance code quality through tasks such as refactoring and code smell detection and many other tasks. Code smells are poor design decisions that can be resolved by changing the internal structure of the code without affecting its output, a process known as refactoring.

Objective:

This study systematically reviews the use of LLMs in code quality enhancement, focusing on techniques such as refactoring, smell detection, and other code improvement methods.

Method:

Using SLR techniques, we reviewed 49 studies up to September 2024, analyzing both qualitative and quantitative data to assess trends and effectiveness.

Results:

The field is active, with refactoring as the most common task, followed by smell detection. Refactored code by LLMs is not reliable. Prompting is used more frequently than fine-tuning, with few-shot learning as the leading prompting method. Java and Python are the most represented languages, while F1, Precision, Recall, and Accuracy are common evaluation metrics, along with BLEU and EM for generation tasks. Open-source and general language models are preferred, with validation datasets as the primary validation approach.

Conclusions:

LLMs show promise for code quality improvement, but challenges in optimization and reliability remain. Future research should prioritize fine-tuning for refactoring, linking LLMs to specific quality attributes, developing benchmark datasets, constructing datasets for diverse programming languages, and exploring a wider range of promoting techniques.
背景:大型语言模型(llm)越来越多地用于软件工程,通过重构和代码气味检测等任务来提高代码质量。代码气味是糟糕的设计决策,可以通过改变代码的内部结构而不影响其输出来解决,这个过程称为重构。目的:本研究系统回顾了法学硕士在代码质量提升中的应用,重点介绍了重构、气味检测等代码改进方法。方法:采用单反技术,回顾截至2024年9月的49项研究,分析定性和定量数据,评估趋势和有效性。结果:该领域是活跃的,重构是最常见的任务,其次是气味检测。llm重构的代码是不可靠的。提示比微调使用得更频繁,少射学习是主要的提示方法。Java和Python是最具代表性的语言,而F1、Precision、Recall和Accuracy是常见的评估指标,BLEU和EM用于生成任务。开源和通用语言模型是首选,验证数据集是主要的验证方法。结论:llm显示了代码质量改进的希望,但在优化和可靠性方面仍然存在挑战。未来的研究应该优先考虑重构的微调,将llm与特定的质量属性联系起来,开发基准数据集,为不同的编程语言构建数据集,并探索更广泛的促进技术。
{"title":"Using LLMs to enhance code quality: A systematic literature review","authors":"Nawaf Alomari ,&nbsp;Moussa Redah ,&nbsp;Ahmad Ashraf ,&nbsp;Mohammad Alshayeb","doi":"10.1016/j.infsof.2025.107960","DOIUrl":"10.1016/j.infsof.2025.107960","url":null,"abstract":"<div><h3>Context:</h3><div>Large Language Models (LLMs) are increasingly used in software engineering to enhance code quality through tasks such as refactoring and code smell detection and many other tasks. Code smells are poor design decisions that can be resolved by changing the internal structure of the code without affecting its output, a process known as refactoring.</div></div><div><h3>Objective:</h3><div>This study systematically reviews the use of LLMs in code quality enhancement, focusing on techniques such as refactoring, smell detection, and other code improvement methods.</div></div><div><h3>Method:</h3><div>Using SLR techniques, we reviewed 49 studies up to September 2024, analyzing both qualitative and quantitative data to assess trends and effectiveness.</div></div><div><h3>Results:</h3><div>The field is active, with refactoring as the most common task, followed by smell detection. Refactored code by LLMs is not reliable. Prompting is used more frequently than fine-tuning, with few-shot learning as the leading prompting method. Java and Python are the most represented languages, while F1, Precision, Recall, and Accuracy are common evaluation metrics, along with BLEU and EM for generation tasks. Open-source and general language models are preferred, with validation datasets as the primary validation approach.</div></div><div><h3>Conclusions:</h3><div>LLMs show promise for code quality improvement, but challenges in optimization and reliability remain. Future research should prioritize fine-tuning for refactoring, linking LLMs to specific quality attributes, developing benchmark datasets, constructing datasets for diverse programming languages, and exploring a wider range of promoting techniques.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107960"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145571427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CPMT: A collaborative metamorphic relations and test cases prioritization approach for Metamorphic Testing CPMT:一种用于变形测试的协作的变形关系和测试用例优先化方法
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-21 DOI: 10.1016/j.infsof.2025.107975
Chang-ai Sun, Shifan Liu, An Fu, Jiaming Zhang

Context:

Metamorphic Testing (MT) is a widely adopted software testing technique that addresses the oracle problem by leveraging Metamorphic Relations (MRs). Various test case prioritization (TCP) techniques have been developed to improve the fault detection efficiency by scheduling the execution order of test cases. However, these techniques cannot be directly applied to MT due to its unique features, such as involving the execution of source and follow-up test cases, application of MRs, and the result verification depends on the availability of the corresponding outputs.

Objective:

This study aims to improve the fault detection efficiency of MT by developing a collaborative prioritization approach called CPMT that considers the scheduling of both MRs and test cases.

Methods:

We first formulate the prioritization problem in MT and then propose to schedule the execution of MRs and test cases based on three strategies, which prioritize the execution of MRs and test cases with a higher potential of fault detection from the different perspectives, including the contributions to specification/implementation coverage, the strictness of output relation, and the earlier detection opportunities.

Results:

Extensive experiments were conducted on seven subject programs to evaluate the effectiveness of CPMT. The experimental results have demonstrated that the proposed approach significantly improved the fault detection efficiency and outperformed the baseline techniques.

Conclusion:

CPMT provides a promising way to improve the fault detection efficiency of MT.
背景:变形测试(MT)是一种被广泛采用的软件测试技术,它通过利用变形关系(MRs)来解决oracle问题。各种测试用例优先级(TCP)技术被开发出来,通过调度测试用例的执行顺序来提高故障检测效率。然而,这些技术不能直接应用于机器翻译,因为其独特的特点,如涉及源和后续测试用例的执行,MRs的应用,结果验证取决于相应输出的可用性。目的:本研究旨在通过开发一种考虑MRs和测试用例调度的称为CPMT的协作优先级方法来提高MT的故障检测效率。方法:首先在机器翻译中提出优先级问题,然后提出基于三种策略的机器翻译和测试用例执行调度,这三种策略从不同的角度,包括对规范/实现覆盖率的贡献、输出关系的严格性和早期检测机会,对具有更高故障检测潜力的机器翻译和测试用例的执行进行优先级排序。结果:在7个实验项目中进行了大量的实验来评估CPMT的有效性。实验结果表明,该方法显著提高了故障检测效率,优于基线技术。结论:CPMT为提高MT的故障检测效率提供了一种很有前途的方法。
{"title":"CPMT: A collaborative metamorphic relations and test cases prioritization approach for Metamorphic Testing","authors":"Chang-ai Sun,&nbsp;Shifan Liu,&nbsp;An Fu,&nbsp;Jiaming Zhang","doi":"10.1016/j.infsof.2025.107975","DOIUrl":"10.1016/j.infsof.2025.107975","url":null,"abstract":"<div><h3>Context:</h3><div>Metamorphic Testing (MT) is a widely adopted software testing technique that addresses the <em>oracle problem</em> by leveraging Metamorphic Relations (MRs). Various test case prioritization (TCP) techniques have been developed to improve the fault detection efficiency by scheduling the execution order of test cases. However, these techniques cannot be directly applied to MT due to its unique features, such as involving the execution of source and follow-up test cases, application of MRs, and the result verification depends on the availability of the corresponding outputs.</div></div><div><h3>Objective:</h3><div>This study aims to improve the fault detection efficiency of MT by developing a collaborative prioritization approach called <em>CPMT</em> that considers the scheduling of both MRs and test cases.</div></div><div><h3>Methods:</h3><div>We first formulate the prioritization problem in MT and then propose to schedule the execution of MRs and test cases based on three strategies, which prioritize the execution of MRs and test cases with a higher potential of fault detection from the different perspectives, including the contributions to specification/implementation coverage, the strictness of output relation, and the earlier detection opportunities.</div></div><div><h3>Results:</h3><div>Extensive experiments were conducted on seven subject programs to evaluate the effectiveness of <em>CPMT</em>. The experimental results have demonstrated that the proposed approach significantly improved the fault detection efficiency and outperformed the baseline techniques.</div></div><div><h3>Conclusion:</h3><div><em>CPMT</em> provides a promising way to improve the fault detection efficiency of MT.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107975"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
KIBO, a new hybrid software development method with enhanced information systems auditing capability KIBO是一种新的混合软件开发方法,具有增强的信息系统审计能力
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-01 DOI: 10.1016/j.infsof.2025.107958
Ioannis Kirpitsas, Theodore Pachidis

Context

The agile revolution has transformed software development, accelerating the adoption of hybrid methods that combine agile and traditional plan-driven practices. These methods offer IT organizations speed and flexibility. However, they can increase governance complexity and complicate value assurance, which is defined here as protecting business value through auditable compliance.

Objective

This paper introduces KIBO, a hybrid software development method. This method embeds auditability that is aligned with COBIT 2019 objectives, the industry-standard framework for IT governance and management, offering more robust control processes than existing methods.

Method

KIBO integrates Scrum, DevOps, and Disciplined Agile Delivery with governance principles through three practices. First, it incorporates IT controls from sprint planning. Second, it applies structured application lifecycle management elements for transparency and traceability. Third, it offers a centralized information systems audit hub that enables continuous auditing. The method was validated in two stages. Initially, the theoretical model was assessed by 90 experts against existing hybrid methods using selected COBIT 2019 objectives as reference criteria. Subsequently, a practical evaluation validated expectations derived from this theoretical assessment.

Results

Theoretical assessment indicates that KIBO enhances IT governance, risk management, and quality. During a 7-week deployment at a leading Southeastern European digital services organization, delivery accelerated by 6.4 % to reach the 95 % epic completion threshold, while issue density decreased by 17.9 % compared to organizational averages. Additionally, the team reported improved quality, process reliability, and net developer satisfaction in practice.

Conclusion

The evaluation suggests KIBO can enhance governance in agile development by supporting alignment with business objectives, improving process efficiency, and strengthening audit readiness. Although challenges remain in role integration and security oversight, KIBO helps balance agility and control. We recommend broader deployments to validate and further extend these positive outcomes.
敏捷革命已经改变了软件开发,加速了结合敏捷和传统计划驱动实践的混合方法的采用。这些方法为IT组织提供了速度和灵活性。然而,它们会增加治理的复杂性并使价值保证复杂化,这里将其定义为通过可审计的遵从性来保护业务价值。目的介绍一种混合软件开发方法KIBO。该方法嵌入了与COBIT 2019目标(IT治理和管理的行业标准框架)一致的可审核性,提供了比现有方法更健壮的控制过程。MethodKIBO通过三个实践将Scrum、DevOps和有纪律的敏捷交付与治理原则集成在一起。首先,它结合了sprint计划中的it控制。其次,它应用结构化的应用程序生命周期管理元素来实现透明性和可追溯性。第三,它提供了一个集中的信息系统审计中心,支持持续的审计。该方法分两个阶段进行验证。最初,90名专家以选定的COBIT 2019目标作为参考标准,对现有的混合方法进行了理论模型评估。随后,实际评估验证了从理论评估中得出的期望。结果理论评估表明KIBO增强了IT治理、风险管理和质量。在一家领先的东南欧数字服务组织的7周部署期间,交付速度加快了6.4%,达到了95%的史诗完成阈值,而问题密度与组织平均水平相比下降了17.9%。此外,团队报告了在实践中改进的质量、过程可靠性和网络开发人员满意度。评估表明KIBO可以通过支持与业务目标的一致性、提高流程效率和加强审计准备来增强敏捷开发中的治理。尽管在角色集成和安全监督方面仍然存在挑战,KIBO有助于平衡敏捷性和控制。我们建议进行更广泛的部署,以验证并进一步扩展这些积极成果。
{"title":"KIBO, a new hybrid software development method with enhanced information systems auditing capability","authors":"Ioannis Kirpitsas,&nbsp;Theodore Pachidis","doi":"10.1016/j.infsof.2025.107958","DOIUrl":"10.1016/j.infsof.2025.107958","url":null,"abstract":"<div><h3>Context</h3><div>The agile revolution has transformed software development, accelerating the adoption of hybrid methods that combine agile and traditional plan-driven practices. These methods offer IT organizations speed and flexibility. However, they can increase governance complexity and complicate value assurance, which is defined here as protecting business value through auditable compliance.</div></div><div><h3>Objective</h3><div>This paper introduces KIBO, a hybrid software development method. This method embeds auditability that is aligned with COBIT 2019 objectives, the industry-standard framework for IT governance and management, offering more robust control processes than existing methods.</div></div><div><h3>Method</h3><div>KIBO integrates Scrum, DevOps, and Disciplined Agile Delivery with governance principles through three practices. First, it incorporates IT controls from sprint planning. Second, it applies structured application lifecycle management elements for transparency and traceability. Third, it offers a centralized information systems audit hub that enables continuous auditing. The method was validated in two stages. Initially, the theoretical model was assessed by 90 experts against existing hybrid methods using selected COBIT 2019 objectives as reference criteria. Subsequently, a practical evaluation validated expectations derived from this theoretical assessment.</div></div><div><h3>Results</h3><div>Theoretical assessment indicates that KIBO enhances IT governance, risk management, and quality. During a 7-week deployment at a leading Southeastern European digital services organization, delivery accelerated by 6.4 % to reach the 95 % epic completion threshold, while issue density decreased by 17.9 % compared to organizational averages. Additionally, the team reported improved quality, process reliability, and net developer satisfaction in practice.</div></div><div><h3>Conclusion</h3><div>The evaluation suggests KIBO can enhance governance in agile development by supporting alignment with business objectives, improving process efficiency, and strengthening audit readiness. Although challenges remain in role integration and security oversight, KIBO helps balance agility and control. We recommend broader deployments to validate and further extend these positive outcomes.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107958"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145468024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The role of voice and silence behaviors in software development: a structural equation modeling analysis 声音和沉默行为在软件开发中的作用:结构方程建模分析
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-14 DOI: 10.1016/j.infsof.2025.107972
Mary Sánchez-Gordón , Ricardo Colomo-Palacios , Alex Sanchez Gordon , Aliaksandr Hubin

Context

Most software companies strive to have high-performing teams and mitigate withdrawal behaviors like being present but unproductive. In this context, psychological safety and developers’ perceived impact are suggested as potential drivers of voice and silence behaviors. However, understanding these social aspects of software development entails the incorporation of social science theories.

Objective

This study aims to empirically explore the relationships among software professionals’ perceived impact, psychological safety, voice and silence behaviors, burnout particularly withdrawal and performance using a theoretical model.

Method

A survey questionnaire was conducted, resulting in 158 valid responses from software development teams. Then, we analyzed the responses using structural equation modeling (SEM) and a novel semi-confirmatory SEM. All variables were measured using pre-validated instruments.

Results

The findings supported the theoretical model, showing that psychological safety was more related to silence than voice, whereas perceived impact showed a stronger relationship with voice than with silence. Furthermore, silence contributed to higher burnout, whereas voice alleviated it. In contrast, silence showed a negative association with performance, whereas voice was positively associated with it.

Conclusions

This study examines the direct effects of perceived impact and psychological safety on voice and silence behaviors. Additionally, it examines the direct effects of these behaviors on burnout and task performance among software professionals. Consequently, the findings offer both theoretical and practical insights.
大多数软件公司都努力拥有高绩效的团队,并减少诸如出现但没有效率的退缩行为。在此背景下,心理安全和开发者感知影响被认为是发声和沉默行为的潜在驱动因素。然而,理解软件开发的这些社会方面需要结合社会科学理论。目的运用理论模型,实证探讨软件专业人员的感知影响、心理安全、发声沉默行为、倦怠尤其是退缩与绩效的关系。方法进行问卷调查,得到158个软件开发团队的有效回复。然后,我们使用结构方程模型(SEM)和一种新型的半验证性SEM来分析响应。使用预先验证的仪器测量所有变量。结果研究结果支持了理论模型,表明心理安全与沉默的关系大于与声音的关系,而感知影响与声音的关系大于与沉默的关系。此外,沉默会导致更高的倦怠,而声音则会缓解倦怠。相比之下,沉默与表现呈负相关,而声音与表现呈正相关。结论本研究考察了感知冲击和心理安全对建言和沉默行为的直接影响。此外,它还研究了这些行为对软件专业人员的倦怠和任务绩效的直接影响。因此,研究结果提供了理论和实践的见解。
{"title":"The role of voice and silence behaviors in software development: a structural equation modeling analysis","authors":"Mary Sánchez-Gordón ,&nbsp;Ricardo Colomo-Palacios ,&nbsp;Alex Sanchez Gordon ,&nbsp;Aliaksandr Hubin","doi":"10.1016/j.infsof.2025.107972","DOIUrl":"10.1016/j.infsof.2025.107972","url":null,"abstract":"<div><h3>Context</h3><div>Most software companies strive to have high-performing teams and mitigate withdrawal behaviors like being present but unproductive. In this context, psychological safety and developers’ perceived impact are suggested as potential drivers of voice and silence behaviors. However, understanding these social aspects of software development entails the incorporation of social science theories.</div></div><div><h3>Objective</h3><div>This study aims to empirically explore the relationships among software professionals’ perceived impact, psychological safety, voice and silence behaviors, burnout particularly withdrawal and performance using a theoretical model.</div></div><div><h3>Method</h3><div>A survey questionnaire was conducted, resulting in 158 valid responses from software development teams. Then, we analyzed the responses using structural equation modeling (SEM) and a novel semi-confirmatory SEM. All variables were measured using pre-validated instruments.</div></div><div><h3>Results</h3><div>The findings supported the theoretical model, showing that psychological safety was more related to silence than voice, whereas perceived impact showed a stronger relationship with voice than with silence. Furthermore, silence contributed to higher burnout, whereas voice alleviated it. In contrast, silence showed a negative association with performance, whereas voice was positively associated with it.</div></div><div><h3>Conclusions</h3><div>This study examines the direct effects of perceived impact and psychological safety on voice and silence behaviors. Additionally, it examines the direct effects of these behaviors on burnout and task performance among software professionals. Consequently, the findings offer both theoretical and practical insights.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107972"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detecting Slow Loop smell by using deep learning: From direct-learning to transfer-learning in cross-language settings 利用深度学习检测慢循环气味:从直接学习到跨语言环境下的迁移学习
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-20 DOI: 10.1016/j.infsof.2025.107976
Yuqi Li , Ruonan Ma , Yixin Bian , Weijie Chen , Lin Huang , Jiaying Li , Federica Sarro

Context:

Code smells are symptoms of poor quality code that hinder software maintenance and evolution. Although most existing studies have focused on Object-Oriented code smells, limited attention has been paid to Android-specific smells, particularly in Kotlin-based Android applications.

Objective:

To identify Android-specific code smells, with a focus on the Slow Loop smell in both Java-based and Kotlin-based Android apps.

Methods:

This paper presents a novel deep learning–based approach for identifying the Slow Loop smell in both Java-based and Kotlin-based Android apps. We propose a two-phase strategy that integrates direct-learning and transfer-learning. In the first phase, Slow Loop smells are identified in Java-based Android applications using a custom tool, Android-specific Smell Detection, which generates labeled training samples. These samples are tokenized and used to train five deep learning models, namely CNN, RNN, CNN-LSTM, and two variants of autoencoders. In the second phase, a transfer-learning mechanism adapts Java-trained models to Kotlin-based applications, thus enabling cross-language detection. The effectiveness of the approach is empirically validated in 80 Java and 90 Kotlin applications, using precision, recall, F1-score, and MCC as evaluation metrics.

Results:

The findings indicate that deep learning methods can be used to detect Android-specific code smells in Java-developed Android applications. In addition, transfer-learning successfully enables Kotlin smell detection, achieving competitive performance without the need for Kotlin-specific training from scratch.

Conclusions:

This study demonstrates the feasibility of applying transfer-learning to cross-language code smell detection. It contributes a scalable framework to identify Android-specific smells and provides new insights into improving code quality in mobile software development.
上下文:代码气味是妨碍软件维护和发展的低质量代码的症状。尽管大多数现有的研究都集中在面向对象的代码气味上,但对Android特定气味的关注有限,特别是在基于kotlin的Android应用程序中。目的:识别Android特定的代码气味,重点关注基于java和基于kotlin的Android应用程序中的慢循环气味。方法:本文提出了一种新的基于深度学习的方法,用于识别基于java和基于kotlin的Android应用程序中的慢循环气味。我们提出了一种整合直接学习和迁移学习的两阶段策略。在第一阶段,在基于java的Android应用程序中使用定制工具(Android特定的气味检测)识别慢循环气味,该工具生成标记的训练样本。这些样本被标记化并用于训练五种深度学习模型,即CNN、RNN、CNN- lstm和两种自编码器变体。在第二阶段,迁移学习机制将java训练的模型适应基于kotlin的应用程序,从而实现跨语言检测。该方法的有效性在80个Java和90个Kotlin应用程序中得到了经验验证,使用精度、召回率、f1分数和MCC作为评估指标。结果:研究结果表明,深度学习方法可以用于检测java开发的Android应用程序中Android特定代码的气味。此外,迁移学习成功地实现了Kotlin气味检测,无需从头开始进行Kotlin特定的培训即可获得竞争性能。结论:本研究证明了将迁移学习应用于跨语言代码气味检测的可行性。它提供了一个可扩展的框架来识别android特定的气味,并为提高移动软件开发中的代码质量提供了新的见解。
{"title":"Detecting Slow Loop smell by using deep learning: From direct-learning to transfer-learning in cross-language settings","authors":"Yuqi Li ,&nbsp;Ruonan Ma ,&nbsp;Yixin Bian ,&nbsp;Weijie Chen ,&nbsp;Lin Huang ,&nbsp;Jiaying Li ,&nbsp;Federica Sarro","doi":"10.1016/j.infsof.2025.107976","DOIUrl":"10.1016/j.infsof.2025.107976","url":null,"abstract":"<div><h3>Context:</h3><div>Code smells are symptoms of poor quality code that hinder software maintenance and evolution. Although most existing studies have focused on Object-Oriented code smells, limited attention has been paid to Android-specific smells, particularly in Kotlin-based Android applications.</div></div><div><h3>Objective:</h3><div>To identify Android-specific code smells, with a focus on the Slow Loop smell in both Java-based and Kotlin-based Android apps.</div></div><div><h3>Methods:</h3><div>This paper presents a novel deep learning–based approach for identifying the Slow Loop smell in both Java-based and Kotlin-based Android apps. We propose a two-phase strategy that integrates direct-learning and transfer-learning. In the first phase, Slow Loop smells are identified in Java-based Android applications using a custom tool, Android-specific Smell Detection, which generates labeled training samples. These samples are tokenized and used to train five deep learning models, namely CNN, RNN, CNN-LSTM, and two variants of autoencoders. In the second phase, a transfer-learning mechanism adapts Java-trained models to Kotlin-based applications, thus enabling cross-language detection. The effectiveness of the approach is empirically validated in 80 Java and 90 Kotlin applications, using precision, recall, F1-score, and MCC as evaluation metrics.</div></div><div><h3>Results:</h3><div>The findings indicate that deep learning methods can be used to detect Android-specific code smells in Java-developed Android applications. In addition, transfer-learning successfully enables Kotlin smell detection, achieving competitive performance without the need for Kotlin-specific training from scratch.</div></div><div><h3>Conclusions:</h3><div>This study demonstrates the feasibility of applying transfer-learning to cross-language code smell detection. It contributes a scalable framework to identify Android-specific smells and provides new insights into improving code quality in mobile software development.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107976"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145617877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GreenDFL: A framework for assessing the sustainability of Decentralized Federated Learning systems greenfl:一个评估分散式联邦学习系统可持续性的框架
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-10-18 DOI: 10.1016/j.infsof.2025.107937
Chao Feng , Alberto Huertas Celdrán , Xi Cheng , Gérôme Bovet , Burkhard Stiller

Context:

Decentralized Federated Learning (DFL) is an emerging paradigm that enables collaborative model training without centralized data and model aggregation, enhancing privacy and resilience. However, its sustainability remains underexplored, as energy consumption and carbon emissions vary across different system configurations. Understanding the environmental impact of DFL is crucial for optimizing its design and deployment.

Objective:

This work aims to develop a comprehensive and operational framework for assessing the sustainability of DFL systems. To address it, this work provides a systematic method for quantifying energy consumption and carbon emissions, offering insights into improving the sustainability of DFL.

Methods:

This work proposes GreenDFL, a fully implementable framework that has been integrated into a real-world DFL platform. GreenDFL systematically analyzes the impact of various factors, including hardware accelerators, model architecture, communication medium, data distribution, network topology, and federation size, on the sustainability of DFL systems. Besides, a sustainability-aware aggregation algorithm (GreenDFL-SA) and a node selection algorithm (GreenDFL-SN) are developed to optimize energy efficiency and reduce carbon emissions in DFL training.

Results:

Empirical experiments are conducted on multiple datasets, measuring energy consumption and carbon emissions at different phases of the DFL lifecycle. Results indicate that local training dominates energy consumption and carbon emissions, while communication has a relatively minor impact. Optimizing model complexity, using GPUs instead of CPUs, and strategically selecting participating nodes significantly improve sustainability. Additionally, using wired communication, particularly optical fiber, effectively reduces energy consumption during the communication phase, while integrating early stopping mechanisms further minimizes overall emissions.

Conclusion:

The proposed GreenDFL provides a comprehensive and practical approach for assessing the sustainability of DFL systems. Furthermore, it offers best practices for improving environmental efficiency in DFL, making sustainability considerations more actionable in real-world deployments.
上下文:去中心化联邦学习(DFL)是一种新兴的范例,可以在没有集中数据和模型聚合的情况下进行协作模型训练,增强隐私和弹性。然而,由于能源消耗和碳排放在不同的系统配置中有所不同,其可持续性仍未得到充分探索。了解DFL对环境的影响对于优化其设计和部署至关重要。目的:本工作旨在开发一个全面的和可操作的框架来评估DFL系统的可持续性。为了解决这一问题,本研究提供了一种量化能源消耗和碳排放的系统方法,为提高DFL的可持续性提供了见解。方法:本工作提出了greenfl,这是一个完全可实现的框架,已集成到现实世界的DFL平台中。greenfl系统地分析了硬件加速器、模型架构、通信介质、数据分布、网络拓扑和联邦规模等各种因素对DFL系统可持续性的影响。此外,为了优化DFL训练中的能源效率和减少碳排放,提出了可持续感知聚合算法(greenfl - sa)和节点选择算法(greenfl - sn)。结果:在多个数据集上进行了实证实验,测量了DFL生命周期不同阶段的能耗和碳排放。结果表明,当地培训在能源消耗和碳排放中占主导地位,而通信的影响相对较小。优化模型复杂度,使用gpu代替cpu,策略性地选择参与节点,显著提高了可持续性。此外,使用有线通信,特别是光纤,有效地降低了通信阶段的能耗,同时集成早期停止机制进一步减少了总排放量。结论:提出的greenfl为评估DFL系统的可持续性提供了一种全面实用的方法。此外,它还提供了提高DFL环境效率的最佳实践,使可持续性考虑在实际部署中更具可操作性。
{"title":"GreenDFL: A framework for assessing the sustainability of Decentralized Federated Learning systems","authors":"Chao Feng ,&nbsp;Alberto Huertas Celdrán ,&nbsp;Xi Cheng ,&nbsp;Gérôme Bovet ,&nbsp;Burkhard Stiller","doi":"10.1016/j.infsof.2025.107937","DOIUrl":"10.1016/j.infsof.2025.107937","url":null,"abstract":"<div><h3>Context:</h3><div>Decentralized Federated Learning (DFL) is an emerging paradigm that enables collaborative model training without centralized data and model aggregation, enhancing privacy and resilience. However, its sustainability remains underexplored, as energy consumption and carbon emissions vary across different system configurations. Understanding the environmental impact of DFL is crucial for optimizing its design and deployment.</div></div><div><h3>Objective:</h3><div>This work aims to develop a comprehensive and operational framework for assessing the sustainability of DFL systems. To address it, this work provides a systematic method for quantifying energy consumption and carbon emissions, offering insights into improving the sustainability of DFL.</div></div><div><h3>Methods:</h3><div>This work proposes <em>GreenDFL</em>, a fully implementable framework that has been integrated into a real-world DFL platform. <em>GreenDFL</em> systematically analyzes the impact of various factors, including hardware accelerators, model architecture, communication medium, data distribution, network topology, and federation size, on the sustainability of DFL systems. Besides, a sustainability-aware aggregation algorithm (<em>GreenDFL-SA</em>) and a node selection algorithm (<em>GreenDFL-SN</em>) are developed to optimize energy efficiency and reduce carbon emissions in DFL training.</div></div><div><h3>Results:</h3><div>Empirical experiments are conducted on multiple datasets, measuring energy consumption and carbon emissions at different phases of the DFL lifecycle. Results indicate that local training dominates energy consumption and carbon emissions, while communication has a relatively minor impact. Optimizing model complexity, using GPUs instead of CPUs, and strategically selecting participating nodes significantly improve sustainability. Additionally, using wired communication, particularly optical fiber, effectively reduces energy consumption during the communication phase, while integrating early stopping mechanisms further minimizes overall emissions.</div></div><div><h3>Conclusion:</h3><div>The proposed <em>GreenDFL</em> provides a comprehensive and practical approach for assessing the sustainability of DFL systems. Furthermore, it offers best practices for improving environmental efficiency in DFL, making sustainability considerations more actionable in real-world deployments.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107937"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145371152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cross-project software defects prediction using fuzzy embedding and deep learning 基于模糊嵌入和深度学习的跨项目软件缺陷预测
IF 4.3 2区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS Pub Date : 2026-02-01 Epub Date: 2025-11-08 DOI: 10.1016/j.infsof.2025.107968
Mohammad Azzeh , Mohammad J. Abdel-Rahman

Context:

Cross-project defect prediction (CPDP) aims to predict software defects in a target project using data from related source projects, especially when defect data for the target project is limited or unavailable. A key challenge in CPDP is data heterogeneity and distributional differences across projects, which often lead to poor performance and unreliable predictions.

Objectives:

This study proposes a new CPDP model that improves prediction accuracy and robustness by introducing fuzzy embedding and deep learning to better capture similarities and differences across projects. The method is designed to mitigate mismatches in data distribution that hinder existing transfer learning and transformation-based approaches.

Methods:

The fuzzy embedding technique is built on fuzzy clustering and fuzzy set theory, which map each data point into a two-dimensional space of membership degrees across clusters. This representation models complex relationships with partial memberships and preserves contextual information that is often lost in conventional transformations. A deep learning model based on convolutional neural networks (CNN) processes the embedding matrices to learn discriminative defect patterns. The framework is evaluated against state-of-the-art CPDP models across multiple datasets, and sensitivity analysis is conducted on the number of clusters used in fuzzy embeddings.

Results:

Empirical evaluation shows that the proposed model consistently outperforms advanced CPDP approaches that rely on standard transformation or weighting methods. Improvements are observed in AUC and other key metrics across diverse datasets, demonstrating that fuzzy embeddings enhance the ability of deep learning to generalize knowledge across projects with varying characteristics.

Conclusion:

This work contributes a practical and effective solution for addressing heterogeneity in CPDP. By combining fuzzy embeddings with deep learning, the model not only achieves higher predictive accuracy but also improves reliability in real-world scenarios where project data distributions differ significantly. These findings highlight the potential of fuzzy embedding to support more resilient software quality assurance practices and provide actionable insights for practitioners dealing with limited or imbalanced defect data.
上下文:跨项目缺陷预测(CPDP)旨在使用来自相关源项目的数据来预测目标项目中的软件缺陷,特别是当目标项目的缺陷数据有限或不可用时。CPDP的一个关键挑战是项目之间的数据异质性和分布差异,这通常会导致较差的性能和不可靠的预测。本研究提出了一个新的CPDP模型,该模型通过引入模糊嵌入和深度学习来更好地捕捉项目之间的相似性和差异性,从而提高了预测的准确性和鲁棒性。该方法旨在减轻数据分布中的不匹配,这些不匹配阻碍了现有的迁移学习和基于转换的方法。方法:基于模糊聚类和模糊集理论建立模糊嵌入技术,将每个数据点映射到聚类间具有隶属度的二维空间中。这种表示对部分成员关系的复杂关系进行建模,并保留了在常规转换中经常丢失的上下文信息。基于卷积神经网络(CNN)的深度学习模型对嵌入矩阵进行处理以学习判别缺陷模式。该框架针对跨多个数据集的最先进的CPDP模型进行评估,并对模糊嵌入中使用的聚类数量进行敏感性分析。结果:实证评估表明,所提出的模型始终优于依赖于标准转换或加权方法的先进CPDP方法。在不同数据集的AUC和其他关键指标中观察到改进,表明模糊嵌入增强了深度学习在具有不同特征的项目中泛化知识的能力。结论:本研究为解决CPDP的异质性提供了一个切实有效的解决方案。通过将模糊嵌入与深度学习相结合,该模型不仅实现了更高的预测精度,而且在项目数据分布差异较大的现实场景中提高了可靠性。这些发现突出了模糊嵌入的潜力,以支持更有弹性的软件质量保证实践,并为处理有限或不平衡缺陷数据的从业者提供可操作的见解。
{"title":"Cross-project software defects prediction using fuzzy embedding and deep learning","authors":"Mohammad Azzeh ,&nbsp;Mohammad J. Abdel-Rahman","doi":"10.1016/j.infsof.2025.107968","DOIUrl":"10.1016/j.infsof.2025.107968","url":null,"abstract":"<div><h3>Context:</h3><div>Cross-project defect prediction (CPDP) aims to predict software defects in a target project using data from related source projects, especially when defect data for the target project is limited or unavailable. A key challenge in CPDP is data heterogeneity and distributional differences across projects, which often lead to poor performance and unreliable predictions.</div></div><div><h3>Objectives:</h3><div>This study proposes a new CPDP model that improves prediction accuracy and robustness by introducing fuzzy embedding and deep learning to better capture similarities and differences across projects. The method is designed to mitigate mismatches in data distribution that hinder existing transfer learning and transformation-based approaches.</div></div><div><h3>Methods:</h3><div>The fuzzy embedding technique is built on fuzzy clustering and fuzzy set theory, which map each data point into a two-dimensional space of membership degrees across clusters. This representation models complex relationships with partial memberships and preserves contextual information that is often lost in conventional transformations. A deep learning model based on convolutional neural networks (CNN) processes the embedding matrices to learn discriminative defect patterns. The framework is evaluated against state-of-the-art CPDP models across multiple datasets, and sensitivity analysis is conducted on the number of clusters used in fuzzy embeddings.</div></div><div><h3>Results:</h3><div>Empirical evaluation shows that the proposed model consistently outperforms advanced CPDP approaches that rely on standard transformation or weighting methods. Improvements are observed in AUC and other key metrics across diverse datasets, demonstrating that fuzzy embeddings enhance the ability of deep learning to generalize knowledge across projects with varying characteristics.</div></div><div><h3>Conclusion:</h3><div>This work contributes a practical and effective solution for addressing heterogeneity in CPDP. By combining fuzzy embeddings with deep learning, the model not only achieves higher predictive accuracy but also improves reliability in real-world scenarios where project data distributions differ significantly. These findings highlight the potential of fuzzy embedding to support more resilient software quality assurance practices and provide actionable insights for practitioners dealing with limited or imbalanced defect data.</div></div>","PeriodicalId":54983,"journal":{"name":"Information and Software Technology","volume":"190 ","pages":"Article 107968"},"PeriodicalIF":4.3,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145468021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Information and Software Technology
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1