首页 > 最新文献

Journal of Software-Evolution and Process最新文献

英文 中文
Evaluating the Test Adequacy of Benchmarks for LLMs on Code Generation 评估llm代码生成基准测试的充分性
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-25 DOI: 10.1002/smr.70034
Xiangyue Liu, Xiaobing Sun, Lili Bo, Yufei Hu, Xinwei Liu, Zhenlei Ye

Code generation for users' intent has become increasingly prevalent with the large language models (LLMs). To automatically evaluate the effectiveness of these models, multiple execution-based benchmarks are proposed, including specially crafted tasks, accompanied by some test cases and a ground truth solution. LLMs are regarded as well-performed in code generation tasks if they can pass the test cases corresponding to most tasks in these benchmarks. However, it is unknown whether the test cases have sufficient test adequacy and whether the test adequacy can affect the evaluation. In this paper, we conducted an empirical study to evaluate the test adequacy of the execution-based benchmarks and to explore their effects during evaluation for LLMs. Based on the evaluation of the widely used benchmarks, HumanEval, MBPP, and two enhanced benchmarks HumanEval+ and MBPP+, we obtained the following results: (1) All the evaluated benchmarks have high statement coverage (above 99.16%), low branch coverage (74.39%) and low mutation score (87.69%). Especially for the tasks with higher cyclomatic complexities in the HumanEval and MBPP, the mutation score of test cases is lower. (2) No significant correlation exists between test adequacy (statement coverage, branch coverage and mutation score) of benchmarks and evaluating results on LLMs at the individual task level. (3) There is a significant positive correlation between mutation score-based evaluation and another execution-based evaluation metric (AvgPassRatio$$ AvgPassRatio $$) on LLMs at the individual task level. (4) The existing test case augmentation techniques have limited improvement in the coverage of test cases in the benchmark, while significantly improving the mutation score by approximately 34.60% and also can bring a more rigorous evaluation to LLMs on code generation. (5) The LLM-based test case generation technique (EvalPlus) performs better than the traditional search-based technique (Pynguin) in improving the benchmarks' test quality and evaluation ability of code generation.

针对用户意图的代码生成在大型语言模型(llm)中变得越来越普遍。为了自动评估这些模型的有效性,提出了多个基于执行的基准,包括特别制作的任务,伴随着一些测试用例和一个基本的真实解决方案。如果llm能够通过与这些基准测试中的大多数任务相对应的测试用例,则认为它们在代码生成任务中表现良好。然而,测试用例是否具有足够的测试充分性,以及测试充分性是否会影响评估,这是未知的。在本文中,我们进行了一项实证研究,以评估基于执行的基准测试的充分性,并探讨其在llm评估中的作用。通过对目前广泛使用的基准HumanEval、MBPP以及两个增强基准HumanEval+和MBPP+的评价,得到以下结果:(1)评价的基准语句覆盖率均较高(99.16%以上),分支覆盖率较低(74.39%),突变评分较低(87.69%)。特别是对于HumanEval和MBPP中圈复杂度较高的任务,测试用例的突变分数较低。(2)基准的测试充分性(语句覆盖率、分支覆盖率和突变得分)与llm在个体任务水平上的评价结果不存在显著相关。(3)基于突变分数的评价与另一项基于执行力的评价指标(a vgPassRatio $$ AvgPassRatio $$)在llm个体任务水平上存在显著正相关。(4)现有的测试用例增强技术对基准测试用例覆盖率的提高有限,但显著提高了约34.60%的突变分数,也可以对llm在代码生成方面进行更严格的评估。(5)基于llm的测试用例生成技术(EvalPlus)在提高基准测试质量和代码生成的评估能力方面优于传统的基于搜索的技术(Pynguin)。
{"title":"Evaluating the Test Adequacy of Benchmarks for LLMs on Code Generation","authors":"Xiangyue Liu,&nbsp;Xiaobing Sun,&nbsp;Lili Bo,&nbsp;Yufei Hu,&nbsp;Xinwei Liu,&nbsp;Zhenlei Ye","doi":"10.1002/smr.70034","DOIUrl":"https://doi.org/10.1002/smr.70034","url":null,"abstract":"<div>\u0000 \u0000 <p>Code generation for users' intent has become increasingly prevalent with the large language models (LLMs). To automatically evaluate the effectiveness of these models, multiple execution-based benchmarks are proposed, including specially crafted tasks, accompanied by some test cases and a ground truth solution. LLMs are regarded as well-performed in code generation tasks if they can pass the test cases corresponding to most tasks in these benchmarks. However, it is unknown whether the test cases have sufficient test adequacy and whether the test adequacy can affect the evaluation. In this paper, we conducted an empirical study to evaluate the test adequacy of the execution-based benchmarks and to explore their effects during evaluation for LLMs. Based on the evaluation of the widely used benchmarks, HumanEval, MBPP, and two enhanced benchmarks HumanEval+ and MBPP+, we obtained the following results: (1) All the evaluated benchmarks have high statement coverage (above 99.16%), low branch coverage (74.39%) and low mutation score (87.69%). Especially for the tasks with higher cyclomatic complexities in the HumanEval and MBPP, the mutation score of test cases is lower. (2) No significant correlation exists between test adequacy (statement coverage, branch coverage and mutation score) of benchmarks and evaluating results on LLMs at the individual task level. (3) There is a significant positive correlation between mutation score-based evaluation and another execution-based evaluation metric (<span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mi>A</mi>\u0000 <mi>v</mi>\u0000 <mi>g</mi>\u0000 <mi>P</mi>\u0000 <mi>a</mi>\u0000 <mi>s</mi>\u0000 <mi>s</mi>\u0000 <mi>R</mi>\u0000 <mi>a</mi>\u0000 <mi>t</mi>\u0000 <mi>i</mi>\u0000 <mi>o</mi>\u0000 </mrow>\u0000 <annotation>$$ AvgPassRatio $$</annotation>\u0000 </semantics></math>) on LLMs at the individual task level. (4) The existing test case augmentation techniques have limited improvement in the coverage of test cases in the benchmark, while significantly improving the mutation score by approximately 34.60% and also can bring a more rigorous evaluation to LLMs on code generation. (5) The LLM-based test case generation technique (EvalPlus) performs better than the traditional search-based technique (Pynguin) in improving the benchmarks' test quality and evaluation ability of code generation.</p>\u0000 </div>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 7","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144482208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring the Effectiveness of Open-Source Donation Platform: An Empirical Study on Opencollective 探索开源捐赠平台的有效性:基于Opencollective的实证研究
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-25 DOI: 10.1002/smr.70033
Shuoxiao Zhang, Enyi Tang, Xinyu Gao, Zhekai Zhang, Yixiao Shan, Haofeng Zhang, Ziyang He, Jianhua Zhao, Xuandong Li

In recent years, with the development of the open-source community, various open-source donation platforms have emerged. These platforms effectively alleviate the financial pressures faced by open-source projects through diversified funding sources and flexible donation methods. As one of the most representative open-source donation platforms, Opencollective has garnered widespread attention from both the open-source community and academia. Although Opencollective claims to provide more funding opportunities for open-source projects, the extent to which it effectively addresses the financial challenges faced by these projects remains unclear. While there have been studies on the effectiveness of traditional donation models, research on the effectiveness of emerging donation platforms such as Opencollective is still limited. Given that a large number of open-source projects are urgently seeking donations, understanding the effectiveness of donations through Opencollective is crucial for these projects. To address this gap, we have made an early step in this direction. This paper conducts a comprehensive study on the effectiveness of donations through the Opencollective, employing a combination of quantitative and qualitative analysis and identifies the following key findings: (1) Opencollective attracts a diverse group of participants, including individual donors, sponsors, contributors, and project managers, with individual donors constituting the largest group. Most donations are concentrated in the range of $5 to $10, indicating that the platform largely relies on small but frequent donations from individuals. (2) Only about 26.61% of open-source projects receive donations through Opencollective, with approximately 64.38% of these projects receiving a total donation amount of less than $50,000. The likelihood of receiving donations increases with project scale, maturity and the number of stars. Among projects that have received donations, larger projects with stronger social media promotion, greater attention and more issues are more likely to receive additional donations. (3) The positive impact of donations on project development and spend activities is significant only in the short term, with no notable long-term effects. In contrast, donations do not have a significant short-term impact on community engagement. Although the long-term effect is slightly positive, it is not statistically significant. (4) The main shortcomings of Opencollective include insufficient project management and collaboration features, inadequate user experience and interface design, high transaction fees, and a lack of transparency in fund allocation and usage. Our findings provide significant theoretical support and practical recommendations for the effectiveness of emerging donation platforms and the sustainable development of open-source projects.

近年来,随着开源社区的发展,各种开源捐赠平台应运而生。这些平台通过多元化的资金来源和灵活的捐赠方式,有效缓解了开源项目面临的资金压力。Opencollective作为最具代表性的开源捐赠平台之一,受到了开源社区和学术界的广泛关注。尽管Opencollective声称为开源项目提供更多的融资机会,但它在多大程度上有效地解决了这些项目面临的财务挑战仍不清楚。虽然对传统捐赠模式的有效性进行了研究,但对Opencollective等新兴捐赠平台的有效性研究仍然有限。鉴于大量的开源项目迫切需要捐赠,了解通过Opencollective捐赠的有效性对这些项目至关重要。为了弥补这一差距,我们已经朝这个方向迈出了早期的一步。本文采用定量分析和定性分析相结合的方法,对Opencollective的捐赠有效性进行了全面研究,得出以下主要结论:(1)Opencollective吸引了包括个人捐赠者、发起人、贡献者和项目经理在内的多元化参与者群体,其中个人捐赠者构成了最大的群体。大多数捐款集中在5美元到10美元之间,这表明该平台在很大程度上依赖于小额但频繁的个人捐款。(2)只有约26.61%的开源项目通过Opencollective获得捐赠,其中约64.38%的项目获得的总捐赠金额低于5万美元。接受捐赠的可能性随着项目规模、成熟度和星级的增加而增加。在接受捐赠的项目中,更大的项目,社交媒体推广力度更大,关注度更高,问题更多,更有可能获得额外的捐赠。(3)捐赠对项目开发和支出活动的积极影响仅在短期内显著,长期影响不显著。相比之下,捐赠对社区参与没有显著的短期影响。虽然长期效果略显积极,但在统计学上并不显著。(4) Opencollective的主要缺点是项目管理和协作功能不足,用户体验和界面设计不完善,交易费用高,资金分配和使用缺乏透明度。研究结果为新兴捐赠平台的有效性和开源项目的可持续发展提供了重要的理论支持和实践建议。
{"title":"Exploring the Effectiveness of Open-Source Donation Platform: An Empirical Study on Opencollective","authors":"Shuoxiao Zhang,&nbsp;Enyi Tang,&nbsp;Xinyu Gao,&nbsp;Zhekai Zhang,&nbsp;Yixiao Shan,&nbsp;Haofeng Zhang,&nbsp;Ziyang He,&nbsp;Jianhua Zhao,&nbsp;Xuandong Li","doi":"10.1002/smr.70033","DOIUrl":"https://doi.org/10.1002/smr.70033","url":null,"abstract":"<div>\u0000 \u0000 <p>In recent years, with the development of the open-source community, various open-source donation platforms have emerged. These platforms effectively alleviate the financial pressures faced by open-source projects through diversified funding sources and flexible donation methods. As one of the most representative open-source donation platforms, Opencollective has garnered widespread attention from both the open-source community and academia. Although Opencollective claims to provide more funding opportunities for open-source projects, the extent to which it effectively addresses the financial challenges faced by these projects remains unclear. While there have been studies on the effectiveness of traditional donation models, research on the effectiveness of emerging donation platforms such as Opencollective is still limited. Given that a large number of open-source projects are urgently seeking donations, understanding the effectiveness of donations through Opencollective is crucial for these projects. To address this gap, we have made an early step in this direction. This paper conducts a comprehensive study on the effectiveness of donations through the Opencollective, employing a combination of quantitative and qualitative analysis and identifies the following key findings: (1) Opencollective attracts a diverse group of participants, including individual donors, sponsors, contributors, and project managers, with individual donors constituting the largest group. Most donations are concentrated in the range of $5 to $10, indicating that the platform largely relies on small but frequent donations from individuals. (2) Only about 26.61% of open-source projects receive donations through Opencollective, with approximately 64.38% of these projects receiving a total donation amount of less than $50,000. The likelihood of receiving donations increases with project scale, maturity and the number of stars. Among projects that have received donations, larger projects with stronger social media promotion, greater attention and more issues are more likely to receive additional donations. (3) The positive impact of donations on project development and spend activities is significant only in the short term, with no notable long-term effects. In contrast, donations do not have a significant short-term impact on community engagement. Although the long-term effect is slightly positive, it is not statistically significant. (4) The main shortcomings of Opencollective include insufficient project management and collaboration features, inadequate user experience and interface design, high transaction fees, and a lack of transparency in fund allocation and usage. Our findings provide significant theoretical support and practical recommendations for the effectiveness of emerging donation platforms and the sustainable development of open-source projects.</p>\u0000 </div>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 7","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144482206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating Security Controls in DevSecOps: Challenges, Solutions, and Future Research Directions 在DevSecOps中集成安全控制:挑战、解决方案和未来研究方向
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-06-09 DOI: 10.1002/smr.70029
Maysa Sinan, Mojtaba Shahin, Iqbal Gondal

Cybersecurity has become a top priority for most organizations to protect their applications. The rapid increase in cyberattacks has necessitated a comprehensive repositioning of how security should be implemented within the software development lifecycle (SDLC). Development, Security, Operations (DevSecOps) is one of the trendy security methodologies and fastest growing development methods promoting shared responsibility for security and automating security practices at every step of the SDLC. DevSecOps is a cultural shift that integrates security controls into DevOps pipelines aiming to upscale overall security. Therefore, many organizations started to incorporate security controls within the deployment of DevSecOps through conducting continuous practices, for example, automated security testing, infrastructure as code (IaC), compliance as code, and continuous monitoring. This study aims to organize the knowledge and shed light on challenges concerning security controls during the adoption of DevSecOps, along with associated solutions and remediation workarounds reported in the literature. Further, the study aims to provide clear insights into the areas that require further investigation and research in the future. A systematic literature review (SLR) of 45 primary studies was carried out to extract data, and subsequently, the extracted data was analyzed using the thematic analysis method. This paper identifies 19 challenges related to security controls that could be experienced by security practitioners while implementing a DevSecOps model, along with 18 solutions and remediation actions suggested in literature to address and overcome some of the enlisted challenges. In addition, some gap areas are identified as opportunities for future research in this domain with the aim of improving the integration of security controls in a DevSecOps environment. Based on findings, this paper points out the importance of automation in software engineering practices, for example, continuous automation, continuous delivery, and continuous feedback, to embed security controls at the early stages of the development process.

网络安全已成为大多数组织保护其应用程序的首要任务。网络攻击的快速增加已经需要对软件开发生命周期(SDLC)中如何实现安全性进行全面的重新定位。开发、安全、运营(DevSecOps)是一种流行的安全方法,也是发展最快的开发方法之一,它促进了安全责任共享,并在SDLC的每一步自动化安全实践。DevSecOps是一种文化转变,它将安全控制集成到DevOps管道中,旨在提升整体安全性。因此,许多组织开始在DevSecOps的部署中加入安全控制,通过执行连续的实践,例如,自动化安全测试、基础设施即代码(IaC)、法规遵从性即代码,以及持续的监控。本研究旨在组织知识,并阐明在采用DevSecOps期间有关安全控制的挑战,以及文献中报道的相关解决方案和补救方法。此外,该研究旨在为未来需要进一步调查和研究的领域提供清晰的见解。通过系统文献综述(SLR)对45项初步研究进行数据提取,随后采用专题分析法对提取的数据进行分析。本文确定了安全从业人员在实现DevSecOps模型时可能遇到的19个与安全控制相关的挑战,以及文献中建议的18个解决方案和补救措施,以解决和克服所列出的一些挑战。此外,一些空白区域被确定为该领域未来研究的机会,目的是改进DevSecOps环境中安全控制的集成。基于这些发现,本文指出了自动化在软件工程实践中的重要性,例如,持续的自动化,持续的交付,以及持续的反馈,在开发过程的早期阶段嵌入安全控制。
{"title":"Integrating Security Controls in DevSecOps: Challenges, Solutions, and Future Research Directions","authors":"Maysa Sinan,&nbsp;Mojtaba Shahin,&nbsp;Iqbal Gondal","doi":"10.1002/smr.70029","DOIUrl":"https://doi.org/10.1002/smr.70029","url":null,"abstract":"<p>Cybersecurity has become a top priority for most organizations to protect their applications. The rapid increase in cyberattacks has necessitated a comprehensive repositioning of how security should be implemented within the software development lifecycle (SDLC). Development, Security, Operations (DevSecOps) is one of the trendy security methodologies and fastest growing development methods promoting shared responsibility for security and automating security practices at every step of the SDLC. DevSecOps is a cultural shift that integrates security controls into DevOps pipelines aiming to upscale overall security. Therefore, many organizations started to incorporate security controls within the deployment of DevSecOps through conducting continuous practices, for example, automated security testing, infrastructure as code (IaC), compliance as code, and continuous monitoring. This study aims to organize the knowledge and shed light on challenges concerning security controls during the adoption of DevSecOps, along with associated solutions and remediation workarounds reported in the literature. Further, the study aims to provide clear insights into the areas that require further investigation and research in the future. A systematic literature review (SLR) of 45 primary studies was carried out to extract data, and subsequently, the extracted data was analyzed using the thematic analysis method. This paper identifies 19 challenges related to security controls that could be experienced by security practitioners while implementing a DevSecOps model, along with 18 solutions and remediation actions suggested in literature to address and overcome some of the enlisted challenges. In addition, some gap areas are identified as opportunities for future research in this domain with the aim of improving the integration of security controls in a DevSecOps environment. Based on findings, this paper points out the importance of automation in software engineering practices, for example, continuous automation, continuous delivery, and continuous feedback, to embed security controls at the early stages of the development process.</p>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 6","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/smr.70029","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144244310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improvement of Software Testing Processes With Test Maturity Model Integration 基于测试成熟度模型集成的软件测试过程改进
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-22 DOI: 10.1002/smr.70031
Gökhan Şit, Süleyman Ersöz, Mehmet Burak Bilgin

In this study, a maturity-level determination and assessment method developed for companies operating in the software industry to perform TMMi Levels 2 and 3 assessments in-house, with the goal of improving testing processes, is presented. With this method, it is aimed to help companies to conduct their own self-assessments and improve their testing processes before participating in high-budget audits. The validity of this method was tested in practice for TMMi Levels 2 and 3 assessments. Companies can prepare for a formal TMMi audit by using the test maturity-level determination methodology developed in this study, or they can simply improve their processes to produce higher quality products. Additionally, if they already have TMMi certification, they can regularly self-audit to ensure continuity and compliance. Demonstrating how TMMi can be applied in practice and providing a guide to test process maturity determination are key contributions of this work.

在本研究中,提出了一种成熟度级别的确定和评估方法,该方法是为在软件行业中运营的公司开发的,用于在内部执行TMMi级别2和3的评估,其目标是改进测试过程。通过这种方法,它旨在帮助公司在参与高预算审计之前进行自己的自我评估并改进他们的测试流程。在TMMi 2级和3级评估的实践中检验了该方法的有效性。公司可以通过使用本研究中开发的测试成熟度级别确定方法来准备正式的TMMi审核,或者他们可以简单地改进他们的过程以生产更高质量的产品。此外,如果他们已经拥有TMMi认证,他们可以定期进行自我审核,以确保连续性和合规性。演示TMMi如何在实践中应用,并为测试过程成熟度确定提供指南是这项工作的关键贡献。
{"title":"Improvement of Software Testing Processes With Test Maturity Model Integration","authors":"Gökhan Şit,&nbsp;Süleyman Ersöz,&nbsp;Mehmet Burak Bilgin","doi":"10.1002/smr.70031","DOIUrl":"https://doi.org/10.1002/smr.70031","url":null,"abstract":"<div>\u0000 \u0000 <p>In this study, a maturity-level determination and assessment method developed for companies operating in the software industry to perform TMMi Levels 2 and 3 assessments in-house, with the goal of improving testing processes, is presented. With this method, it is aimed to help companies to conduct their own self-assessments and improve their testing processes before participating in high-budget audits. The validity of this method was tested in practice for TMMi Levels 2 and 3 assessments. Companies can prepare for a formal TMMi audit by using the test maturity-level determination methodology developed in this study, or they can simply improve their processes to produce higher quality products. Additionally, if they already have TMMi certification, they can regularly self-audit to ensure continuity and compliance. Demonstrating how TMMi can be applied in practice and providing a guide to test process maturity determination are key contributions of this work.</p>\u0000 </div>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144117832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Functional Size Measurement With Conceptual Models: A Systematic Literature Review 概念模型的功能大小测量:系统文献综述
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-21 DOI: 10.1002/smr.70030
Ala Arman, Emiliano Di Reto, Massimo Mecella, Giuseppe Santucci
<p>The demand for efficient functional size measurement (FSM) methods in the competitive software market today is undeniable. However, incomplete and imprecise system specifications pose significant challenges, particularly in scenarios that require fast, flexible, and accurate software size estimation, such as public tenders. Although the integration of conceptual models within FSMs offers a promising solution to these issues, a systematic exploration of such methods remains largely unexplored. This work evaluates FSM methods that integrate conceptual models by analyzing studies from the past 20 years. It highlights key contributions and advances in proposed conceptual model-based FSM methods. In addition, the study examines their limitations and challenges, offering insights for future improvements. A systematic literature review (SLR) was conducted to guide the research process. The review was organized around three research questions, each targeting the study's key objectives: (1) to explore FSM methods utilizing conceptual models, (2) to summarize proposals for their improvement, and (3) to identify the limitations of the proposed enhancements. Primary studies span two decades (2004–2024), with peaks in 2008 and 2015, averaging one to two studies annually. Of the 1371 initial studies, 13 were selected using strict criteria. These studies are categorized into <i>Measurement Techniques</i> (30.77%), <i>Automation</i> (38.46%), and <i>Application-Specific</i> topics (30.77%). The contributions of the primary studies are analyzed in terms of their approaches <i>Repeatability</i> and <i>Validation</i>. <i>Repeatability</i> is assessed by examining whether the primary studies proposed a formal model when using real datasets. In contrast, <i>Validation</i> focuses on whether the studies were tested in real-world projects. A total of 46.15% of the primary studies utilize formal models, whereas 53.85% rely on nonformal models, although dataset size is often unspecified. Most studies validate their methods using 1 to 30 projects. Common Software Measurement International Consortium (COSMIC) is the most widely used FSM method (69.23%), followed by the Function Point Analysis (FPA) (15.38%) and custom Methods (15.38%), with conceptual UML models appearing in 84.61% of the studies. Key limitations, including <i>Scalability and Generalizability</i>, <i>Complexity Robustness</i>, and <i>Flexibility</i>, persist across all categories. Notably, <i>Scalability and Generalizability</i> was identified as a limitation in 75% of <i>Measurement Techniques</i> studies, 80% of <i>Automation</i> studies, and 75% of <i>Application-Specific</i> studies, while <i>Flexibility</i> challenges were most pronounced, affecting 100% of <i>Application-Specific</i> studies. The limited number of primary studies underscores a substantial research gap in conceptual model-based FSM methods. Future research should focus on developing formal models to enhance theoretical rigor, lever
在当今竞争激烈的软件市场中,对高效功能大小度量(FSM)方法的需求是不可否认的。然而,不完整和不精确的系统规范带来了重大的挑战,特别是在需要快速、灵活和准确的软件大小估计的场景中,例如公开招标。尽管在fsm中整合概念模型为这些问题提供了一个有希望的解决方案,但对这些方法的系统探索在很大程度上仍未被探索。本工作通过分析过去20年的研究来评估FSM方法整合概念模型。它强调了提出的基于概念模型的FSM方法的关键贡献和进展。此外,该研究还考察了它们的局限性和挑战,为未来的改进提供了见解。通过系统文献综述(SLR)来指导研究过程。该综述围绕三个研究问题进行组织,每个问题都针对研究的关键目标:(1)利用概念模型探索FSM方法,(2)总结改进建议,(3)确定建议增强的局限性。主要研究跨越二十年(2004-2024),2008年和2015年达到高峰,平均每年一到两项研究。在1371项初步研究中,有13项是根据严格的标准选出的。这些研究分为测量技术(30.77%)、自动化(38.46%)和特定应用主题(30.77%)。从方法、可重复性和验证性方面分析了初步研究的贡献。通过检查主要研究是否在使用真实数据集时提出正式模型来评估可重复性。相比之下,验证侧重于研究是否在现实世界的项目中进行了测试。总共46.15%的初步研究使用正式模型,而53.85%依赖于非正式模型,尽管数据集大小通常未指定。大多数研究用1到30个项目来验证他们的方法。通用软件度量国际联盟(COSMIC)是使用最广泛的FSM方法(69.23%),其次是功能点分析(FPA)(15.38%)和定制方法(15.38%),概念UML模型出现在84.61%的研究中。关键的限制,包括可伸缩性和泛化性、复杂性、健壮性和灵活性,在所有类别中都存在。值得注意的是,在75%的测量技术研究、80%的自动化研究和75%的特定应用研究中,可扩展性和通用性被认为是一个限制,而灵活性挑战是最明显的,影响了100%的特定应用研究。有限数量的初步研究强调了基于概念模型的FSM方法的实质性研究差距。未来的研究应侧重于开发正式模型以增强理论严谨性,利用真实世界的数据集进行验证,提供全面的方法描述,并标准化验证实践。此外,通过提高可伸缩性、泛化性和灵活性来优先考虑FSM方法的进步是至关重要的。这些增强将使FSM方法能够有效地管理复杂的系统,适应不同的软件领域,并解决特定于应用程序的需求,确保它们在动态和不断发展的软件开发环境中的持续相关性。
{"title":"Functional Size Measurement With Conceptual Models: A Systematic Literature Review","authors":"Ala Arman,&nbsp;Emiliano Di Reto,&nbsp;Massimo Mecella,&nbsp;Giuseppe Santucci","doi":"10.1002/smr.70030","DOIUrl":"https://doi.org/10.1002/smr.70030","url":null,"abstract":"&lt;p&gt;The demand for efficient functional size measurement (FSM) methods in the competitive software market today is undeniable. However, incomplete and imprecise system specifications pose significant challenges, particularly in scenarios that require fast, flexible, and accurate software size estimation, such as public tenders. Although the integration of conceptual models within FSMs offers a promising solution to these issues, a systematic exploration of such methods remains largely unexplored. This work evaluates FSM methods that integrate conceptual models by analyzing studies from the past 20 years. It highlights key contributions and advances in proposed conceptual model-based FSM methods. In addition, the study examines their limitations and challenges, offering insights for future improvements. A systematic literature review (SLR) was conducted to guide the research process. The review was organized around three research questions, each targeting the study's key objectives: (1) to explore FSM methods utilizing conceptual models, (2) to summarize proposals for their improvement, and (3) to identify the limitations of the proposed enhancements. Primary studies span two decades (2004–2024), with peaks in 2008 and 2015, averaging one to two studies annually. Of the 1371 initial studies, 13 were selected using strict criteria. These studies are categorized into &lt;i&gt;Measurement Techniques&lt;/i&gt; (30.77%), &lt;i&gt;Automation&lt;/i&gt; (38.46%), and &lt;i&gt;Application-Specific&lt;/i&gt; topics (30.77%). The contributions of the primary studies are analyzed in terms of their approaches &lt;i&gt;Repeatability&lt;/i&gt; and &lt;i&gt;Validation&lt;/i&gt;. &lt;i&gt;Repeatability&lt;/i&gt; is assessed by examining whether the primary studies proposed a formal model when using real datasets. In contrast, &lt;i&gt;Validation&lt;/i&gt; focuses on whether the studies were tested in real-world projects. A total of 46.15% of the primary studies utilize formal models, whereas 53.85% rely on nonformal models, although dataset size is often unspecified. Most studies validate their methods using 1 to 30 projects. Common Software Measurement International Consortium (COSMIC) is the most widely used FSM method (69.23%), followed by the Function Point Analysis (FPA) (15.38%) and custom Methods (15.38%), with conceptual UML models appearing in 84.61% of the studies. Key limitations, including &lt;i&gt;Scalability and Generalizability&lt;/i&gt;, &lt;i&gt;Complexity Robustness&lt;/i&gt;, and &lt;i&gt;Flexibility&lt;/i&gt;, persist across all categories. Notably, &lt;i&gt;Scalability and Generalizability&lt;/i&gt; was identified as a limitation in 75% of &lt;i&gt;Measurement Techniques&lt;/i&gt; studies, 80% of &lt;i&gt;Automation&lt;/i&gt; studies, and 75% of &lt;i&gt;Application-Specific&lt;/i&gt; studies, while &lt;i&gt;Flexibility&lt;/i&gt; challenges were most pronounced, affecting 100% of &lt;i&gt;Application-Specific&lt;/i&gt; studies. The limited number of primary studies underscores a substantial research gap in conceptual model-based FSM methods. Future research should focus on developing formal models to enhance theoretical rigor, lever","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/smr.70030","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144108838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Overcoming Data Shortage in Critical Domains With Data Augmentation for Natural Language Software Requirements 基于自然语言软件需求的数据扩充克服关键领域的数据短缺
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-12 DOI: 10.1002/smr.70027
Robin Korfmann, Patrick Beyersdorffer, Rainer Gerlich, Jürgen Münch, Marco Kuhrmann

Natural language processing (NLP) offers the potential to automate quality assurance of software requirement specifications. In particular, large-scale projects involving numerous suppliers can benefit from this improvement. However, due to privacy restrictions especially in highly restrictive industries, the availability of software requirements specification documents for training NLP tools is severely limited. Also, domain- and project-specific vocabulary, as such in the aerospace domain, require specialized models for processing effectively. To provide a sufficient amount of data to train such models, we studied algorithms for the augmentation of textual data. Four algorithms have been investigated by expanding a given set of requirements from the European Space projects generating correct and incorrect requirements. The initial study yielded data of poor quality due to the particularities of the domain-specific vocabulary, yet laid the foundation for the algorithms' improvement, which, eventually, resulted in an increased set of requirements, which is 20 times the size of the seed set. A complementing experiment demonstrated the usability of augmented requirements to support AI-based quality assurance of software requirements. Furthermore, a selected improvement of the augmentation algorithms demonstrated notable quality improvements by doubling the number of correctly augmented requirements.

自然语言处理(NLP)提供了自动化软件需求规范质量保证的潜力。特别是涉及众多供应商的大型项目可以从这种改进中受益。然而,由于隐私限制,特别是在高度限制的行业中,培训NLP工具的软件需求规范文档的可用性受到严重限制。此外,特定于领域和项目的词汇表,如航空航天领域中的词汇表,需要专门的模型来进行有效的处理。为了提供足够的数据量来训练这样的模型,我们研究了文本数据增强的算法。通过扩展来自欧洲空间项目的一组给定需求,研究了四种算法,生成了正确和不正确的需求。由于领域特定词汇表的特殊性,最初的研究产生的数据质量较差,但为算法的改进奠定了基础,最终导致需求集的增加,这是种子集大小的20倍。一个补充实验证明了增强需求的可用性,以支持基于人工智能的软件需求质量保证。此外,通过将正确增强的需求数量增加一倍,对增强算法的选择改进显示出显著的质量改进。
{"title":"Overcoming Data Shortage in Critical Domains With Data Augmentation for Natural Language Software Requirements","authors":"Robin Korfmann,&nbsp;Patrick Beyersdorffer,&nbsp;Rainer Gerlich,&nbsp;Jürgen Münch,&nbsp;Marco Kuhrmann","doi":"10.1002/smr.70027","DOIUrl":"https://doi.org/10.1002/smr.70027","url":null,"abstract":"<p>Natural language processing (NLP) offers the potential to automate quality assurance of software requirement specifications. In particular, large-scale projects involving numerous suppliers can benefit from this improvement. However, due to privacy restrictions especially in highly restrictive industries, the availability of software requirements specification documents for training NLP tools is severely limited. Also, domain- and project-specific vocabulary, as such in the aerospace domain, require specialized models for processing effectively. To provide a sufficient amount of data to train such models, we studied algorithms for the augmentation of textual data. Four algorithms have been investigated by expanding a given set of requirements from the European Space projects generating correct and incorrect requirements. The initial study yielded data of poor quality due to the particularities of the domain-specific vocabulary, yet laid the foundation for the algorithms' improvement, which, eventually, resulted in an increased set of requirements, which is 20 times the size of the seed set. A complementing experiment demonstrated the usability of augmented requirements to support AI-based quality assurance of software requirements. Furthermore, a selected improvement of the augmentation algorithms demonstrated notable quality improvements by doubling the number of correctly augmented requirements.</p>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/smr.70027","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143939432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Approach Based on Metadata to Implement Convention Over Configuration Decoupled From Framework Logic 基于元数据实现与框架逻辑解耦的约定优于配置的方法
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-05-10 DOI: 10.1002/smr.70028
Everaldo Gomes, Eduardo Guerra, Phyllipe Lima, Paulo Meirelles

Frameworks are essential for software development, providing code design and facilitating reuse for their users. Well-known Java frameworks and APIs rely on metadata configuration through code annotations, using Reflection API to consume and process them. Code elements that share the same annotations often exhibit similarities, creating the opportunity to use conventions as a metadata source. This paper proposes a model for defining Convention over Configuration (CoC) for annotation usage, decoupled from the metadata reading logic. With this model, if a convention is present, the framework will automatically consider that element to be annotated. We implemented this model in the Esfinge Metadata API and evaluated it in an experiment where participants implemented the CoC pattern using two approaches: our proposed one and the Java Reflection API. As a result, 75% of participants implemented our approach faster than with just the Reflection API, and we observed a higher failure rate with the Reflection API than with the Esfinge API. Moreover, the code produced with our approach also resulted in fewer lines of code. Based on these results, we confirmed that the proposed approach fulfilled its goal of supporting the definition of conventions decoupled from the framework logic, thereby improving code readability and maintainability.

框架对于软件开发至关重要,它为用户提供代码设计并促进重用。众所周知的Java框架和API依赖于通过代码注释配置元数据,并使用反射API来消费和处理它们。共享相同注释的代码元素通常表现出相似性,从而创造了使用约定作为元数据源的机会。本文提出了一个定义用于注释使用的约定优于配置(CoC)的模型,该模型与元数据读取逻辑解耦。使用此模型,如果存在约定,框架将自动认为该元素需要注释。我们在esfinger元数据API中实现了这个模型,并在一个实验中对其进行了评估,在这个实验中,参与者使用两种方法实现了CoC模式:我们提出的方法和Java反射API。结果,75%的参与者实现我们的方法比只使用反射API更快,我们观察到反射API比使用esfinger API的失败率更高。此外,用我们的方法生成的代码也减少了代码行数。基于这些结果,我们证实了所提出的方法实现了支持从框架逻辑解耦的约定定义的目标,从而提高了代码的可读性和可维护性。
{"title":"An Approach Based on Metadata to Implement Convention Over Configuration Decoupled From Framework Logic","authors":"Everaldo Gomes,&nbsp;Eduardo Guerra,&nbsp;Phyllipe Lima,&nbsp;Paulo Meirelles","doi":"10.1002/smr.70028","DOIUrl":"https://doi.org/10.1002/smr.70028","url":null,"abstract":"<p>Frameworks are essential for software development, providing code design and facilitating reuse for their users. Well-known Java frameworks and APIs rely on metadata configuration through code annotations, using Reflection API to consume and process them. Code elements that share the same annotations often exhibit similarities, creating the opportunity to use conventions as a metadata source. This paper proposes a model for defining Convention over Configuration (CoC) for annotation usage, decoupled from the metadata reading logic. With this model, if a convention is present, the framework will automatically consider that element to be annotated. We implemented this model in the Esfinge Metadata API and evaluated it in an experiment where participants implemented the CoC pattern using two approaches: our proposed one and the Java Reflection API. As a result, 75% of participants implemented our approach faster than with just the Reflection API, and we observed a higher failure rate with the Reflection API than with the Esfinge API. Moreover, the code produced with our approach also resulted in fewer lines of code. Based on these results, we confirmed that the proposed approach fulfilled its goal of supporting the definition of conventions decoupled from the framework logic, thereby improving code readability and maintainability.</p>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1002/smr.70028","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143930498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Novel Vulnerability-Detection Method Based on the Semantic Features of Source Code and the LLVM Intermediate Representation 基于源代码语义特征和LLVM中间表示的漏洞检测新方法
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-04-29 DOI: 10.1002/smr.70026
Jinfu Chen, Jiapeng Zhou, Wei Lin, Dave Towey, Saihua Cai, Haibo Chen, Jingyi Chen, Yemin Yin

With the increasingly frequent attacks on software systems, software security is an issue that must be addressed. Within software security, automated detection of software vulnerabilities is an important subject. Most existing vulnerability detectors rely on the features of a single code type (e.g., source code or intermediate representation [IR]), which may lead to both the global features of the code slices and the memory operation information not being captured or considered. In particular, vulnerability detection based on source-code features cannot usually include some macro or type definition content. In this paper, we propose a vulnerability-detection method that combines the semantic features of source code and the low level virtual machine (LLVM) IR. Our proposed approach starts by slicing (C/C++) source files using improved slicing techniques to cover more comprehensive code information. It then extracts semantic information from the LLVM IR based on the executable source code. This can enrich the features fed to the artificial neural network (ANN) model for learning. We conducted an experimental evaluation using a publicly-available dataset of 11,381 C/C++ programs. The experimental results show the vulnerability-detection accuracy of our proposed method to reach over 96% for code slices generated according to four different slicing criteria. This outperforms most other compared detection methods.

随着对软件系统的攻击日益频繁,软件安全是一个必须解决的问题。在软件安全中,软件漏洞的自动检测是一个重要的课题。大多数现有漏洞检测器依赖于单一代码类型的特征(例如,源代码或中间表示[IR]),这可能导致代码片的全局特征和内存操作信息都没有被捕获或考虑。特别是,基于源代码特性的漏洞检测通常不能包含一些宏或类型定义内容。本文提出了一种结合源代码语义特征和低级虚拟机(LLVM) IR的漏洞检测方法。我们提出的方法首先使用改进的切片技术对(C/ c++)源文件进行切片,以覆盖更全面的代码信息。然后,它根据可执行源代码从LLVM IR中提取语义信息。这可以丰富特征馈送到人工神经网络(ANN)模型进行学习。我们使用11,381个C/ c++程序的公开数据集进行了实验评估。实验结果表明,根据四种不同的切片标准生成的代码切片,本文方法的漏洞检测准确率达到96%以上。这优于大多数其他比较检测方法。
{"title":"A Novel Vulnerability-Detection Method Based on the Semantic Features of Source Code and the LLVM Intermediate Representation","authors":"Jinfu Chen,&nbsp;Jiapeng Zhou,&nbsp;Wei Lin,&nbsp;Dave Towey,&nbsp;Saihua Cai,&nbsp;Haibo Chen,&nbsp;Jingyi Chen,&nbsp;Yemin Yin","doi":"10.1002/smr.70026","DOIUrl":"https://doi.org/10.1002/smr.70026","url":null,"abstract":"<div>\u0000 \u0000 <p>With the increasingly frequent attacks on software systems, software security is an issue that must be addressed. Within software security, automated detection of software vulnerabilities is an important subject. Most existing vulnerability detectors rely on the features of a single code type (e.g., source code or intermediate representation [IR]), which may lead to both the global features of the code slices and the memory operation information not being captured or considered. In particular, vulnerability detection based on source-code features cannot usually include some macro or type definition content. In this paper, we propose a vulnerability-detection method that combines the semantic features of source code and the low level virtual machine (LLVM) IR. Our proposed approach starts by slicing (C/C++) source files using improved slicing techniques to cover more comprehensive code information. It then extracts semantic information from the LLVM IR based on the executable source code. This can enrich the features fed to the artificial neural network (ANN) model for learning. We conducted an experimental evaluation using a publicly-available dataset of 11,381 C/C++ programs. The experimental results show the vulnerability-detection accuracy of our proposed method to reach over 96% for code slices generated according to four different slicing criteria. This outperforms most other compared detection methods.</p>\u0000 </div>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 5","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143888815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ECP: Coprocessor Architecture to Protect Program Logic Consistency 保护程序逻辑一致性的协处理器架构
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-04-23 DOI: 10.1002/smr.70023
Yang Gao, Siqi Lu, Yongjuan Wang, Haopeng Fan, Qingdi Han, Jingsheng Li

Contemporary program protection methods focus on safeguarding either program generation, storage, or execution; however, no unified protection strategy exists for ensuring the security of a full program lifecycle. In this study, we combine the static security of program generation with the dynamic security of process execution and propose a novel program logic consistency security property. An encryption core processing (ECP) architecture is presented that provides coprocessor solutions to protect the program logic consistency at the granularity of instructions and data flows. The new authenticated encryption mode in the architecture uses the offset value of the program's instructions and data in relation to the segment-based address as its encryption parameters. Lightweight cryptographic primitives are adopted to ensure that the hardware burden added by the ECP is limited, especially under ×$$ times $$64 architectures. We prove that the proposed scheme in the ECP architecture satisfies indistinguishability under chosen plaintext attack and demonstrate the effectiveness of the architecture against various attacks. Additionally, a theoretical performance analysis is provided for estimating the overhead introduced by the ECP architecture.

当代程序保护方法侧重于保护程序的生成、存储或执行;然而,没有统一的保护策略来确保整个程序生命周期的安全性。本文将程序生成的静态安全性与进程执行的动态安全性相结合,提出了一种新的程序逻辑一致性安全特性。提出了一种加密核心处理(ECP)体系结构,提供了协处理器解决方案,以保护指令和数据流粒度上的程序逻辑一致性。该体系结构中新的身份验证加密模式使用程序指令和数据相对于基于段的地址的偏移值作为其加密参数。采用轻量级加密原语来确保ECP增加的硬件负担是有限的,特别是在x $$ times $$ 64体系结构下。我们证明了该方案在ECP体系结构中满足选择明文攻击下的不可区分性,并证明了该体系结构对各种攻击的有效性。此外,对ECP体系结构引入的开销进行了理论性能分析。
{"title":"ECP: Coprocessor Architecture to Protect Program Logic Consistency","authors":"Yang Gao,&nbsp;Siqi Lu,&nbsp;Yongjuan Wang,&nbsp;Haopeng Fan,&nbsp;Qingdi Han,&nbsp;Jingsheng Li","doi":"10.1002/smr.70023","DOIUrl":"https://doi.org/10.1002/smr.70023","url":null,"abstract":"<div>\u0000 \u0000 <p>Contemporary program protection methods focus on safeguarding either program generation, storage, or execution; however, no unified protection strategy exists for ensuring the security of a full program lifecycle. In this study, we combine the static security of program generation with the dynamic security of process execution and propose a novel program logic consistency security property. An encryption core processing (ECP) architecture is presented that provides coprocessor solutions to protect the program logic consistency at the granularity of instructions and data flows. The new authenticated encryption mode in the architecture uses the offset value of the program's instructions and data in relation to the segment-based address as its encryption parameters. Lightweight cryptographic primitives are adopted to ensure that the hardware burden added by the ECP is limited, especially under <span></span><math>\u0000 <semantics>\u0000 <mrow>\u0000 <mo>×</mo>\u0000 </mrow>\u0000 <annotation>$$ times $$</annotation>\u0000 </semantics></math>64 architectures. We prove that the proposed scheme in the ECP architecture satisfies indistinguishability under chosen plaintext attack and demonstrate the effectiveness of the architecture against various attacks. Additionally, a theoretical performance analysis is provided for estimating the overhead introduced by the ECP architecture.</p>\u0000 </div>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 4","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143865784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Community Detection of Directed Network for Software Ecosystems Based on a Two-Step Information Dissemination Model 基于两步信息传播模型的软件生态系统有向网络群落检测
IF 1.7 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-04-23 DOI: 10.1002/smr.70025
Huijie Tu, Xiangjuan Yao, Tingting Hou, Dunwei Gong, Mengyi Yang

A software ecosystem is a complex system that allows developers to cooperate with each other. Community is a universal and important topological property of networks. Detecting the communities of the software ecosystem is of great significance for analyzing its structural characteristics, discovering its hidden patterns, and predicting its behavior. Traditional community detection algorithms of complex networks are mostly for undirected networks. For the social network, the direction of information dissemination between developers cannot be ignored. In addition, the existing algorithms of community detection usually only consider direct influence between individuals while neglecting indirect relationships. To solve these problems, this paper presents a community detection method based on a two-step information dissemination model for the software ecosystem. First, a two-step information dissemination model is established to calculate the information gain of nodes. Second, a ranking method of developers' comprehensive influence is given through their influence vectors and information gains. Finally, communities are detected by taking the influential nodes as the cluster centers and the probability of information dissemination as the clustering direction. The proposed method is applied to community detection of typical software ecosystems in GitHub. The experimental results show that our method has good performance in the identification of community structure.

软件生态系统是一个允许开发人员相互合作的复杂系统。共同体是网络的一个普遍而重要的拓扑性质。检测软件生态系统的群落对于分析软件生态系统的结构特征、发现软件生态系统的隐藏模式、预测软件生态系统的行为具有重要意义。传统的复杂网络社团检测算法多针对无向网络。对于社交网络来说,开发者之间的信息传播方向是不容忽视的。此外,现有的社区检测算法通常只考虑个体之间的直接影响,而忽略了间接关系。针对这些问题,本文提出了一种基于两步信息传播模型的软件生态系统社区检测方法。首先,建立两步信息传播模型,计算节点的信息增益;其次,通过开发者的影响向量和信息增益,给出了开发者综合影响力的排序方法。最后,以影响节点为聚类中心,以信息传播概率为聚类方向,进行社区检测。将该方法应用于GitHub中典型软件生态系统的社区检测。实验结果表明,该方法具有较好的群落结构识别性能。
{"title":"Community Detection of Directed Network for Software Ecosystems Based on a Two-Step Information Dissemination Model","authors":"Huijie Tu,&nbsp;Xiangjuan Yao,&nbsp;Tingting Hou,&nbsp;Dunwei Gong,&nbsp;Mengyi Yang","doi":"10.1002/smr.70025","DOIUrl":"https://doi.org/10.1002/smr.70025","url":null,"abstract":"<div>\u0000 \u0000 <p>A software ecosystem is a complex system that allows developers to cooperate with each other. Community is a universal and important topological property of networks. Detecting the communities of the software ecosystem is of great significance for analyzing its structural characteristics, discovering its hidden patterns, and predicting its behavior. Traditional community detection algorithms of complex networks are mostly for undirected networks. For the social network, the direction of information dissemination between developers cannot be ignored. In addition, the existing algorithms of community detection usually only consider direct influence between individuals while neglecting indirect relationships. To solve these problems, this paper presents a community detection method based on a two-step information dissemination model for the software ecosystem. First, a two-step information dissemination model is established to calculate the information gain of nodes. Second, a ranking method of developers' comprehensive influence is given through their influence vectors and information gains. Finally, communities are detected by taking the influential nodes as the cluster centers and the probability of information dissemination as the clustering direction. The proposed method is applied to community detection of typical software ecosystems in GitHub. The experimental results show that our method has good performance in the identification of community structure.</p>\u0000 </div>","PeriodicalId":48898,"journal":{"name":"Journal of Software-Evolution and Process","volume":"37 4","pages":""},"PeriodicalIF":1.7,"publicationDate":"2025-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143866021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Software-Evolution and Process
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1