首页 > 最新文献

Science of Computer Programming最新文献

英文 中文
Multi-objective optimization of cloud systems 云系统的多目标优化
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-29 DOI: 10.1016/j.scico.2026.103447
Miguel Pérez , Pablo C. Cañizares , Alberto Núñez
Currently, enormous amounts of data are continuously processed to support our daily activities, such as managing bank accounts, streaming movies, or interacting on social networks. In recent years, cloud infrastructures have proven to be a reliable solution, not only for processing this data but also for enabling users worldwide to access it remotely. However, this processing demands vast computing resources, leading to significant energy consumption.
In this paper, we present a strategy to address this problem by combining multi-objective optimization techniques with Metamorphic Testing (MT) and simulation tools to optimize cloud systems, focusing on both performance and energy consumption. To achieve this, several multi-objective genetic algorithms (MOGAs) have been integrated into the MT-EA4Cloud framework, a solution that previously applied single-objective evolutionary algorithms with MT. To determine the suitability of the proposed approach, an empirical study was conducted to analyze the behavior of the different MOGAs included in the framework. In this study, various test sets and two distinct workloads – inspired by big data analytics operations – were created to represent multiple cloud scenarios.
The results clearly demonstrate that MOGAs can be effectively combined with MT to optimize cloud systems while considering multiple objectives – in this case, performance and energy consumption. A careful analysis of the results indicates that increasing the mutation rate leads to the best outcomes. In general, the NSGA-II algorithm has produced the best results in the experiments conducted in this study.
目前,为了支持我们的日常活动,如管理银行账户、流媒体电影或在社交网络上互动,需要不断处理大量数据。近年来,云基础设施已被证明是一种可靠的解决方案,不仅可以处理这些数据,还可以使全球用户远程访问这些数据。然而,这种处理需要大量的计算资源,导致大量的能源消耗。在本文中,我们提出了一种解决这一问题的策略,通过将多目标优化技术与变形测试(MT)和仿真工具相结合来优化云系统,重点关注性能和能耗。为了实现这一目标,将几种多目标遗传算法(MOGAs)集成到MT- ea4cloud框架中,该解决方案之前将单目标进化算法与MT结合使用。为了确定所提出方法的适用性,进行了一项实证研究,分析了框架中包含的不同MOGAs的行为。在这项研究中,受大数据分析操作的启发,创建了各种测试集和两种不同的工作负载,以表示多种云场景。结果清楚地表明,MOGAs可以有效地与MT相结合,以优化云系统,同时考虑多个目标-在这种情况下,性能和能耗。对结果的仔细分析表明,增加突变率会导致最好的结果。总的来说,NSGA-II算法在本研究的实验中取得了最好的结果。
{"title":"Multi-objective optimization of cloud systems","authors":"Miguel Pérez ,&nbsp;Pablo C. Cañizares ,&nbsp;Alberto Núñez","doi":"10.1016/j.scico.2026.103447","DOIUrl":"10.1016/j.scico.2026.103447","url":null,"abstract":"<div><div>Currently, enormous amounts of data are continuously processed to support our daily activities, such as managing bank accounts, streaming movies, or interacting on social networks. In recent years, cloud infrastructures have proven to be a reliable solution, not only for processing this data but also for enabling users worldwide to access it remotely. However, this processing demands vast computing resources, leading to significant energy consumption.</div><div>In this paper, we present a strategy to address this problem by combining multi-objective optimization techniques with Metamorphic Testing (MT) and simulation tools to optimize cloud systems, focusing on both performance and energy consumption. To achieve this, several multi-objective genetic algorithms (MOGAs) have been integrated into the MT-EA4Cloud framework, a solution that previously applied single-objective evolutionary algorithms with MT. To determine the suitability of the proposed approach, an empirical study was conducted to analyze the behavior of the different MOGAs included in the framework. In this study, various test sets and two distinct workloads – inspired by big data analytics operations – were created to represent multiple cloud scenarios.</div><div>The results clearly demonstrate that MOGAs can be effectively combined with MT to optimize cloud systems while considering multiple objectives – in this case, performance and energy consumption. A careful analysis of the results indicates that increasing the mutation rate leads to the best outcomes. In general, the NSGA-II algorithm has produced the best results in the experiments conducted in this study.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"252 ","pages":"Article 103447"},"PeriodicalIF":1.4,"publicationDate":"2026-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146102614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DATVD: A novel vulnerability detection method based on dynamic attention and hybrid convolutional pooling DATVD:一种基于动态关注和混合卷积池的新型漏洞检测方法
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-21 DOI: 10.1016/j.scico.2026.103443
Jinfu Chen , Jinyu Mu , Saihua Cai , Jiapeng Zhou , Ziyan Liu , Xinping Shi
Nowadays, the digitization process is constantly advancing. While software has become an indispensable part of people’s lives, software vulnerabilities have also become a serious security threat. With the expansion of software scale and the rapid development of artificial intelligence technology, deep learning technology has been widely used in vulnerability detection. However, it performs poorly in the task of detecting whether there are vulnerabilities in real-world code. The reason is that the amount of real-world code has increased and its structure has become more complex. It is difficult for deep learning models to learn the relationship between code snippets and vulnerability triggers, so that the existing vulnerability detection models have low accuracy in real-world code vulnerability detection tasks. In order to solve the above problems, we propose a software vulnerability detection system DATVD based on dynamic attention. The model consists of three components. The graph embedding component extracts code semantics and structural information and generates a graph representation of the code; the GGNN_DAT component uses the dynamic attention mechanism to learn code features and generate a one-dimensional vector representation; the hybrid convolution pooling component performs graph classification. Due to the limited availability of real-world code datasets, we conducted experiments on the widely recognized public datasets-Debian, Chrome, and Hybrid. Experimental results show that compared with existing neural networks, the proposed DATVD model can effectively improve the accuracy of source code vulnerability detection. When compared to Devign model, the accuracy of the model on these datasets is improved by an average of 3.13 %. The proposed DATVD also demonstrates better detection stability.
如今,数字化进程在不断推进。在软件成为人们生活中不可或缺的一部分的同时,软件漏洞也成为了严重的安全威胁。随着软件规模的扩大和人工智能技术的快速发展,深度学习技术在漏洞检测中得到了广泛的应用。然而,它在检测真实代码中是否存在漏洞的任务中表现不佳。原因是实际代码的数量增加了,其结构变得更加复杂。深度学习模型难以学习代码片段与漏洞触发器之间的关系,使得现有的漏洞检测模型在现实世界的代码漏洞检测任务中准确率较低。为了解决上述问题,我们提出了一种基于动态关注的软件漏洞检测系统DATVD。该模型由三个部分组成。图嵌入组件提取代码语义和结构信息,并生成代码的图表示;GGNN_DAT组件使用动态关注机制学习代码特征并生成一维向量表示;混合卷积池化组件执行图分类。由于实际代码数据集的可用性有限,我们在广泛认可的公共数据集(debian、Chrome和Hybrid)上进行了实验。实验结果表明,与现有神经网络相比,所提出的DATVD模型能有效提高源代码漏洞检测的准确性。与Devign模型相比,该模型在这些数据集上的准确率平均提高了3.13%。所提出的DATVD也显示出更好的检测稳定性。
{"title":"DATVD: A novel vulnerability detection method based on dynamic attention and hybrid convolutional pooling","authors":"Jinfu Chen ,&nbsp;Jinyu Mu ,&nbsp;Saihua Cai ,&nbsp;Jiapeng Zhou ,&nbsp;Ziyan Liu ,&nbsp;Xinping Shi","doi":"10.1016/j.scico.2026.103443","DOIUrl":"10.1016/j.scico.2026.103443","url":null,"abstract":"<div><div>Nowadays, the digitization process is constantly advancing. While software has become an indispensable part of people’s lives, software vulnerabilities have also become a serious security threat. With the expansion of software scale and the rapid development of artificial intelligence technology, deep learning technology has been widely used in vulnerability detection. However, it performs poorly in the task of detecting whether there are vulnerabilities in real-world code. The reason is that the amount of real-world code has increased and its structure has become more complex. It is difficult for deep learning models to learn the relationship between code snippets and vulnerability triggers, so that the existing vulnerability detection models have low accuracy in real-world code vulnerability detection tasks. In order to solve the above problems, we propose a software vulnerability detection system DATVD based on dynamic attention. The model consists of three components. The graph embedding component extracts code semantics and structural information and generates a graph representation of the code; the GGNN_DAT component uses the dynamic attention mechanism to learn code features and generate a one-dimensional vector representation; the hybrid convolution pooling component performs graph classification. Due to the limited availability of real-world code datasets, we conducted experiments on the widely recognized public datasets-Debian, Chrome, and Hybrid. Experimental results show that compared with existing neural networks, the proposed DATVD model can effectively improve the accuracy of source code vulnerability detection. When compared to Devign model, the accuracy of the model on these datasets is improved by an average of 3.13 %. The proposed DATVD also demonstrates better detection stability.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103443"},"PeriodicalIF":1.4,"publicationDate":"2026-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146078409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Has cross-project defect prediction truly progressed? A five-year field diagnosis challenging the state-of-the-art 跨项目缺陷预测真的有进展吗?五年的现场诊断挑战了最先进的技术
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-20 DOI: 10.1016/j.scico.2026.103449
Kangjian Zhou

Background

Cross-Project Defect Prediction (CPDP) is a critically active area of software engineering. The past five years have witnessed a surge of proposed techniques, from complex feature representation learning to the application of Large Language Models (LLMs), all claiming state-of-the-art (SOTA).

Problem

However, this apparent innovation faces a fundamental challenge to the current research paradigm. The field has largely disregarded the established baseline, ManualDown, advocated years ago for its strong performance. This neglect makes it impossible to discern whether reported advancements constitute genuine progress or a false prosperity driven by inadequate comparisons.

Objective

This study aims to serve as a five-year field diagnosis. Our goal is to determine whether the CPDP field has truly progressed by empirically evaluating the most prominent recent approaches (2020–2024) against the ManualDown baseline, thereby correcting the course of the field.

Method

We conduct a large-scale, reproducible comparison by systematically selecting recent SOTA CPDP models. To ensure a fair and unambiguous assessment, we compare them against ManualDown using the identical target test projects and the exact same performance metrics as reported in the original studies of these models.

Results

Our diagnosis reveals a striking discrepancy: approximately half of the recently proposed SOTA models show no substantial improvement over ManualDown (exhibiting a small effect size at best), while the other half perform even worse. This finding poses a serious challenge to the current research paradigm, suggesting that the field's trajectory may be misguided. A notable exception is the emerging potential of LLMs, whose contextual understanding may hold the key to meaningful future gains.

Conclusion

ManualDown remains a robust, competitive baseline for both classification and the more practical effort-aware ranking tasks. Therefore, this field diagnosis establishes the formal adoption of ManualDown as a foundational baseline. This practice is essential to ensure that future CPDP research demonstrates verifiable, substantial improvements, thereby correcting the course of the field and steering it toward meaningful advancements.
跨项目缺陷预测(CPDP)是软件工程中一个非常活跃的领域。在过去的五年中,从复杂的特征表示学习到大型语言模型(llm)的应用,所有这些技术都声称是最先进的(SOTA)。然而,这种明显的创新面临着对当前研究范式的根本性挑战。该领域在很大程度上忽视了多年前因其强劲表现而提倡的既定基准ManualDown。这种忽视使得人们无法辨别所报道的进步是真正的进步,还是由不充分的比较所驱动的虚假繁荣。目的本研究旨在为5年的现场诊断提供依据。我们的目标是通过经验评估最近最突出的方法(2020-2024)来确定CPDP领域是否真正取得了进展,从而纠正该领域的进程。方法系统选择SOTA近期的CPDP模型,进行大规模、可重复的比较。为了确保公平和明确的评估,我们将它们与ManualDown进行比较,使用相同的目标测试项目和在这些模型的原始研究中报告的完全相同的性能指标。结果我们的诊断揭示了一个惊人的差异:最近提出的SOTA模型中,大约有一半没有显示出比ManualDown有实质性的改善(最多显示出很小的效应大小),而另一半的表现甚至更糟。这一发现对当前的研究范式提出了严峻的挑战,表明该领域的发展轨迹可能被误导了。一个值得注意的例外是法学硕士的新兴潜力,其上下文理解可能是未来有意义收益的关键。manualdown仍然是一个稳健的、有竞争力的基线,无论是分类还是更实际的努力意识排序任务。因此,该现场诊断确立了ManualDown作为基础基线的正式采用。这一实践对于确保未来的CPDP研究显示出可验证的、实质性的改进,从而纠正该领域的进程并将其引向有意义的进步至关重要。
{"title":"Has cross-project defect prediction truly progressed? A five-year field diagnosis challenging the state-of-the-art","authors":"Kangjian Zhou","doi":"10.1016/j.scico.2026.103449","DOIUrl":"10.1016/j.scico.2026.103449","url":null,"abstract":"<div><h3>Background</h3><div>Cross-Project Defect Prediction (CPDP) is a critically active area of software engineering. The past five years have witnessed a surge of proposed techniques, from complex feature representation learning to the application of Large Language Models (LLMs), all claiming state-of-the-art (SOTA).</div></div><div><h3>Problem</h3><div>However, this apparent innovation faces a fundamental challenge to the current research paradigm. The field has largely disregarded the established baseline, ManualDown, advocated years ago for its strong performance. This neglect makes it impossible to discern whether reported advancements constitute genuine progress or a false prosperity driven by inadequate comparisons.</div></div><div><h3>Objective</h3><div>This study aims to serve as a five-year field diagnosis. Our goal is to determine whether the CPDP field has truly progressed by empirically evaluating the most prominent recent approaches (2020–2024) against the ManualDown baseline, thereby correcting the course of the field.</div></div><div><h3>Method</h3><div>We conduct a large-scale, reproducible comparison by systematically selecting recent SOTA CPDP models. To ensure a fair and unambiguous assessment, we compare them against ManualDown using the identical target test projects and the exact same performance metrics as reported in the original studies of these models.</div></div><div><h3>Results</h3><div>Our diagnosis reveals a striking discrepancy: approximately half of the recently proposed SOTA models show no substantial improvement over ManualDown (exhibiting a small effect size at best), while the other half perform even worse. This finding poses a serious challenge to the current research paradigm, suggesting that the field's trajectory may be misguided. A notable exception is the emerging potential of LLMs, whose contextual understanding may hold the key to meaningful future gains.</div></div><div><h3>Conclusion</h3><div>ManualDown remains a robust, competitive baseline for both classification and the more practical effort-aware ranking tasks. Therefore, this field diagnosis establishes the formal adoption of ManualDown as a foundational baseline. This practice is essential to ensure that future CPDP research demonstrates verifiable, substantial improvements, thereby correcting the course of the field and steering it toward meaningful advancements.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103449"},"PeriodicalIF":1.4,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146037773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Introducing a novel technique for call graph visualization and design pattern detection through runtime data profiling and dynamic warping 介绍了一种通过运行时数据分析和动态翘曲实现调用图可视化和设计模式检测的新技术
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-20 DOI: 10.1016/j.scico.2026.103445
Tarik Houichime, Younes el Amrani
Automated design patterns recovery (ADPR) is a significant academic endeavor focused on identifying and methodically recording design patterns found within software codebases. This process typically involves a thorough examination of source code to find characteristics indicative of specific patterns. Despite sophisticated techniques, persistent challenges remain. These challenges include the complexity of static analysis and variations in pattern appearance across languages. Critically, static methods are fundamentally ill-suited for capturing the temporal, interactive nature of behavioral design patterns. This has led to a field where dynamic methods, while promising, have seen limited exploration regarding their integration with modern, state-of-the-art classifiers, complicating the achievement of comprehensive results. This gap highlights a clear need for novel approaches that can effectively model and analyze runtime behavior directly. In response, this study introduces a dynamic, language-portable two-stage framework. First, we present a novel method for visualizing runtime data as a perceptually-tuned sinusoidal signal. This signal acts as a discovery tool for human analysts, encoding the “local” context of a call (e.g., method’s type) as amplitude and its “global” context (e.g., object interactions) as frequency. Second, we demonstrate how this visualization provides the foundational basis for the symbolic sequencing used in pattern detection. The signal acts as a procedural bridge, it allows an analyst to identify a ’Region of Interest’ from the signal, which then guides the extraction of the corresponding event snippet from the raw log. This snippet is then translated into a compact, symbolic “behavio-stuctural signature”, providing a robust and analyzable representation. Importantly, this work also studies the nature of these sequences, such as their optimal length, and how these properties impact the classification process, thereby validating the foundational basis of the sequential representation.
自动设计模式恢复(ADPR)是一项重要的学术研究,专注于识别和系统地记录在软件代码库中发现的设计模式。这个过程通常包括对源代码的彻底检查,以找到指示特定模式的特征。尽管技术成熟,但挑战依然存在。这些挑战包括静态分析的复杂性和跨语言模式外观的变化。关键的是,静态方法从根本上不适合捕捉行为设计模式的时间和交互特性。这导致了一个领域,动态方法,虽然有希望,已经看到有限的探索,他们与现代的,最先进的分类器的集成,复杂的实现全面的结果。这一差距突出了对能够直接有效地建模和分析运行时行为的新方法的明确需求。作为回应,本研究引入了一个动态的、语言可移植的两阶段框架。首先,我们提出了一种将运行时数据可视化为感知调谐正弦信号的新方法。该信号作为人类分析人员的发现工具,将调用的“本地”上下文(例如,方法的类型)编码为幅度,并将其“全局”上下文(例如,对象交互)编码为频率。其次,我们演示了这种可视化如何为模式检测中使用的符号排序提供基础基础。信号作为一个程序桥梁,它允许分析师从信号中识别一个“感兴趣的区域”,然后指导从原始日志中提取相应的事件片段。然后,这个片段被翻译成一个紧凑的、象征性的“行为结构签名”,提供一个健壮的、可分析的表示。重要的是,这项工作还研究了这些序列的性质,例如它们的最佳长度,以及这些属性如何影响分类过程,从而验证了序列表示的基础。
{"title":"Introducing a novel technique for call graph visualization and design pattern detection through runtime data profiling and dynamic warping","authors":"Tarik Houichime,&nbsp;Younes el Amrani","doi":"10.1016/j.scico.2026.103445","DOIUrl":"10.1016/j.scico.2026.103445","url":null,"abstract":"<div><div>Automated design patterns recovery (ADPR) is a significant academic endeavor focused on identifying and methodically recording design patterns found within software codebases. This process typically involves a thorough examination of source code to find characteristics indicative of specific patterns. Despite sophisticated techniques, persistent challenges remain. These challenges include the complexity of static analysis and variations in pattern appearance across languages. Critically, static methods are fundamentally ill-suited for capturing the temporal, interactive nature of behavioral design patterns. This has led to a field where dynamic methods, while promising, have seen limited exploration regarding their integration with modern, state-of-the-art classifiers, complicating the achievement of comprehensive results. This gap highlights a clear need for novel approaches that can effectively model and analyze runtime behavior directly. In response, this study introduces a dynamic, language-portable two-stage framework. First, we present a novel method for visualizing runtime data as a perceptually-tuned sinusoidal signal. This signal acts as a discovery tool for human analysts, encoding the “local” context of a call (e.g., method’s type) as amplitude and its “global” context (e.g., object interactions) as frequency. Second, we demonstrate how this visualization provides the foundational basis for the symbolic sequencing used in pattern detection. The signal acts as a procedural bridge, it allows an analyst to identify a ’Region of Interest’ from the signal, which then guides the extraction of the corresponding event snippet from the raw log. This snippet is then translated into a compact, symbolic “behavio-stuctural signature”, providing a robust and analyzable representation. Importantly, this work also studies the nature of these sequences, such as their optimal length, and how these properties impact the classification process, thereby validating the foundational basis of the sequential representation.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103445"},"PeriodicalIF":1.4,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146078412","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Securing LLM code generation: Leveraging prompt engineering to mitigate vulnerabilities across models and languages 保护LLM代码生成:利用提示工程来减轻模型和语言之间的漏洞
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-19 DOI: 10.1016/j.scico.2026.103446
Shaykhah S. Aldosari , Layla S. Aldawsari
Large Language Models (LLMs) represent a significant advancement in artificial intelligence (AI) capabilities, enabling natural and intuitive human-machine interactions. One rapidly evolving AI application involves LLM code generation, which can expedite software development by automating code writing, debugging, and optimizing. However, despite these enhanced capabilities, essential questions remain regarding the security implications of code generated by these models. This study addresses three key research questions to examine the security risks in LLM-generated code. It examines whether code generated by different open-source LLMs exhibits measurable variation in vulnerability prevalence. Furthermore, it also investigates how the choice of programming languages influences the security of LLM-generated code. Finally, it explores the degree to which prompt specificity and construction shape the security of the generated code. Our findings demonstrate differences across all dimensions: LLMs exhibited a variance of up to 136.06, programming languages showed a maximum performance gap of 56%, and prompt engineering achieved up to 77% improvement in security.
大型语言模型(llm)代表了人工智能(AI)能力的重大进步,实现了自然和直观的人机交互。一个快速发展的AI应用程序涉及LLM代码生成,它可以通过自动化代码编写、调试和优化来加快软件开发。然而,尽管有这些增强的功能,关于这些模型生成的代码的安全性问题仍然存在。本研究解决了三个关键的研究问题,以检查法学硕士生成的代码中的安全风险。它检查了由不同的开源llm生成的代码是否在漏洞流行方面表现出可测量的差异。此外,它还研究了编程语言的选择如何影响法学硕士生成代码的安全性。最后,它探讨了提示的特异性和结构在多大程度上塑造了生成代码的安全性。我们的研究结果显示了所有维度的差异:llm表现出高达136.06的差异,编程语言表现出56%的最大性能差距,而提示工程在安全性方面实现了高达77%的改进。
{"title":"Securing LLM code generation: Leveraging prompt engineering to mitigate vulnerabilities across models and languages","authors":"Shaykhah S. Aldosari ,&nbsp;Layla S. Aldawsari","doi":"10.1016/j.scico.2026.103446","DOIUrl":"10.1016/j.scico.2026.103446","url":null,"abstract":"<div><div>Large Language Models (LLMs) represent a significant advancement in artificial intelligence (AI) capabilities, enabling natural and intuitive human-machine interactions. One rapidly evolving AI application involves LLM code generation, which can expedite software development by automating code writing, debugging, and optimizing. However, despite these enhanced capabilities, essential questions remain regarding the security implications of code generated by these models. This study addresses three key research questions to examine the security risks in LLM-generated code. It examines whether code generated by different open-source LLMs exhibits measurable variation in vulnerability prevalence. Furthermore, it also investigates how the choice of programming languages influences the security of LLM-generated code. Finally, it explores the degree to which prompt specificity and construction shape the security of the generated code. Our findings demonstrate differences across all dimensions: LLMs exhibited a variance of up to 136.06, programming languages showed a maximum performance gap of 56%, and prompt engineering achieved up to 77% improvement in security.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103446"},"PeriodicalIF":1.4,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146078411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cost-adaptive multi-level semantic feature learning for source code based bug severity prediction 基于成本自适应多级语义特征学习的源代码bug严重性预测
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-17 DOI: 10.1016/j.scico.2026.103444
Xiaoke Zhu , Yufeng Shi , Xiaopan Chen , Caihong Yuan , Fumin Qi , Xiao-Yuan Jing
Bug severity prediction plays a crucial role in software development by enabling timely defect management. Traditional approaches that rely on bug reports are prone to subjective bias, often leading to inaccurate severity assessments. In contrast, source code-based methods can directly learn code representations to more accurately identify potential defects. However, existing source code-based models don’t make full use of the hierarchical deep semantic information, and don’t pay enough attention on the intrinsic class imbalance issue. To overcome these challenges, this paper presents the Cost-Adaptive Multi-level sEmantic feature Learning (CAMEL) framework for bug severity prediction. The framework comprises three core modules: the feature extraction module, the Multi-level Semantic Information Fusion (MSIF) module, and the Cost Weight Optimization (CWO) module. Specifically, the feature extraction module leverages CodeBERT to capture multi-level semantic information from source code. The MSIF then dynamically aggregates layer-specific features from each CodeBERT layer using an LSTM combined with a hierarchical attention mechanism, thereby preserving global semantic integrity. Finally, the CWO module mitigates the influence of class imbalance issue by dynamically adjusting class weight parameters. Experiments conducted on a dataset of 3342 method-level code snippets with varying bug severity levels demonstrate that CAMEL significantly outperforms state-of-the-art methods across key metrics, including F1-Weighted, Precision, Recall, and MCC.
Bug严重性预测在软件开发中起着至关重要的作用,它支持及时的缺陷管理。依赖bug报告的传统方法容易产生主观偏见,经常导致不准确的严重性评估。相比之下,基于源代码的方法可以直接学习代码表示,从而更准确地识别潜在的缺陷。然而,现有的基于源代码的模型没有充分利用分层深层语义信息,对固有的类不平衡问题重视不够。为了克服这些挑战,本文提出了用于漏洞严重性预测的成本自适应多层次语义特征学习(CAMEL)框架。该框架包括三个核心模块:特征提取模块、多层语义信息融合(MSIF)模块和成本权重优化(CWO)模块。具体来说,特征提取模块利用CodeBERT从源代码中捕获多层次语义信息。然后,MSIF使用LSTM和分层注意机制来动态聚合来自每个CodeBERT层的特定层的特征,从而保持全局语义完整性。最后,CWO模块通过动态调整类权重参数来缓解类不平衡问题的影响。在3342个方法级代码片段的数据集上进行的实验表明,CAMEL在关键指标上明显优于最先进的方法,包括f1加权、精度、召回率和MCC。
{"title":"Cost-adaptive multi-level semantic feature learning for source code based bug severity prediction","authors":"Xiaoke Zhu ,&nbsp;Yufeng Shi ,&nbsp;Xiaopan Chen ,&nbsp;Caihong Yuan ,&nbsp;Fumin Qi ,&nbsp;Xiao-Yuan Jing","doi":"10.1016/j.scico.2026.103444","DOIUrl":"10.1016/j.scico.2026.103444","url":null,"abstract":"<div><div>Bug severity prediction plays a crucial role in software development by enabling timely defect management. Traditional approaches that rely on bug reports are prone to subjective bias, often leading to inaccurate severity assessments. In contrast, source code-based methods can directly learn code representations to more accurately identify potential defects. However, existing source code-based models don’t make full use of the hierarchical deep semantic information, and don’t pay enough attention on the intrinsic class imbalance issue. To overcome these challenges, this paper presents the Cost-Adaptive Multi-level sEmantic feature Learning (CAMEL) framework for bug severity prediction. The framework comprises three core modules: the feature extraction module, the Multi-level Semantic Information Fusion (MSIF) module, and the Cost Weight Optimization (CWO) module. Specifically, the feature extraction module leverages CodeBERT to capture multi-level semantic information from source code. The MSIF then dynamically aggregates layer-specific features from each CodeBERT layer using an LSTM combined with a hierarchical attention mechanism, thereby preserving global semantic integrity. Finally, the CWO module mitigates the influence of class imbalance issue by dynamically adjusting class weight parameters. Experiments conducted on a dataset of 3342 method-level code snippets with varying bug severity levels demonstrate that CAMEL significantly outperforms state-of-the-art methods across key metrics, including F1-Weighted, Precision, Recall, and MCC.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103444"},"PeriodicalIF":1.4,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146037864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive urgency-based real-time task scheduling in ADAS systems ADAS系统中基于自适应紧急度的实时任务调度
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2026-01-09 DOI: 10.1016/j.scico.2026.103436
Mahdi Seyfipoor, Sayyed Muhammad Jaffry, Siamak Mohammadi
Advanced driver assistance systems (ADAS) are quintessential examples of real-time Cyber-Physical Systems (CPS), where physical processes and computational elements interact in real-time to enhance safety and automation in transportation. Efficient task scheduling is one of the most important software aspects in real-time systems. In this paper, we propose a CPS-oriented framework for real-time task scheduling in ADAS with a real-time scheduler. Our design integrates object detection, distance estimation, and an Adaptive Urgency scheduler that fuses normalized laxity, dynamic priority, and computational load into a unified metric to manage aperiodic tasks with strict temporal requirements. By adjusting the number of frames between each tracking based on the environment stress, we reduce unneeded object detection tasks, relying on tracking instead. Focusing on aperiodic tasks, the proposed multi-core task scheduler design handles sensor-triggered events and adapts scheduling dynamically while reducing context switch overhead by limiting unnecessary preemption. Experimental results demonstrate improved deadline adherence and improved priority reinforcement, validating the approach for real-time CPS implementations in automotive domains. This paper focuses on the performance of the scheduler from the aspect of deadline misses, context switches, and stability, as well as the ratio of high-priority deadline misses to the total number of deadline misses. We use software simulation to evaluate the algorithms, where the results show an improvement over classical real-time scheduling algorithms, as well as newer algorithms that have contributed to this field. Our proposed algorithm achieved a proportional miss rate of 3 % for critical tasks, which is a 16 % improvement over baselines such as EDF and MLLF.
高级驾驶辅助系统(ADAS)是实时网络物理系统(CPS)的典型例子,物理过程和计算元素实时交互,以提高交通运输的安全性和自动化程度。高效的任务调度是实时系统中最重要的软件方面之一。本文提出了一种基于cps的ADAS实时任务调度框架。我们的设计集成了目标检测、距离估计和一个自适应紧急调度程序,该调度程序将规范化松弛性、动态优先级和计算负载融合到一个统一的度量中,以管理具有严格时间要求的非周期性任务。通过根据环境压力调整每次跟踪之间的帧数,减少不必要的目标检测任务,转而依赖于跟踪。针对非周期任务,提出的多核任务调度器设计处理传感器触发事件并动态调整调度,同时通过限制不必要的抢占来减少上下文切换开销。实验结果表明,改进的截止日期遵守和改进的优先级强化,验证了该方法在汽车领域的实时CPS实施。本文从截止日期缺失、上下文切换、稳定性以及高优先级截止日期缺失与总截止日期缺失的比例三个方面对调度程序的性能进行了研究。我们使用软件模拟来评估算法,结果显示优于经典的实时调度算法,以及对该领域做出贡献的新算法。我们提出的算法在关键任务中实现了3%的比例缺失率,比EDF和MLLF等基准提高了16%。
{"title":"Adaptive urgency-based real-time task scheduling in ADAS systems","authors":"Mahdi Seyfipoor,&nbsp;Sayyed Muhammad Jaffry,&nbsp;Siamak Mohammadi","doi":"10.1016/j.scico.2026.103436","DOIUrl":"10.1016/j.scico.2026.103436","url":null,"abstract":"<div><div>Advanced driver assistance systems (ADAS) are quintessential examples of real-time Cyber-Physical Systems (CPS), where physical processes and computational elements interact in real-time to enhance safety and automation in transportation. Efficient task scheduling is one of the most important software aspects in real-time systems. In this paper, we propose a CPS-oriented framework for real-time task scheduling in ADAS with a real-time scheduler. Our design integrates object detection, distance estimation, and an Adaptive Urgency scheduler that fuses normalized laxity, dynamic priority, and computational load into a unified metric to manage aperiodic tasks with strict temporal requirements. By adjusting the number of frames between each tracking based on the environment stress, we reduce unneeded object detection tasks, relying on tracking instead. Focusing on aperiodic tasks, the proposed multi-core task scheduler design handles sensor-triggered events and adapts scheduling dynamically while reducing context switch overhead by limiting unnecessary preemption. Experimental results demonstrate improved deadline adherence and improved priority reinforcement, validating the approach for real-time CPS implementations in automotive domains. This paper focuses on the performance of the scheduler from the aspect of deadline misses, context switches, and stability, as well as the ratio of high-priority deadline misses to the total number of deadline misses. We use software simulation to evaluate the algorithms, where the results show an improvement over classical real-time scheduling algorithms, as well as newer algorithms that have contributed to this field. Our proposed algorithm achieved a proportional miss rate of 3 % for critical tasks, which is a 16 % improvement over baselines such as EDF and MLLF.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103436"},"PeriodicalIF":1.4,"publicationDate":"2026-01-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146078410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Event-B formalisation of a chat system: A case study 聊天系统的事件- b形式化:一个案例研究
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-27 DOI: 10.1016/j.scico.2025.103432
Néstor Cataño
This paper presents the formal modelling and refinement of a chat system using the Event-B formal method. We elicit software requirements as User Stories and manually map them into Event-B. We model core chat functionalities, including user creation, chat session creation, message sending, message forwarding, and message deletion, while ensuring consistency via invariants and proof obligations in Rodin. We discuss challenges, lessons learnt, and propose several best modelling practices for the design and verification of similar event-driven messaging systems. Our work outlines directions for future integration with tool-supported code generation.
本文采用Event-B形式化方法对聊天系统进行形式化建模和改进。我们以用户故事的形式引出软件需求,并手动将它们映射到Event-B中。我们为核心聊天功能建模,包括用户创建、聊天会话创建、消息发送、消息转发和消息删除,同时通过Rodin中的不变量和证明义务确保一致性。我们讨论了挑战和经验教训,并为类似事件驱动的消息传递系统的设计和验证提出了几个最佳建模实践。我们的工作概述了未来与工具支持的代码生成集成的方向。
{"title":"Event-B formalisation of a chat system: A case study","authors":"Néstor Cataño","doi":"10.1016/j.scico.2025.103432","DOIUrl":"10.1016/j.scico.2025.103432","url":null,"abstract":"<div><div>This paper presents the formal modelling and refinement of a chat system using the <span>Event-B</span> formal method. We elicit software requirements as User Stories and manually map them into <span>Event-B</span>. We model core chat functionalities, including user creation, chat session creation, message sending, message forwarding, and message deletion, while ensuring consistency via invariants and proof obligations in Rodin. We discuss challenges, lessons learnt, and propose several best modelling practices for the design and verification of similar event-driven messaging systems. Our work outlines directions for future integration with tool-supported code generation.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103432"},"PeriodicalIF":1.4,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145927623","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhancing software quality attributes through multi-dimensional refactoring at source-level 通过源代码级别的多维重构来增强软件质量属性
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-25 DOI: 10.1016/j.scico.2025.103434
Morteza Zakeri , Fatemeh Abdi , Fatemeh Bagheri
Cyber-Physical Systems (CPSs) increasingly depend on complex, high-level software components for coordination, integration, and control logic. As these components evolve, maintaining key quality attributes—such as modularity, testability, and architectural stability—becomes essential. Automated source-level refactoring offers a practical and systematic way to maintain software quality in dynamic CPS environments, where evolution occurs through ongoing development rather than autonomous runtime adaptation. Search-based refactoring methods identify optimal refactoring sequences to enhance software quality automatically. However, the multiplicity of quality attributes, the lack of formal definitions for them, and their non-correlation make it challenging to measure, reconcile, and appropriately apply quality attributes in search-based refactoring. This paper introduces an automated refactoring engine, CodART, which utilizes compiler principles to perform 18 different refactoring operations at the source code level, generating compilable code. Additionally, nine quality attributes are defined and evaluated to guide search-based refactoring effectively. The novel RNSGA-III algorithm is employed to better balance objectives in the nine-dimensional space. Many existing refactoring tools apply transformations at simplified code, UML, or AST level and do not directly output compilable, transformed source code. In contrast, CodART applies all transformations at the source level and produces compilable Java programs as output - a key requirement for integration into high-assurance CPS software pipelines. Compared to existing approaches, the proposed method enhances the number of quality attributes, refactorings, and optimization algorithms. The proposed algorithm improves software quality by an average of 9%, 12%, and 18% in large, medium, and small projects, respectively, surpassing state-of-the-art methods.
信息物理系统(cps)越来越依赖于复杂的高级软件组件来进行协调、集成和控制逻辑。随着这些组件的发展,维护关键的质量属性——比如模块化、可测试性和架构稳定性——变得至关重要。自动化的源代码级重构提供了一种在动态CPS环境中维护软件质量的实用而系统的方法,在这种环境中,进化是通过持续的开发而不是自主的运行时适应发生的。基于搜索的重构方法识别最佳重构序列,自动提高软件质量。然而,质量属性的多样性、缺乏它们的正式定义以及它们的非相关性使得在基于搜索的重构中度量、协调和适当地应用质量属性变得具有挑战性。本文介绍了一个自动重构引擎,CodART,它利用编译器原理在源代码级别执行18种不同的重构操作,生成可编译的代码。此外,还定义和评估了9个质量属性,以有效地指导基于搜索的重构。采用新颖的RNSGA-III算法在九维空间中更好地平衡目标。许多现有的重构工具在简化代码、UML或AST级别应用转换,并且不直接输出可编译的、转换的源代码。相反,CodART在源代码级别应用所有转换,并生成可编译的Java程序作为输出——这是集成到高保证CPS软件管道中的关键需求。与现有方法相比,该方法增加了质量属性、重构和优化算法的数量。所提出的算法在大型、中型和小型项目中分别平均提高了9%、12%和18%的软件质量,超过了最先进的方法。
{"title":"Enhancing software quality attributes through multi-dimensional refactoring at source-level","authors":"Morteza Zakeri ,&nbsp;Fatemeh Abdi ,&nbsp;Fatemeh Bagheri","doi":"10.1016/j.scico.2025.103434","DOIUrl":"10.1016/j.scico.2025.103434","url":null,"abstract":"<div><div>Cyber-Physical Systems (CPSs) increasingly depend on complex, high-level software components for coordination, integration, and control logic. As these components evolve, maintaining key quality attributes—such as modularity, testability, and architectural stability—becomes essential. Automated source-level refactoring offers a practical and systematic way to maintain software quality in dynamic CPS environments, where evolution occurs through ongoing development rather than autonomous runtime adaptation. Search-based refactoring methods identify optimal refactoring sequences to enhance software quality automatically. However, the multiplicity of quality attributes, the lack of formal definitions for them, and their non-correlation make it challenging to measure, reconcile, and appropriately apply quality attributes in search-based refactoring. This paper introduces an automated refactoring engine, CodART, which utilizes compiler principles to perform 18 different refactoring operations at the source code level, generating compilable code. Additionally, nine quality attributes are defined and evaluated to guide search-based refactoring effectively. The novel RNSGA-III algorithm is employed to better balance objectives in the nine-dimensional space. Many existing refactoring tools apply transformations at simplified code, UML, or AST level and do not directly output compilable, transformed source code. In contrast, CodART applies all transformations at the source level and produces compilable Java programs as output - a key requirement for integration into high-assurance CPS software pipelines. Compared to existing approaches, the proposed method enhances the number of quality attributes, refactorings, and optimization algorithms. The proposed algorithm improves software quality by an average of 9%, 12%, and 18% in large, medium, and small projects, respectively, surpassing state-of-the-art methods.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103434"},"PeriodicalIF":1.4,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145885477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MulAD: A log-based anomaly detection approach for distributed systems using multi-pattern and multi-model fusion MulAD:一种基于日志的分布式系统异常检测方法,采用多模式和多模型融合
IF 1.4 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-23 DOI: 10.1016/j.scico.2025.103433
Xinjie Wei , Chang-Ai Sun , Xiaoyi Zhang , Dave Towey
Context:Log-based anomaly detection (LAD) techniques examine whether or not continuously-generated logs match historically-normal patterns: This helps to ensure reliability in distributed systems using DevOps. However, complex anomalies can span multiple log-pattern types and thus may only be detected by combining these patterns: Relying only on any single pattern may cause anomalies to be missed. These are false negatives in anomaly detection.
Objective:In this paper, we propose an Anomaly-Detection approach based on Multi-type log-pattern fusion and Multi-model integration (MulAD): MulAD fuses multi-type log patterns into a synthetic representation to detect complex anomalies.
Method:MulAD first rearranges logs by source parameters to decouple interleaving logs and isolate relevant events. It then derives log patterns across five dimensions — semantic, sequential, quantitative, temporal (chronological), and parametric — and fuses them into a unified synthesized pattern. Finally, to detect anomalies, MulAD integrates the MABi-LSTM, Transformer, and graph neural network (GNN) models together: Each of these models is specifically designed to capture temporal and sequential dependencies, contextual information, and structural dependencies.
Result:We evaluated MulAD on three public datasets (HDFS, BGL, and ThunderBird) and one industrial one, from the Ray system. Experimental results show that MulAD outperforms all state-of-the-art techniques.
Conclusion:We conclude that MulAD is a promising anomaly-detection technique for complex anomalies in distributed systems.
上下文:基于日志的异常检测(LAD)技术检查连续生成的日志是否与历史正常模式匹配:这有助于确保使用DevOps的分布式系统的可靠性。然而,复杂的异常可以跨越多个日志模式类型,因此可能只能通过组合这些模式来检测:仅依赖任何单一模式可能会导致错过异常。这些是异常检测中的假阴性。目的:本文提出了一种基于多类型日志模式融合和多模型集成(MulAD)的异常检测方法:MulAD将多类型日志模式融合成一个综合表示来检测复杂异常。方法:MulAD首先按源参数重新排列日志,以解耦交错的日志并隔离相关事件。然后,它从五个维度(语义、顺序、数量、时间(时间顺序)和参数)派生日志模式,并将它们融合到统一的合成模式中。最后,为了检测异常,MulAD将MABi-LSTM、Transformer和图形神经网络(GNN)模型集成在一起:这些模型中的每一个都专门用于捕获时间和顺序依赖关系、上下文信息和结构依赖关系。结果:我们在来自Ray系统的三个公共数据集(HDFS、BGL和ThunderBird)和一个工业数据集上评估了MulAD。实验结果表明,MulAD优于所有最先进的技术。结论:我们认为MulAD是一种很有前途的分布式系统复杂异常检测技术。
{"title":"MulAD: A log-based anomaly detection approach for distributed systems using multi-pattern and multi-model fusion","authors":"Xinjie Wei ,&nbsp;Chang-Ai Sun ,&nbsp;Xiaoyi Zhang ,&nbsp;Dave Towey","doi":"10.1016/j.scico.2025.103433","DOIUrl":"10.1016/j.scico.2025.103433","url":null,"abstract":"<div><div><strong>Context:</strong>Log-based anomaly detection (LAD) techniques examine whether or not continuously-generated logs match historically-normal patterns: This helps to ensure reliability in distributed systems using DevOps. However, complex anomalies can span multiple log-pattern types and thus may only be detected by combining these patterns: Relying only on any single pattern may cause anomalies to be missed. These are false negatives in anomaly detection.</div><div><strong>Objective:</strong>In this paper, we propose an Anomaly-Detection approach based on Multi-type log-pattern fusion and Multi-model integration (MulAD): MulAD fuses multi-type log patterns into a synthetic representation to detect complex anomalies.</div><div><strong>Method:</strong>MulAD first rearranges logs by source parameters to decouple interleaving logs and isolate relevant events. It then derives log patterns across five dimensions — semantic, sequential, quantitative, temporal (chronological), and parametric — and fuses them into a unified <em>synthesized pattern</em>. Finally, to detect anomalies, MulAD integrates the MABi-LSTM, Transformer, and graph neural network (GNN) models together: Each of these models is specifically designed to capture temporal and sequential dependencies, contextual information, and structural dependencies.</div><div><strong>Result:</strong>We evaluated MulAD on three public datasets (HDFS, BGL, and ThunderBird) and one industrial one, from the Ray system. Experimental results show that MulAD outperforms all state-of-the-art techniques.</div><div><strong>Conclusion:</strong>We conclude that MulAD is a promising anomaly-detection technique for complex anomalies in distributed systems.</div></div>","PeriodicalId":49561,"journal":{"name":"Science of Computer Programming","volume":"251 ","pages":"Article 103433"},"PeriodicalIF":1.4,"publicationDate":"2025-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145885478","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Science of Computer Programming
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1