首页 > 最新文献

Future Generation Computer Systems-The International Journal of Escience最新文献

英文 中文
ZAD-ML: Dual-layer Learning for zero-Day attack detection in multivariate time series 多元时间序列零日攻击检测的双层学习
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-02-05 DOI: 10.1016/j.future.2026.108422
Edward Kwadwo Boahen , Ahmad Salehi Shahraki
Zero-day attack detection in multivariate time series is an increasingly vital field, driven by its significance in security-sensitive environments, such as network monitoring. Traditional anomaly detection methods often depend on predefined patterns, which are ineffective against zero-day attacks that exploit previously unidentified vulnerabilities. To address this limitation, we introduce ZAD-ML, an unsupervised learning framework designed specifically for detecting zero-day attacks by utilizing behavioral analytics without prior knowledge of attack signatures. ZAD-ML incorporates a dual-layer neural network system where the first layer learns to compress and encode normal temporal behavioral patterns into a dense representation, facilitating efficient anomaly detection and allowing it to adapt and update its understanding of normal behavior continuously. The second layer is enhanced with attention mechanisms to analyze temporal sequences for behavioral deviations, allowing the system to adaptively update its baseline for normal behavior based on emerging data trends. We incorporate deep learning techniques to enhance the model’s ability to learn complex patterns and anomalies in data behavior. We evaluated our framework on four public Network datasets, demonstrating its capability to detect zero-day attacks with high accuracy and significantly reduced false positives as compared with existing methods. ZAD-ML provides a robust, adaptable solution for real-time anomaly detection. The implementation of our proposed method is now publicly accessible at https://github.com/don2c/ZAD-ML.
基于多变量时间序列的零日攻击检测在网络监控等安全敏感环境中的重要性,使得零日攻击检测成为一个越来越重要的领域。传统的异常检测方法通常依赖于预定义的模式,这对于利用先前未识别漏洞的零日攻击是无效的。为了解决这一限制,我们引入了ZAD-ML,这是一种无监督学习框架,专门用于通过利用行为分析来检测零日攻击,而无需事先了解攻击特征。ZAD-ML采用双层神经网络系统,其中第一层学习压缩和编码正常的时间行为模式为密集表示,促进有效的异常检测,并允许它不断适应和更新其对正常行为的理解。第二层增强了注意力机制,用于分析行为偏差的时间序列,允许系统根据新出现的数据趋势自适应地更新其正常行为的基线。我们采用深度学习技术来增强模型学习复杂模式和数据行为异常的能力。我们在四个公共网络数据集上评估了我们的框架,证明了与现有方法相比,它能够高精度地检测零日攻击,并显着减少误报。ZAD-ML为实时异常检测提供了一个强大的、适应性强的解决方案。我们提出的方法的实现现在可以在https://github.com/don2c/ZAD-ML上公开访问。
{"title":"ZAD-ML: Dual-layer Learning for zero-Day attack detection in multivariate time series","authors":"Edward Kwadwo Boahen ,&nbsp;Ahmad Salehi Shahraki","doi":"10.1016/j.future.2026.108422","DOIUrl":"10.1016/j.future.2026.108422","url":null,"abstract":"<div><div>Zero-day attack detection in multivariate time series is an increasingly vital field, driven by its significance in security-sensitive environments, such as network monitoring. Traditional anomaly detection methods often depend on predefined patterns, which are ineffective against zero-day attacks that exploit previously unidentified vulnerabilities. To address this limitation, we introduce ZAD-ML, an unsupervised learning framework designed specifically for detecting zero-day attacks by utilizing behavioral analytics without prior knowledge of attack signatures. ZAD-ML incorporates a dual-layer neural network system where the first layer learns to compress and encode normal temporal behavioral patterns into a dense representation, facilitating efficient anomaly detection and allowing it to adapt and update its understanding of normal behavior continuously. The second layer is enhanced with attention mechanisms to analyze temporal sequences for behavioral deviations, allowing the system to adaptively update its baseline for normal behavior based on emerging data trends. We incorporate deep learning techniques to enhance the model’s ability to learn complex patterns and anomalies in data behavior. We evaluated our framework on four public Network datasets, demonstrating its capability to detect zero-day attacks with high accuracy and significantly reduced false positives as compared with existing methods. ZAD-ML provides a robust, adaptable solution for real-time anomaly detection. The implementation of our proposed method is now publicly accessible at <span><span>https://github.com/don2c/ZAD-ML</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108422"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146135263","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Striking the balance between speed and compression ratio: A fast bit-grouping algorithm and adaptive compressor selection for scientific data 在速度和压缩比之间取得平衡:科学数据的快速位分组算法和自适应压缩器选择
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-10 DOI: 10.1016/j.future.2026.108370
Michael Middlezong
High-performance computing (HPC) systems have enabled unprecedented advancements in scientific simulation, producing larger and larger quantities of data to be analyzed. The resulting storage and I/O overheads present a significant bottleneck to scientific workflows. While many compression algorithms have been developed to address the issue, achieving the optimal balance between compression ratio and throughput remains a challenge. Furthermore, strict error bound requirements are inadequately addressed by current solutions. This paper introduces GRASP, a fast bit-grouping compressor that leverages the local smoothness of data to achieve high throughput while maintaining competitive compression ratios under tight error constraints. For the purposes of compressor selection, we also propose a novel efficiency metric that considers both compression and I/O performance, allowing the user to make an informed decision about which compressor to use. We also develop an adaptive compression selection framework based on this metric, using sampling to determine at runtime the optimal compressor for specific use cases. Experimental results across six diverse datasets demonstrate that GRASP outperforms traditional error-bounded compressors such as SZ3 and ZFP in speed while achieving similar compression ratios under tight error bounds. Additionally, we assess scenarios in which a naive compressor selection fails to select the optimal compressor, demonstrating the importance of an adaptive compressor selection framework. These contributions provide a practical approach to balancing speed and compression ratio in modern scientific data management.
高性能计算(HPC)系统使科学模拟取得了前所未有的进步,产生了越来越多的需要分析的数据。由此产生的存储和I/O开销是科学工作流程的一个重要瓶颈。虽然已经开发了许多压缩算法来解决这个问题,但在压缩比和吞吐量之间实现最佳平衡仍然是一个挑战。此外,当前的解决方案没有充分解决严格的错误边界要求。本文介绍了一种快速的位分组压缩器GRASP,它利用数据的局部平滑性来实现高吞吐量,同时在严格的错误约束下保持有竞争力的压缩比。为了选择压缩机,我们还提出了一种考虑压缩和I/O性能的新型效率指标,允许用户对使用哪种压缩机做出明智的决定。我们还基于该指标开发了一个自适应压缩选择框架,使用采样在运行时确定特定用例的最佳压缩器。在六个不同数据集上的实验结果表明,GRASP在速度上优于传统的错误有界压缩器(如SZ3和ZFP),同时在严格的错误界限下获得相似的压缩比。此外,我们还评估了原始压缩机选择无法选择最佳压缩机的情况,证明了自适应压缩机选择框架的重要性。这些贡献为现代科学数据管理中平衡速度和压缩比提供了一种实用的方法。
{"title":"Striking the balance between speed and compression ratio: A fast bit-grouping algorithm and adaptive compressor selection for scientific data","authors":"Michael Middlezong","doi":"10.1016/j.future.2026.108370","DOIUrl":"10.1016/j.future.2026.108370","url":null,"abstract":"<div><div>High-performance computing (HPC) systems have enabled unprecedented advancements in scientific simulation, producing larger and larger quantities of data to be analyzed. The resulting storage and I/O overheads present a significant bottleneck to scientific workflows. While many compression algorithms have been developed to address the issue, achieving the optimal balance between compression ratio and throughput remains a challenge. Furthermore, strict error bound requirements are inadequately addressed by current solutions. This paper introduces GRASP, a fast bit-grouping compressor that leverages the local smoothness of data to achieve high throughput while maintaining competitive compression ratios under tight error constraints. For the purposes of compressor selection, we also propose a novel efficiency metric that considers both compression and I/O performance, allowing the user to make an informed decision about which compressor to use. We also develop an adaptive compression selection framework based on this metric, using sampling to determine at runtime the optimal compressor for specific use cases. Experimental results across six diverse datasets demonstrate that GRASP outperforms traditional error-bounded compressors such as SZ3 and ZFP in speed while achieving similar compression ratios under tight error bounds. Additionally, we assess scenarios in which a naive compressor selection fails to select the optimal compressor, demonstrating the importance of an adaptive compressor selection framework. These contributions provide a practical approach to balancing speed and compression ratio in modern scientific data management.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108370"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145957071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High performance graph-parallel accelerator design 高性能图形并行加速器设计
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-23 DOI: 10.1016/j.future.2026.108385
Cemil Kaan Akyol , Muhammet Mustafa Ozdal , Ozcan Ozturk
Graph applications are becoming increasingly important with their widespread usage and the amounts of data they deal with. Biological and social web graphs are well-known examples that show the importance of efficiently processing graph analytic applications and problems. Due to limited resources, efficiency and performance are much more critical in embedded systems. We propose an efficient source-to-source-based methodology for graph applications that gives the freedom of not knowing the low-level details of parallelization and distribution by translating any vertex-centric C++ graph application into a pipelined SystemC model. High-Level Synthesis (HLS) tools can synthesize the generated SystemC model to obtain the design of the hardware. To support different types of graph applications, we have implemented features like non-standard application support, active set functionality, asynchronous execution support, conditional pipeline support, non-neighbor data access support, multiple pipeline support, and user-defined data type functionality. Our accelerator development flow can generate better-performing accelerators than OpenCL. Furthermore, it dramatically reduces the design time compared to using HLS tools. Therefore, the proposed methodology can generate fast accelerators with minimal effort using a high-level language description from the user.
由于图形应用程序的广泛使用和处理的数据量,它们正变得越来越重要。生物和社会网络图是众所周知的例子,显示了有效处理图分析应用程序和问题的重要性。由于资源有限,在嵌入式系统中效率和性能更为重要。我们为图形应用程序提出了一种高效的基于源到源的方法,通过将任何以顶点为中心的c++图形应用程序转换为流水线的SystemC模型,可以自由地不知道并行化和分布的底层细节。高级综合(High-Level Synthesis, HLS)工具可以综合生成的SystemC模型,从而得到硬件的设计。为了支持不同类型的图形应用程序,我们实现了一些特性,如非标准应用程序支持、活动集功能、异步执行支持、条件管道支持、非邻居数据访问支持、多管道支持和用户定义的数据类型功能。我们的加速器开发流程可以生成比OpenCL性能更好的加速器。此外,与使用HLS工具相比,它大大缩短了设计时间。因此,所提出的方法可以使用来自用户的高级语言描述以最小的努力生成快速加速器。
{"title":"High performance graph-parallel accelerator design","authors":"Cemil Kaan Akyol ,&nbsp;Muhammet Mustafa Ozdal ,&nbsp;Ozcan Ozturk","doi":"10.1016/j.future.2026.108385","DOIUrl":"10.1016/j.future.2026.108385","url":null,"abstract":"<div><div>Graph applications are becoming increasingly important with their widespread usage and the amounts of data they deal with. Biological and social web graphs are well-known examples that show the importance of efficiently processing graph analytic applications and problems. Due to limited resources, efficiency and performance are much more critical in embedded systems. We propose an efficient source-to-source-based methodology for graph applications that gives the freedom of not knowing the low-level details of parallelization and distribution by translating any vertex-centric C++ graph application into a pipelined SystemC model. High-Level Synthesis (HLS) tools can synthesize the generated SystemC model to obtain the design of the hardware. To support different types of graph applications, we have implemented features like non-standard application support, active set functionality, asynchronous execution support, conditional pipeline support, non-neighbor data access support, multiple pipeline support, and user-defined data type functionality. Our accelerator development flow can generate better-performing accelerators than OpenCL. Furthermore, it dramatically reduces the design time compared to using HLS tools. Therefore, the proposed methodology can generate fast accelerators with minimal effort using a high-level language description from the user.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108385"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
RoWD: Automated rogue workload detector for HPC security RoWD:用于高性能计算安全的自动流氓工作负载检测器
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-22 DOI: 10.1016/j.future.2026.108392
Francesco Antici , Jens Domke , Andrea Bartolini , Zeynep Kiziltan , Satoshi Matsuoka
The increasing reliance on High-Performance Computing (HPC) systems to execute complex scientific and industrial workloads raises significant security concerns related to the misuse of HPC resources for unauthorized or malicious activities. Rogue job executions can threaten the integrity, confidentiality, and availability of HPC infrastructures. Given the scale and heterogeneity of HPC job submissions, manual or ad hoc monitoring is inadequate to effectively detect such misuse. Therefore, automated solutions capable of systematically analyzing job submissions are essential to detect rogue workloads. To address this challenge, we present RoWD (Rogue Workload Detector), the first framework for automated and systematic security screening of the HPC job-submission pipeline. RoWD is composed of modular plug-ins that classify different types of workloads and enable the detection of rogue jobs through the analysis of job scripts and associated metadata. We deploy RoWD on the Supercomputer Fugaku to classify AI workloads and release SCRIPT-AI, the first dataset of annotated job scripts labeled with workload characteristics. We evaluate RoWD on approximately 50K previously unseen jobs executed on Fugaku between 2021 and 2025. Our results show that RoWD accurately classifies AI jobs (achieving an F1 score of 95%), is robust against adversarial behavior, and incurs low runtime overhead, making it suitable for strengthening the security of HPC environments and for real-time deployment in production systems.
越来越多地依赖高性能计算(HPC)系统来执行复杂的科学和工业工作负载,引发了与滥用HPC资源进行未经授权或恶意活动相关的重大安全问题。恶意作业执行会威胁到HPC基础架构的完整性、机密性和可用性。鉴于高性能计算作业提交的规模和异质性,手工或特别监测不足以有效地检测此类滥用。因此,能够系统地分析作业提交的自动化解决方案对于检测非法工作负载至关重要。为了应对这一挑战,我们提出了RoWD(流氓工作负载检测器),这是第一个对高性能计算作业提交管道进行自动化和系统安全筛选的框架。RoWD由模块化插件组成,这些插件对不同类型的工作负载进行分类,并通过分析作业脚本和相关元数据来检测流氓作业。我们在超级计算机Fugaku上部署了RoWD,对人工智能工作负载进行分类,并发布了SCRIPT-AI,这是第一个标有工作负载特征的注释作业脚本数据集。我们评估了2021年至2025年间在Fugaku执行的约5万个以前未见过的作业的RoWD。我们的研究结果表明,RoWD可以准确地对人工智能作业进行分类(达到95%的F1分数),对对抗行为具有鲁棒性,并且运行时开销低,适用于增强高性能计算环境的安全性和在生产系统中的实时部署。
{"title":"RoWD: Automated rogue workload detector for HPC security","authors":"Francesco Antici ,&nbsp;Jens Domke ,&nbsp;Andrea Bartolini ,&nbsp;Zeynep Kiziltan ,&nbsp;Satoshi Matsuoka","doi":"10.1016/j.future.2026.108392","DOIUrl":"10.1016/j.future.2026.108392","url":null,"abstract":"<div><div>The increasing reliance on High-Performance Computing (HPC) systems to execute complex scientific and industrial workloads raises significant security concerns related to the misuse of HPC resources for unauthorized or malicious activities. Rogue job executions can threaten the integrity, confidentiality, and availability of HPC infrastructures. Given the scale and heterogeneity of HPC job submissions, manual or ad hoc monitoring is inadequate to effectively detect such misuse. Therefore, automated solutions capable of systematically analyzing job submissions are essential to detect rogue workloads. To address this challenge, we present RoWD (Rogue Workload Detector), the first framework for automated and systematic security screening of the HPC job-submission pipeline. RoWD is composed of modular plug-ins that classify different types of workloads and enable the detection of rogue jobs through the analysis of job scripts and associated metadata. We deploy RoWD on the Supercomputer Fugaku to classify AI workloads and release SCRIPT-AI, the first dataset of annotated job scripts labeled with workload characteristics. We evaluate RoWD on approximately 50K previously unseen jobs executed on Fugaku between 2021 and 2025. Our results show that RoWD accurately classifies AI jobs (achieving an F1 score of 95%), is robust against adversarial behavior, and incurs low runtime overhead, making it suitable for strengthening the security of HPC environments and for real-time deployment in production systems.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108392"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033300","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vertical auto-scaling mechanism for elastic memory management of containerized applications in Kubernetes Kubernetes中容器化应用弹性内存管理的垂直自动扩展机制
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-22 DOI: 10.1016/j.future.2026.108407
Taeshin Kang, Minwoo Kang, Heonchang Yu
Cloud service providers typically offer containers with fixed resource sizes. However, cloud users often overprovision container resources to prevent service interruptions caused by resource shortages. This practice leads to low utilization of system resources in the cloud. To address this issue, cloud service providers offer container auto-scaling. They primarily support horizontal auto-scaling, which provides horizontal elasticity. However, this approach has limitations in responding promptly to unexpected spikes in resource usage and in optimizing resource utilization. Vertical auto-scaling can help overcome these limitations. Its importance is increasing, particularly for stateful and real-time applications that require immediate resource elasticity. Nevertheless, vertical elasticity remains difficult to achieve and has not been actively researched or widely implemented. This study proposes a vertical auto-scaling mechanism for elastic memory management in container-based applications running in Kubernetes, which is widely recognized as the standard platform for container orchestration. In the proposed approach, high-priority tasks are given priority for scaling up, while tasks that cannot undergo scale-up are suspended using the cgroup freeze feature to prevent further memory allocation. If memory pressure persists and task termination becomes unavoidable, tasks are terminated in ascending order of priority, starting with the lowest. Once memory pressure is relieved, stateful applications are restarted from the point at which they were suspended. Compared to the default Kubernetes environment without vertical elasticity, EVMMv2 reduced the total execution time of stateful applications by up to 40% and improved the request success rate of stateless applications by 37%.
云服务提供商通常提供固定资源大小的容器。但是,云用户经常过度配置容器资源,以防止资源短缺导致的服务中断。这种做法导致云中系统资源的利用率很低。为了解决这个问题,云服务提供商提供了容器自动伸缩功能。它们主要支持水平自动伸缩,从而提供水平弹性。然而,这种方法在迅速响应资源使用的意外高峰和优化资源利用方面存在局限性。垂直自动伸缩可以帮助克服这些限制。它的重要性正在增加,特别是对于需要即时资源弹性的有状态和实时应用程序。然而,垂直弹性仍然很难实现,没有积极研究或广泛实施。本研究提出了一种垂直自动伸缩机制,用于在Kubernetes上运行的基于容器的应用程序中的弹性内存管理,Kubernetes被广泛认为是容器编排的标准平台。在建议的方法中,高优先级的任务被赋予扩展的优先级,而不能进行扩展的任务则使用cgroup冻结特性挂起,以防止进一步的内存分配。如果内存压力持续存在并且任务终止不可避免,则按优先级升序终止任务,从最低优先级开始。一旦内存压力得到缓解,有状态应用程序将从挂起它们的位置重新启动。与没有垂直弹性的默认Kubernetes环境相比,EVMMv2将有状态应用程序的总执行时间减少了40%,并将无状态应用程序的请求成功率提高了37%。
{"title":"Vertical auto-scaling mechanism for elastic memory management of containerized applications in Kubernetes","authors":"Taeshin Kang,&nbsp;Minwoo Kang,&nbsp;Heonchang Yu","doi":"10.1016/j.future.2026.108407","DOIUrl":"10.1016/j.future.2026.108407","url":null,"abstract":"<div><div>Cloud service providers typically offer containers with fixed resource sizes. However, cloud users often overprovision container resources to prevent service interruptions caused by resource shortages. This practice leads to low utilization of system resources in the cloud. To address this issue, cloud service providers offer container auto-scaling. They primarily support horizontal auto-scaling, which provides horizontal elasticity. However, this approach has limitations in responding promptly to unexpected spikes in resource usage and in optimizing resource utilization. Vertical auto-scaling can help overcome these limitations. Its importance is increasing, particularly for stateful and real-time applications that require immediate resource elasticity. Nevertheless, vertical elasticity remains difficult to achieve and has not been actively researched or widely implemented. This study proposes a vertical auto-scaling mechanism for elastic memory management in container-based applications running in Kubernetes, which is widely recognized as the standard platform for container orchestration. In the proposed approach, high-priority tasks are given priority for scaling up, while tasks that cannot undergo scale-up are suspended using the <em>cgroup freeze</em> feature to prevent further memory allocation. If memory pressure persists and task termination becomes unavoidable, tasks are terminated in ascending order of priority, starting with the lowest. Once memory pressure is relieved, stateful applications are restarted from the point at which they were suspended. Compared to the default Kubernetes environment without vertical elasticity, EVMMv2 reduced the total execution time of stateful applications by up to 40% and improved the request success rate of stateless applications by 37%.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108407"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GraalDoss: Direct object snapshotting and sharing for cloud-native applications GraalDoss:云原生应用程序的直接对象快照和共享
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-20 DOI: 10.1016/j.future.2026.108375
Ivan Ristović , Vojin Jovanović , Peter Hofer , Milena Vujošević Janičić
Modern cloud-computing providers operate on a pay-as-you-use billing model, with computing power and memory being the most important and expensive resources. Due to resource costs, cloud-native applications should start fast while minimizing startup time and memory footprint over multiple application instances. However, modern workloads consist of large amounts of data, often requiring initialization which introduces repeated CPU work across application instances. Current cloud-native solutions that pre-initialize application code and data operate at application-build time to enable sharing during execution. However, these solutions do not consider data that becomes available or can only be initialized during application execution.
We present Doss, a direct object snapshotting and sharing mechanism for cloud-native applications. Doss snapshots the state of the object graph directly from the executing language-runtime heap. This allows Doss to achieve constant deserialization overhead with memory mappings. Doss shares warmed-up data snapshots across compatible language-runtime instances, reducing the memory overhead of the system, and avoiding cold starts. We implement GraalDoss in Java as part of GraalVM. GraalDoss maintains a constant data-cache memory overhead across multiple application instances, eliminating costly data initialization. In microservice applications, GraalDoss reduces the total memory footprint by 44% for 8 microservice instances and improves first-response times by 51%. In natural language processing applications, GraalDoss improves total execution times by several orders of magnitude.
现代云计算提供商采用按使用量付费的计费模式,计算能力和内存是最重要也是最昂贵的资源。由于资源成本,云原生应用程序应该快速启动,同时最小化多个应用程序实例的启动时间和内存占用。然而,现代工作负载由大量数据组成,通常需要初始化,这会在应用程序实例之间引入重复的CPU工作。当前的云原生解决方案在构建应用程序时对应用程序代码和数据进行预初始化,以便在执行期间实现共享。但是,这些解决方案不考虑在应用程序执行期间变得可用或只能初始化的数据。我们介绍了Doss,一个用于云原生应用程序的直接对象快照和共享机制。Doss直接从执行的语言运行时堆中快照对象图的状态。这允许Doss通过内存映射实现恒定的反序列化开销。Doss在兼容的语言运行时实例之间共享预热的数据快照,从而减少了系统的内存开销,并避免了冷启动。我们在Java中实现GraalDoss作为GraalVM的一部分。GraalDoss在多个应用程序实例之间保持恒定的数据缓存内存开销,从而消除了昂贵的数据初始化。在微服务应用程序中,GraalDoss为8个微服务实例减少了44%的总内存占用,并将首次响应时间提高了51%。在自然语言处理应用程序中,GraalDoss将总执行时间提高了几个数量级。
{"title":"GraalDoss: Direct object snapshotting and sharing for cloud-native applications","authors":"Ivan Ristović ,&nbsp;Vojin Jovanović ,&nbsp;Peter Hofer ,&nbsp;Milena Vujošević Janičić","doi":"10.1016/j.future.2026.108375","DOIUrl":"10.1016/j.future.2026.108375","url":null,"abstract":"<div><div>Modern cloud-computing providers operate on a pay-as-you-use billing model, with computing power and memory being the most important and expensive resources. Due to resource costs, cloud-native applications should start fast while minimizing startup time and memory footprint over multiple application instances. However, modern workloads consist of large amounts of data, often requiring initialization which introduces repeated CPU work across application instances. Current cloud-native solutions that pre-initialize application code and data operate at application-build time to enable sharing during execution. However, these solutions do not consider data that becomes available or can only be initialized during application execution.</div><div>We present <span>Doss</span>, a direct object snapshotting and sharing mechanism for cloud-native applications. <span>Doss</span> snapshots the state of the object graph directly from the executing language-runtime heap. This allows <span>Doss</span> to achieve constant deserialization overhead with memory mappings. <span>Doss</span> shares warmed-up data snapshots across compatible language-runtime instances, reducing the memory overhead of the system, and avoiding cold starts. We implement Graal<span>Doss</span> in Java as part of GraalVM. Graal<span>Doss</span> maintains a constant data-cache memory overhead across multiple application instances, eliminating costly data initialization. In microservice applications, Graal<span>Doss</span> reduces the total memory footprint by 44% for 8 microservice instances and improves first-response times by 51%. In natural language processing applications, Graal<span>Doss</span> improves total execution times by several orders of magnitude.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108375"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146014881","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MoFormer: A centrality-aware multi-task graph transformer with multi-gate mixture-of-experts for link-level network performance modeling MoFormer:用于链路级网络性能建模的具有多门混合专家的中心性感知多任务图转换器
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-24 DOI: 10.1016/j.future.2026.108406
Hanlin Liu , Aliya Bao , Mingyue Li , Yintan Ai , Hua Li
Link-level network performance modeling (NPM) facilitates efficient traffic control, precise fault localization, and reliable resource management in emerging network paradigms such as Software-Defined Networking and Intent-Based Networking. A variety of models, such as Long Short-Term Memory and Graph Neural Networks (GNNs), are utilized to enhance the effectiveness of NPM. However, a practical NPM requires the generalization ability to adapt to diverse network topologies and prediction tasks without retraining. To meet this requirement, graph Transformer models are a breakthrough by encoding nodes and their structural features into tokens, breaking free from the dependencies on fixed graph structures typical of traditional GNNs. Nevertheless, they mostly focus on node-centric representations, which are insufficient to capture the fine-grained interactions and dependencies between links, thus limiting their applicability in link-level NPM. In this paper, we propose a centrality-aware multi-task graph Transformer with multi-gate mixture-of-experts (MMoE), named MoFormer, for link-level NPM. Specifically, a link-centric tokenized graph representation method is proposed to transform each link and its neighborhood information into a sequence of tokens guided by the routing protocol. A routing-aware betweenness centrality encoding mechanism is further developed to enhance the ability to characterize the tokens considering the relative importance of each link. MoFormer takes advantage of MMoE combined with Transformer to enable joint learning of multiple prediction tasks. Experimental results on both simulated and real-world datasets demonstrate the significant improvements of MoFormer over existing state-of-the-art baselines while maintaining superior generalization ability.
在软件定义网络和基于意图的网络等新兴网络模式中,链路级网络性能建模(NPM)有助于实现高效的流量控制、精确的故障定位和可靠的资源管理。利用长短期记忆和图神经网络(gnn)等多种模型来提高NPM的有效性。然而,一个实际的NPM需要泛化能力来适应不同的网络拓扑和预测任务,而不需要再训练。为了满足这一需求,图转换模型是一种突破,它将节点及其结构特征编码为token,摆脱了传统gnn对固定图结构的依赖。然而,它们主要关注以节点为中心的表示,这不足以捕捉链接之间的细粒度交互和依赖关系,从而限制了它们在链接级NPM中的适用性。在本文中,我们提出了一个具有多门混合专家(MMoE)的中心感知多任务图转换器,称为MoFormer,用于链路级NPM。具体而言,提出了一种以链路为中心的标记化图表示方法,将每个链路及其邻域信息转换为路由协议引导下的令牌序列。考虑到每个链路的相对重要性,进一步开发了路由感知的中间中心性编码机制,以增强表征令牌的能力。MoFormer利用MMoE与Transformer相结合的优势,实现对多个预测任务的联合学习。在模拟和真实数据集上的实验结果表明,MoFormer在保持优越泛化能力的同时,比现有最先进的基线有了显著的改进。
{"title":"MoFormer: A centrality-aware multi-task graph transformer with multi-gate mixture-of-experts for link-level network performance modeling","authors":"Hanlin Liu ,&nbsp;Aliya Bao ,&nbsp;Mingyue Li ,&nbsp;Yintan Ai ,&nbsp;Hua Li","doi":"10.1016/j.future.2026.108406","DOIUrl":"10.1016/j.future.2026.108406","url":null,"abstract":"<div><div>Link-level network performance modeling (NPM) facilitates efficient traffic control, precise fault localization, and reliable resource management in emerging network paradigms such as Software-Defined Networking and Intent-Based Networking. A variety of models, such as Long Short-Term Memory and Graph Neural Networks (GNNs), are utilized to enhance the effectiveness of NPM. However, a practical NPM requires the generalization ability to adapt to diverse network topologies and prediction tasks without retraining. To meet this requirement, graph Transformer models are a breakthrough by encoding nodes and their structural features into tokens, breaking free from the dependencies on fixed graph structures typical of traditional GNNs. Nevertheless, they mostly focus on node-centric representations, which are insufficient to capture the fine-grained interactions and dependencies between links, thus limiting their applicability in link-level NPM. In this paper, we propose a centrality-aware multi-task graph Transformer with multi-gate mixture-of-experts (MMoE), named MoFormer, for link-level NPM. Specifically, a link-centric tokenized graph representation method is proposed to transform each link and its neighborhood information into a sequence of tokens guided by the routing protocol. A routing-aware betweenness centrality encoding mechanism is further developed to enhance the ability to characterize the tokens considering the relative importance of each link. MoFormer takes advantage of MMoE combined with Transformer to enable joint learning of multiple prediction tasks. Experimental results on both simulated and real-world datasets demonstrate the significant improvements of MoFormer over existing state-of-the-art baselines while maintaining superior generalization ability.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108406"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146047990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A reliability- and latency-driven task allocation framework for workflow applications in the edge-hub-cloud continuum 一个可靠性和延迟驱动的任务分配框架,用于边缘中心云连续体中的工作流应用程序
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-02-03 DOI: 10.1016/j.future.2026.108414
Andreas Kouloumpris , Georgios L. Stavrinides , Maria K. Michael , Theocharis Theocharides
A growing number of critical workflow applications leverage a streamlined edge-hub-cloud architecture, which diverges from the conventional edge computing paradigm. An edge device, in collaboration with a hub device and a cloud server, often suffices for their reliable and efficient execution. However, task allocation in this streamlined architecture is challenging due to device limitations and diverse operating conditions. Given the inherent criticality of such workflow applications, where reliability and latency are vital yet conflicting objectives, an exact task allocation approach is typically required to ensure optimal solutions. As no existing method holistically addresses these issues, we propose an exact multi-objective task allocation framework to jointly optimize the overall reliability and latency of a workflow application in the specific edge-hub-cloud architecture. We present a comprehensive binary integer linear programming formulation that considers the relative importance of each objective. It incorporates time redundancy techniques, while accounting for crucial constraints often overlooked in related studies. We evaluate our approach using a relevant real-world workflow application, as well as synthetic workflows varying in structure, size, and criticality. In the real-world application, our method achieved average improvements of 84.19% in reliability and 49.81% in latency over baseline strategies, across relevant objective trade-offs. Overall, the experimental results demonstrate the effectiveness and scalability of our approach across diverse workflow applications for the considered system architecture, highlighting its practicality with runtimes averaging between 0.03 and 50.94 seconds across all examined workflows.
越来越多的关键工作流应用程序利用了流线型的边缘中心云架构,这与传统的边缘计算范式有所不同。边缘设备与集线器设备和云服务器协作,通常足以实现可靠和高效的执行。然而,由于设备限制和不同的操作条件,这种流线型架构中的任务分配具有挑战性。考虑到这些工作流应用程序固有的关键性,其中可靠性和延迟是至关重要的,但又相互冲突的目标,通常需要精确的任务分配方法来确保最佳解决方案。由于现有方法无法从整体上解决这些问题,我们提出了一个精确的多目标任务分配框架,以共同优化特定边缘中心云架构下工作流应用程序的整体可靠性和延迟。我们提出了一个综合的二元整数线性规划公式,它考虑了每个目标的相对重要性。它结合了时间冗余技术,同时考虑了相关研究中经常被忽视的关键约束。我们使用一个相关的现实世界工作流应用程序来评估我们的方法,以及在结构、大小和重要性上变化的合成工作流。在实际应用中,与基线策略相比,我们的方法在可靠性方面平均提高了84.19%,在延迟方面平均提高了49.81%。总的来说,实验结果证明了我们的方法在考虑的系统架构的不同工作流应用程序中的有效性和可扩展性,突出了它的实用性,在所有检查的工作流中平均运行时间在0.03到50.94秒之间。
{"title":"A reliability- and latency-driven task allocation framework for workflow applications in the edge-hub-cloud continuum","authors":"Andreas Kouloumpris ,&nbsp;Georgios L. Stavrinides ,&nbsp;Maria K. Michael ,&nbsp;Theocharis Theocharides","doi":"10.1016/j.future.2026.108414","DOIUrl":"10.1016/j.future.2026.108414","url":null,"abstract":"<div><div>A growing number of critical workflow applications leverage a streamlined edge-hub-cloud architecture, which diverges from the conventional edge computing paradigm. An edge device, in collaboration with a hub device and a cloud server, often suffices for their reliable and efficient execution. However, task allocation in this streamlined architecture is challenging due to device limitations and diverse operating conditions. Given the inherent criticality of such workflow applications, where reliability and latency are vital yet conflicting objectives, an exact task allocation approach is typically required to ensure optimal solutions. As no existing method holistically addresses these issues, we propose an exact multi-objective task allocation framework to jointly optimize the overall reliability and latency of a workflow application in the specific edge-hub-cloud architecture. We present a comprehensive binary integer linear programming formulation that considers the relative importance of each objective. It incorporates time redundancy techniques, while accounting for crucial constraints often overlooked in related studies. We evaluate our approach using a relevant real-world workflow application, as well as synthetic workflows varying in structure, size, and criticality. In the real-world application, our method achieved average improvements of 84.19% in reliability and 49.81% in latency over baseline strategies, across relevant objective trade-offs. Overall, the experimental results demonstrate the effectiveness and scalability of our approach across diverse workflow applications for the considered system architecture, highlighting its practicality with runtimes averaging between 0.03 and 50.94 seconds across all examined workflows.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108414"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146110188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FedAPE: Heterogeneous federated learning with attention-guided aggregation and prototype enhancement 基于注意引导聚合和原型增强的异质联邦学习
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-02-06 DOI: 10.1016/j.future.2026.108417
Xiao Wang , Zhengda Wu , Jinghua Zhu
Prototype-based heterogeneous federated learning has emerged as an efficient paradigm that reduces communication overhead through prototype transmission while preserving client privacy. However, current approaches suffer from three critical limitations: 1) conventional prototype averaging overlooks intra-class variations; 2) sensitivity to imbalanced class distributions in client feature spaces; and 3) restricted generalization capacity in heterogeneous environments. This paper presents FedAPE - a novel Attention-guided aggregation and Prototype Enhancement framework for heterogeneous federated learning. Our technical contributions are threefold: First, we develop an attention-based feature aggregation mechanism that dynamically captures intra-class similarity patterns to learn discriminative class-specific prototypes. Second, we introduce a contrastive learning objective that explicitly enforces inter-class separability and intra-class compactness during local training. Third, we devise an adaptive prototype enhancement strategy that synthesizes representative pseudo-prototypes in under-represented regions of the feature space, effectively compensating for class distribution imbalance. Comprehensive evaluations on CIFAR-10, CIFAR-100, and GTSRB datasets demonstrate FedAPE’s superiority over state-of-the-art methods, achieving accuracy improvements of 2.82%, 5.65%, and 2.94%, respectively, in heterogeneous model configurations. The results validate our framework’s enhanced capability in handling both feature heterogeneity and class imbalance scenarios.
基于原型的异构联邦学习已经成为一种有效的范例,它通过原型传输减少了通信开销,同时保护了客户端的隐私。然而,目前的方法有三个关键的局限性:1)传统的原型平均忽略了类内的变化;2)对客户端特征空间中不平衡类分布的敏感性;3)异构环境下有限的泛化能力。本文提出了一种新的用于异构联邦学习的注意引导聚合和原型增强框架FedAPE。我们的技术贡献有三个方面:首先,我们开发了一种基于注意力的特征聚合机制,该机制动态捕获类内相似性模式,以学习判别类特定的原型。其次,我们引入了一个对比学习目标,该目标在局部训练期间明确地执行类间可分离性和类内紧密性。第三,我们设计了一种自适应原型增强策略,在特征空间中代表性不足的区域合成具有代表性的伪原型,有效地补偿了类分布的不平衡。对CIFAR-10、CIFAR-100和GTSRB数据集的综合评估表明,FedAPE优于最先进的方法,在异构模型配置下,准确率分别提高了2.82%、5.65%和2.94%。结果验证了我们的框架在处理特征异构和类不平衡场景方面的增强能力。
{"title":"FedAPE: Heterogeneous federated learning with attention-guided aggregation and prototype enhancement","authors":"Xiao Wang ,&nbsp;Zhengda Wu ,&nbsp;Jinghua Zhu","doi":"10.1016/j.future.2026.108417","DOIUrl":"10.1016/j.future.2026.108417","url":null,"abstract":"<div><div>Prototype-based heterogeneous federated learning has emerged as an efficient paradigm that reduces communication overhead through prototype transmission while preserving client privacy. However, current approaches suffer from three critical limitations: 1) conventional prototype averaging overlooks intra-class variations; 2) sensitivity to imbalanced class distributions in client feature spaces; and 3) restricted generalization capacity in heterogeneous environments. This paper presents FedAPE - a novel Attention-guided aggregation and Prototype Enhancement framework for heterogeneous federated learning. Our technical contributions are threefold: First, we develop an attention-based feature aggregation mechanism that dynamically captures intra-class similarity patterns to learn discriminative class-specific prototypes. Second, we introduce a contrastive learning objective that explicitly enforces inter-class separability and intra-class compactness during local training. Third, we devise an adaptive prototype enhancement strategy that synthesizes representative pseudo-prototypes in under-represented regions of the feature space, effectively compensating for class distribution imbalance. Comprehensive evaluations on CIFAR-10, CIFAR-100, and GTSRB datasets demonstrate FedAPE’s superiority over state-of-the-art methods, achieving accuracy improvements of 2.82%, 5.65%, and 2.94%, respectively, in heterogeneous model configurations. The results validate our framework’s enhanced capability in handling both feature heterogeneity and class imbalance scenarios.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108417"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146134533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SeHP-CSQ: A secure, high-performance cross-shard queuing model SeHP-CSQ:一个安全、高性能的跨分片队列模型
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-07-01 Epub Date: 2026-01-20 DOI: 10.1016/j.future.2026.108376
Hui Dai , Lingyun Yuan , Haochen Bao , Han Chen
Blockchain sharding parallelises processing to boost throughput. Cross-shard transactions’ low transmission efficiency and security risks limit system scalability. We propose a secure cross-shard high-performance processing queuing model. First, we model hybrid multi-distribution batch arrival-processing and accurately depict transaction arrival and processing dynamics. Second, we construct a cross-shard transaction processing queuing model based on M/M/1/N queuing, along with a metric system for key performance indicators. Modifying the queue capacity to regulate batch control of cross-shard transactions directed at the target shard, thereby improving robustness and scalability. Third, we design a dynamic adaptive malicious transaction analysis bound, which derives an upper bound on the real-time tail probability via Chernoff’s inequality and Hoeffding’s inequality, and prove that the analysis bound can converge at an exponential rate under any shard size, thus effectively limiting the impact of malicious behaviours on the security of the shard system. Experimental results show that the proposed queuing model can reach a maximum throughput of about 8.0 × 104 TPS and achieve load balancing in high concurrency scenarios. The queuing waiting time is less than 0.5 ms, with the overload probability and the system failure probability converging to 0%, which verifies that the model has adequate security While ensuring high processing efficiency.
区块链分片并行处理以提高吞吐量。跨分片交易传输效率低,存在安全风险,限制了系统的可扩展性。提出了一种安全的跨分片高性能处理队列模型。首先,建立了混合多分布批量到达处理模型,准确描述了事务到达和处理动态。其次,构建了基于M/M/1/N排队的跨分片事务处理排队模型,并给出了关键性能指标的度量体系。修改队列容量,以调节针对目标分片的跨分片事务的批处理控制,从而提高鲁棒性和可伸缩性。第三,设计了一个动态自适应的恶意交易分析界,通过Chernoff不等式和Hoeffding不等式推导出实时尾概率的上界,并证明了分析界在任意分片大小下都能以指数速率收敛,从而有效地限制了恶意行为对分片系统安全性的影响。实验结果表明,所提出的队列模型可以达到8.0 × 104 TPS左右的最大吞吐量,并在高并发场景下实现负载均衡。排队等待时间小于0.5 ms,过载概率和系统故障概率收敛到0%,验证了该模型在保证高处理效率的同时具有足够的安全性。
{"title":"SeHP-CSQ: A secure, high-performance cross-shard queuing model","authors":"Hui Dai ,&nbsp;Lingyun Yuan ,&nbsp;Haochen Bao ,&nbsp;Han Chen","doi":"10.1016/j.future.2026.108376","DOIUrl":"10.1016/j.future.2026.108376","url":null,"abstract":"<div><div>Blockchain sharding parallelises processing to boost throughput. Cross-shard transactions’ low transmission efficiency and security risks limit system scalability. We propose a secure cross-shard high-performance processing queuing model. First, we model hybrid multi-distribution batch arrival-processing and accurately depict transaction arrival and processing dynamics. Second, we construct a cross-shard transaction processing queuing model based on M/M/1/N queuing, along with a metric system for key performance indicators. Modifying the queue capacity to regulate batch control of cross-shard transactions directed at the target shard, thereby improving robustness and scalability. Third, we design a dynamic adaptive malicious transaction analysis bound, which derives an upper bound on the real-time tail probability via Chernoff’s inequality and Hoeffding’s inequality, and prove that the analysis bound can converge at an exponential rate under any shard size, thus effectively limiting the impact of malicious behaviours on the security of the shard system. Experimental results show that the proposed queuing model can reach a maximum throughput of about 8.0 × 10<sup>4</sup> TPS and achieve load balancing in high concurrency scenarios. The queuing waiting time is less than 0.5 ms, with the overload probability and the system failure probability converging to 0%, which verifies that the model has adequate security While ensuring high processing efficiency.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108376"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146014877","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Future Generation Computer Systems-The International Journal of Escience
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1