首页 > 最新文献

Future Generation Computer Systems-The International Journal of Escience最新文献

英文 中文
Leveraging cutting-edge high performance computing for large-scale applications 利用尖端的高性能计算大规模应用程序
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2026-01-13 DOI: 10.1016/j.future.2026.108374
Claude Tadonki , Gabriele Mencagli , Leonel Sousa
High Performance Computing (HPC) recently entered into the exascale era, marking an important milestone of its history. High-end supercomputers and clusters with remarkable levels of performance are now commonly available for general and specific computational needs, thereby increasing the focus on HPC and related topics. Leveraging the potential of high-speed processing units is an HPC skillful task that requires in-depth knowledge in both hardware and software domains. In fact, the architectural structure of cutting-edge HPC processors is complex and involves several specialized features provided through specific units/mechanisms, the processing constraint/overhead of which can turn out to be an efficiency bottleneck. Large-scale supercomputers present greater challenges due to the significant overhead associated with interprocessor communication and synchronization. The evolution of HPC appears closely tied to the growing demand for speed from large-scale applications like complex combinatorial problems, big data applications, the training of large-scale AI models and high-precision simulations, to name a few. As a result, the implementation of cutting-edge techniques should remain scalable on large-scale machines for the benefit of end-users.
高性能计算(HPC)最近进入了百亿亿次时代,标志着其历史上的一个重要里程碑。具有卓越性能水平的高端超级计算机和集群现在通常可用于一般和特定的计算需求,从而增加了对HPC和相关主题的关注。利用高速处理单元的潜力是一项HPC技术任务,需要在硬件和软件领域有深入的知识。事实上,尖端HPC处理器的体系结构是复杂的,并且涉及到通过特定单元/机制提供的一些专门功能,其处理约束/开销可能成为效率瓶颈。由于与处理器间通信和同步相关的巨大开销,大型超级计算机提出了更大的挑战。HPC的发展似乎与大规模应用(如复杂的组合问题、大数据应用、大规模人工智能模型的训练和高精度模拟)对速度的日益增长的需求密切相关。因此,为了最终用户的利益,尖端技术的实现应该在大型机器上保持可伸缩性。
{"title":"Leveraging cutting-edge high performance computing for large-scale applications","authors":"Claude Tadonki ,&nbsp;Gabriele Mencagli ,&nbsp;Leonel Sousa","doi":"10.1016/j.future.2026.108374","DOIUrl":"10.1016/j.future.2026.108374","url":null,"abstract":"<div><div>High Performance Computing (HPC) recently entered into the exascale era, marking an important milestone of its history. High-end supercomputers and clusters with remarkable levels of performance are now commonly available for general and specific computational needs, thereby increasing the focus on HPC and related topics. Leveraging the potential of high-speed processing units is an HPC skillful task that requires in-depth knowledge in both hardware and software domains. In fact, the architectural structure of cutting-edge HPC processors is complex and involves several specialized features provided through specific units/mechanisms, the processing constraint/overhead of which can turn out to be an efficiency bottleneck. Large-scale supercomputers present greater challenges due to the significant overhead associated with interprocessor communication and synchronization. The evolution of HPC appears closely tied to the growing demand for speed from large-scale applications like complex combinatorial problems, big data applications, the training of large-scale AI models and high-precision simulations, to name a few. As a result, the implementation of cutting-edge techniques should remain scalable on large-scale machines for the benefit of end-users.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108374"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145962155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A heuristic approach to Spark workflow task scheduling on heterogeneous nodes 异构节点上Spark工作流任务调度的启发式方法
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2025-12-23 DOI: 10.1016/j.future.2025.108335
Mehboob Hussain , Ying Xu , Zeeshan Abbas , Ali Kamran , Amir Rehman , Muhammad Yasir
Big data applications often use workflows shown as Directed Acyclic Graphs (DAGs). These DAGs link jobs and stages with order rules. Regular Spark-based workflow scheduling assumes all executors are identical and expects jobs to run one after another. This approach fails in mixed cloud setups, where machines vary in speed and computing capabilities. Efficient workflow scheduling in heterogeneous Spark nodes is crucial for reducing makespan and achieving load balancing. However, many existing methods overlook the challenge created by DAG-constrained jobs, stage-level dependencies, and varied node performance. This paper addresses efficient workflow scheduling on a mixed Spark cluster. We seek to minimize workflow completion time while keeping node loads balanced. We propose a modified Spark framework. It includes both a job scheduler and a stage scheduler tailored for mixed setups. Our method introduces multi-level node classification based on load status, a Speculative Stage Execution strategy for dynamic scheduling, and a Node Awareness Strategy for real-time task assignment. Compared to Rainbow, SAF, and DSWTS algorithms, SWTS reduces makespan by up to 40%, improves load balancing by 55%, and increases resource utilization by 20%. This demonstrates superior efficiency across all workflows.
大数据应用通常使用有向无环图(dag)表示的工作流。这些dag将作业和阶段与顺序规则联系起来。常规的基于spark的工作流调度假设所有执行器都是相同的,并期望作业一个接一个地运行。这种方法在混合云设置中失败,因为机器的速度和计算能力各不相同。在异构Spark节点中高效的工作流调度对于减少makespan和实现负载平衡至关重要。然而,许多现有方法忽略了受dag约束的作业、阶段级依赖关系和不同节点性能所带来的挑战。本文研究了在混合Spark集群上的高效工作流调度。我们力求在保持节点负载平衡的同时最小化工作流完成时间。我们提出了一个修改后的Spark框架。它包括一个作业调度器和一个为混合设置定制的阶段调度器。该方法引入了基于负载状态的多级节点分类、用于动态调度的推测阶段执行策略和用于实时任务分配的节点感知策略。与Rainbow、SAF和DSWTS算法相比,SWTS将完工时间缩短了40%,将负载平衡提高了55%,并将资源利用率提高了20%。这证明了跨所有工作流的卓越效率。
{"title":"A heuristic approach to Spark workflow task scheduling on heterogeneous nodes","authors":"Mehboob Hussain ,&nbsp;Ying Xu ,&nbsp;Zeeshan Abbas ,&nbsp;Ali Kamran ,&nbsp;Amir Rehman ,&nbsp;Muhammad Yasir","doi":"10.1016/j.future.2025.108335","DOIUrl":"10.1016/j.future.2025.108335","url":null,"abstract":"<div><div>Big data applications often use workflows shown as Directed Acyclic Graphs (DAGs). These DAGs link jobs and stages with order rules. Regular Spark-based workflow scheduling assumes all executors are identical and expects jobs to run one after another. This approach fails in mixed cloud setups, where machines vary in speed and computing capabilities. Efficient workflow scheduling in heterogeneous Spark nodes is crucial for reducing makespan and achieving load balancing. However, many existing methods overlook the challenge created by DAG-constrained jobs, stage-level dependencies, and varied node performance. This paper addresses efficient workflow scheduling on a mixed Spark cluster. We seek to minimize workflow completion time while keeping node loads balanced. We propose a modified Spark framework. It includes both a job scheduler and a stage scheduler tailored for mixed setups. Our method introduces multi-level node classification based on load status, a Speculative Stage Execution strategy for dynamic scheduling, and a Node Awareness Strategy for real-time task assignment. Compared to Rainbow, SAF, and DSWTS algorithms, SWTS reduces makespan by up to 40%, improves load balancing by 55%, and increases resource utilization by 20%. This demonstrates superior efficiency across all workflows.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108335"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145823158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging big data and cloud technology for scalable and interoperable smart farming 利用大数据和云技术实现可扩展和可互操作的智能农业
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2025-12-21 DOI: 10.1016/j.future.2025.108324
Amine Roukh, Saïd Mahmoudi
Agriculture faces escalating demands to increase food production amid shrinking arable land, resource depletion, and climate variability. Existing smart farming solutions often lack scalability, interoperability, real-time analytics, and region-specific adaptability. This paper presents WALLeSmart, a cloud-based smart farming platform designed to address these challenges through a scalable Lambda architecture and a modular plugin system. Hosted on a GDPR-compliant private cloud, WALLeSmart integrates diverse data sources (e.g., IoT sensors, satellite imagery, weather data) to deliver real-time insights and predictive analytics, achieving low-latency processing (e.g., 80 seconds for weather data streams). Key features include a one-stop shop for accessing agricultural platforms (e.g., Myawenet, MyCDL, Cerise), a consent management system for data control, a Walloon Agricultural DataHub for secure data exchange, and a personalized dashboard for farmers. The platform’s unique governance model, led by farmers, ensures autonomy and transparency. Real-world case studies in Wallonia, Belgium, demonstrate its ability to process over 3 million weather measurements and 61,130 dairy farm datasets, supporting applications like SALVE, W@llHerbe, and MyFieldBook. WALLeSmart’s generalizable design enables adaptation to diverse regions, addressing ethical concerns like algorithmic bias and data ownership through transparent AI and user-centric consent mechanisms, fostering efficiency, sustainability, and profitability.
在耕地减少、资源枯竭和气候变化的背景下,农业面临着不断增长的粮食增产需求。现有的智能农业解决方案往往缺乏可扩展性、互操作性、实时分析和特定地区的适应性。本文介绍了WALLeSmart,这是一个基于云的智能农业平台,旨在通过可扩展的Lambda架构和模块化插件系统解决这些挑战。WALLeSmart托管在符合gdpr的私有云上,集成了各种数据源(例如,物联网传感器,卫星图像,天气数据),以提供实时洞察和预测分析,实现低延迟处理(例如,天气数据流80秒)。主要功能包括用于访问农业平台的一站式商店(例如Myawenet, MyCDL, Cerise),用于数据控制的同意管理系统,用于安全数据交换的Walloon农业数据中心,以及用于农民的个性化仪表板。该平台由农民主导的独特治理模式确保了自主性和透明度。在比利时瓦隆的实际案例研究表明,它能够处理超过300万个天气测量和61130个奶牛场数据集,支持SALVE, W@llHerbe和MyFieldBook等应用程序。WALLeSmart的通用设计能够适应不同地区,通过透明的人工智能和以用户为中心的同意机制解决算法偏见和数据所有权等道德问题,提高效率、可持续性和盈利能力。
{"title":"Leveraging big data and cloud technology for scalable and interoperable smart farming","authors":"Amine Roukh,&nbsp;Saïd Mahmoudi","doi":"10.1016/j.future.2025.108324","DOIUrl":"10.1016/j.future.2025.108324","url":null,"abstract":"<div><div>Agriculture faces escalating demands to increase food production amid shrinking arable land, resource depletion, and climate variability. Existing smart farming solutions often lack scalability, interoperability, real-time analytics, and region-specific adaptability. This paper presents <em>WALLeSmart</em>, a cloud-based smart farming platform designed to address these challenges through a scalable Lambda architecture and a modular plugin system. Hosted on a GDPR-compliant private cloud, <em>WALLeSmart</em> integrates diverse data sources (e.g., IoT sensors, satellite imagery, weather data) to deliver real-time insights and predictive analytics, achieving low-latency processing (e.g., 80 seconds for weather data streams). Key features include a one-stop shop for accessing agricultural platforms (e.g., Myawenet, MyCDL, Cerise), a consent management system for data control, a Walloon Agricultural DataHub for secure data exchange, and a personalized dashboard for farmers. The platform’s unique governance model, led by farmers, ensures autonomy and transparency. Real-world case studies in Wallonia, Belgium, demonstrate its ability to process over 3 million weather measurements and 61,130 dairy farm datasets, supporting applications like <em>SALVE, W@llHerbe</em>, and <em>MyFieldBook. WALLeSmart</em>’s generalizable design enables adaptation to diverse regions, addressing ethical concerns like algorithmic bias and data ownership through transparent AI and user-centric consent mechanisms, fostering efficiency, sustainability, and profitability.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108324"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145796206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SP-IFDA-Traj: Optimizing differentially private trajectory publishing for enhanced utility SP-IFDA-Traj:优化不同的私人轨迹发布增强效用
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2026-01-05 DOI: 10.1016/j.future.2025.108359
Yongxin Zhao , Hao Lin , Xiangtian Zheng , Yixuan Song , Jinze Du
Differential privacy (DP) is crucial for trajectory data publication, yet the utility of existing prefix-tree-based mechanisms heavily depends on intricate parameter tuning, often yielding suboptimal performance. To address this, we propose SP-IFDA-Traj, a novel framework for automatically and jointly optimizing critical parameters in personalized noisy prefix-tree trajectory publishing. SP-IFDA-Traj enhances the Flow Direction Algorithm (FDA) for optimizing parameters by: 1) employing chaotic mapping initialization for improved population diversity, 2) incorporating adaptive neighborhood generation to balance exploration and exploitation, and 3) leveraging Spark-based parallelized fitness evaluation for enhanced efficiency. Guided by a meticulously designed fitness function targeting maximal data utility, our framework optimizes Hilbert encoding order, privacy budget allocation, and pruning strategies. Extensive experiments on real-world large-scale trajectory datasets demonstrate that SP-IFDA-Traj substantially improves the privacy-utility trade-off. In convergence tests, the optimization engine improves the average fitness over FDA by 1.74% on BJ-Day3 and 103.98% on BJ-Day7, indicating consistently superior convergence across heterogeneous datasets. In terms of trajectory query accuracy, SP-IFDA-Traj reduces error to about 1% of that of baseline methods and approximately 10% of that of other existing optimization strategy models.
差分隐私(DP)对于轨迹数据发布至关重要,但现有基于前缀树的机制的效用严重依赖于复杂的参数调优,通常会产生次优性能。为了解决这个问题,我们提出了SP-IFDA-Traj,这是一个新的框架,用于自动和联合优化个性化噪声前缀树轨迹发布中的关键参数。SP-IFDA-Traj对Flow Direction Algorithm (FDA)的参数优化进行了改进:1)采用混沌映射初始化来提高种群多样性;2)采用自适应邻域生成来平衡探索和开发;3)利用基于spark的并行适应度评估来提高效率。在精心设计的适应度函数的指导下,以最大的数据效用为目标,我们的框架优化了希尔伯特编码顺序、隐私预算分配和修剪策略。在真实世界的大规模轨迹数据集上进行的大量实验表明,SP-IFDA-Traj极大地改善了隐私-效用权衡。在收敛测试中,优化引擎在BJ-Day3上的平均适应度提高了1.74%,在BJ-Day7上提高了103.98%,表明在异构数据集上始终具有优异的收敛性。在轨迹查询精度方面,SP-IFDA-Traj将误差降低到基线方法的1%左右,将误差降低到其他现有优化策略模型的10%左右。
{"title":"SP-IFDA-Traj: Optimizing differentially private trajectory publishing for enhanced utility","authors":"Yongxin Zhao ,&nbsp;Hao Lin ,&nbsp;Xiangtian Zheng ,&nbsp;Yixuan Song ,&nbsp;Jinze Du","doi":"10.1016/j.future.2025.108359","DOIUrl":"10.1016/j.future.2025.108359","url":null,"abstract":"<div><div>Differential privacy (DP) is crucial for trajectory data publication, yet the utility of existing prefix-tree-based mechanisms heavily depends on intricate parameter tuning, often yielding suboptimal performance. To address this, we propose SP-IFDA-Traj, a novel framework for automatically and jointly optimizing critical parameters in personalized noisy prefix-tree trajectory publishing. SP-IFDA-Traj enhances the Flow Direction Algorithm (FDA) for optimizing parameters by: 1) employing chaotic mapping initialization for improved population diversity, 2) incorporating adaptive neighborhood generation to balance exploration and exploitation, and 3) leveraging Spark-based parallelized fitness evaluation for enhanced efficiency. Guided by a meticulously designed fitness function targeting maximal data utility, our framework optimizes Hilbert encoding order, privacy budget allocation, and pruning strategies. Extensive experiments on real-world large-scale trajectory datasets demonstrate that SP-IFDA-Traj substantially improves the privacy-utility trade-off. In convergence tests, the optimization engine improves the average fitness over FDA by 1.74% on BJ-Day3 and 103.98% on BJ-Day7, indicating consistently superior convergence across heterogeneous datasets. In terms of trajectory query accuracy, SP-IFDA-Traj reduces error to about 1% of that of baseline methods and approximately 10% of that of other existing optimization strategy models.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108359"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145897246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
COZO : A secure and efficient blockchain-enhanced federated learning paradigm with optimized storage and equitable contribution valuation COZO:一种安全高效的区块链增强联邦学习范式,具有优化的存储和公平的贡献评估
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2026-01-05 DOI: 10.1016/j.future.2025.108362
Ning Liu , Yizhi Zhou , Yuchen Qin , Qinzheng Feng , Heng Qi
Federated Learning (FL) serves as a crucial enabler for digital supply chain transformation by supporting secure multi-party data collaboration while preserving privacy. However, its real-world implementation encounters two critical limitations: (1) conventional contribution assessment approaches (e.g., Shapley value) often ignore individual rationality, resulting in biased incentives that undermine long-term engagement; and (2) the underlying blockchain storage infrastructure suffers from inherent performance constraints, including slow read/write operations that cannot keep pace with FL’s need for real-time data exchange and frequent model updates. To overcome these challenges, we propose a framework that leverages Clique Games Solution theory to ensure fair and efficient contribution measurement. Our solution incorporates coalition screening and Pareto optimization to guarantee individual rationality with polynomial computational complexity. Additionally, we integrate COLE-a columnar storage module enhanced with learned indexes and LSM-tree merge mechanisms-to dramatically accelerate data access. Experimental results demonstrate that our approach achieves 5.72e-4 reduction in core distance for contribution assessment stability, improves storage IOPS by 28.7%, and maintains model accuracy within 1.2% of the baseline methods, outperforming mainstream baselines in evaluation stability and storage performance while maintaining competitive model accuracy, thus offering a comprehensive and practical solution for privacy-aware distributed learning in sensitive environments.
联邦学习(FL)通过支持安全的多方数据协作,同时保护隐私,成为数字供应链转型的关键推动者。然而,它在现实世界的实施遇到了两个关键的限制:(1)传统的贡献评估方法(如Shapley值)往往忽视个人理性,导致有偏见的激励,破坏了长期参与;(2)底层区块链存储基础设施受到固有的性能限制,包括读写操作缓慢,无法跟上FL对实时数据交换和频繁模型更新的需求。为了克服这些挑战,我们提出了一个利用派系博弈解决方案理论的框架,以确保公平和有效的贡献测量。我们的解决方案结合了联盟筛选和帕累托优化,以多项式的计算复杂度保证了个体的合理性。此外,我们还集成了cole(一种增强了学习索引和lsm树合并机制的列式存储模块),以显著加快数据访问速度。实验结果表明,我们的方法在贡献评估稳定性方面的核心距离减少了5.72e-4,存储IOPS提高了28.7%,模型精度保持在基线方法的1.2%以内,在评估稳定性和存储性能方面优于主流基线,同时保持了具有竞争力的模型精度,从而为敏感环境下的隐私感知分布式学习提供了全面实用的解决方案。
{"title":"COZO : A secure and efficient blockchain-enhanced federated learning paradigm with optimized storage and equitable contribution valuation","authors":"Ning Liu ,&nbsp;Yizhi Zhou ,&nbsp;Yuchen Qin ,&nbsp;Qinzheng Feng ,&nbsp;Heng Qi","doi":"10.1016/j.future.2025.108362","DOIUrl":"10.1016/j.future.2025.108362","url":null,"abstract":"<div><div>Federated Learning (FL) serves as a crucial enabler for digital supply chain transformation by supporting secure multi-party data collaboration while preserving privacy. However, its real-world implementation encounters two critical limitations: (1) conventional contribution assessment approaches (e.g., Shapley value) often ignore individual rationality, resulting in biased incentives that undermine long-term engagement; and (2) the underlying blockchain storage infrastructure suffers from inherent performance constraints, including slow read/write operations that cannot keep pace with FL’s need for real-time data exchange and frequent model updates. To overcome these challenges, we propose a framework that leverages Clique Games Solution theory to ensure fair and efficient contribution measurement. Our solution incorporates coalition screening and Pareto optimization to guarantee individual rationality with polynomial computational complexity. Additionally, we integrate COLE-a columnar storage module enhanced with learned indexes and LSM-tree merge mechanisms-to dramatically accelerate data access. Experimental results demonstrate that our approach achieves 5.72e-4 reduction in core distance for contribution assessment stability, improves storage IOPS by 28.7%, and maintains model accuracy within 1.2% of the baseline methods, outperforming mainstream baselines in evaluation stability and storage performance while maintaining competitive model accuracy, thus offering a comprehensive and practical solution for privacy-aware distributed learning in sensitive environments.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108362"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145902511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploiting Modular Redundancy for approximating Random Forest classifiers 利用模冗余逼近随机森林分类器
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2025-12-18 DOI: 10.1016/j.future.2025.108330
Antonio Emmanuele , Mario Barbareschi , Alberto Bosio
The deployment of machine learning models at the edge is crucial for enabling low-latency decision-making, optimizing resource utilization, and enhancing data confidentiality. Random Forest classifiers have proven to be highly accurate while offering computationally efficient inference, making them well-suited for resource-constrained edge devices. However, as the volume of training data grows, the complexity and size of these models also increase, limiting their deployment in edge computing scenarios. In order to address this challenge, we propose a novel approximation strategy for Random Forest classifiers leveraging on the concept of modular redundancy. In particular, our approach imposes that each target class is determined by only a subset of trees in a modular redundant fashion. This allows to prune from each tree the leaves related to no-longer relevant classes, significantly reducing the size of the model. To achieve an optimal balance between accuracy and resource savings with minimal computational time, we introduce an heuristic algorithm that determine the best subset of trees for each class. We evaluate our approach on multiple UCI machine learning datasets using a hardware accelerator for tree ensembles, demonstrating its effectiveness. The result shows that, on average, a 2.5 % reduction in accuracy leads to save up to 50 % in hardware overhead and energy consumption.
在边缘部署机器学习模型对于实现低延迟决策、优化资源利用和增强数据机密性至关重要。随机森林分类器已被证明是高度准确的,同时提供计算效率推断,使它们非常适合资源受限的边缘设备。然而,随着训练数据量的增长,这些模型的复杂性和规模也在增加,限制了它们在边缘计算场景中的部署。为了解决这一挑战,我们提出了一种利用模块化冗余概念的随机森林分类器的新近似策略。特别是,我们的方法强制每个目标类仅由树的一个子集以模块化冗余的方式确定。这允许从每棵树上修剪与不再相关的类相关的叶子,从而显着减小模型的大小。为了在最小的计算时间内实现准确性和资源节约之间的最佳平衡,我们引入了一种启发式算法,用于确定每个类的最佳树子集。我们使用树集成硬件加速器在多个UCI机器学习数据集上评估了我们的方法,证明了它的有效性。结果表明,平均而言,精确度降低2.5%可节省高达50%的硬件开销和能耗。
{"title":"Exploiting Modular Redundancy for approximating Random Forest classifiers","authors":"Antonio Emmanuele ,&nbsp;Mario Barbareschi ,&nbsp;Alberto Bosio","doi":"10.1016/j.future.2025.108330","DOIUrl":"10.1016/j.future.2025.108330","url":null,"abstract":"<div><div>The deployment of machine learning models at the edge is crucial for enabling low-latency decision-making, optimizing resource utilization, and enhancing data confidentiality. Random Forest classifiers have proven to be highly accurate while offering computationally efficient inference, making them well-suited for resource-constrained edge devices. However, as the volume of training data grows, the complexity and size of these models also increase, limiting their deployment in edge computing scenarios. In order to address this challenge, we propose a novel approximation strategy for Random Forest classifiers leveraging on the concept of modular redundancy. In particular, our approach imposes that each target class is determined by only a subset of trees in a modular redundant fashion. This allows to prune from each tree the leaves related to no-longer relevant classes, significantly reducing the size of the model. To achieve an optimal balance between accuracy and resource savings with minimal computational time, we introduce an heuristic algorithm that determine the best subset of trees for each class. We evaluate our approach on multiple UCI machine learning datasets using a hardware accelerator for tree ensembles, demonstrating its effectiveness. The result shows that, on average, a 2.5 % reduction in accuracy leads to save up to 50 % in hardware overhead and energy consumption.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108330"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145785036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
interTwin: Advancing Scientific Digital Twins through AI, Federated Computing and Data interTwin:通过人工智能、联邦计算和数据推进科学数字双胞胎
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2025-12-15 DOI: 10.1016/j.future.2025.108312
Andrea Manzi , Raul Bardaji , Ivan Rodero , Germán Moltó , Sandro Fiore , Isabel Campos , Donatello Elia , Francesco Sarandrea , A. Paul Millar , Daniele Spiga , Matteo Bunino , Gabriele Accarino , Lorenzo Asprea , Samuel Bernardo , Miguel Caballer , Charis Chatzikyriakou , Diego Ciangottini , Michele Claus , Andrea Cristofori , Davide Donno , Juraj Zvolensky
The EU project interTwin, co-designed and implemented the prototype of an interdisciplinary Digital Twin Engine (DTE), an open-source platform that provides generic and domain-specific software components for modelling and simulation to integrate application-specific Digital Twins (DTs). The DTE is built upon a co-designed conceptual model - the DTE blueprint architecture - guided by open standards and interoperability principles. The ambition is to develop a unified approach to the implementation of DTs that is applicable across diverse scientific disciplines to foster collaborations and facilitate developments. Co-design involved DT use cases from high-energy physics, radio astronomy, astroparticle physics, climate research, and environmental monitoring, which drove advancements in modelling and simulation by leveraging heterogeneous distributed digital infrastructures, enabling dynamic workflow composition, real-time data management and processing, quality and uncertainty tracing of models, and multi-source data fusion.
欧盟interTwin项目共同设计并实现了跨学科数字孪生引擎(DTE)的原型,这是一个开源平台,提供通用和特定领域的软件组件,用于建模和仿真,以集成特定应用的数字孪生(dt)。DTE建立在一个共同设计的概念模型上——DTE蓝图体系结构——由开放标准和互操作性原则指导。其目标是制定一种适用于不同科学学科的统一方法来实施直接临床试验,以促进合作和促进发展。协同设计涉及来自高能物理、射电天文学、天体粒子物理、气候研究和环境监测的DT用例,通过利用异构分布式数字基础设施,实现动态工作流组成、实时数据管理和处理、模型的质量和不确定性跟踪以及多源数据融合,推动了建模和仿真的进步。
{"title":"interTwin: Advancing Scientific Digital Twins through AI, Federated Computing and Data","authors":"Andrea Manzi ,&nbsp;Raul Bardaji ,&nbsp;Ivan Rodero ,&nbsp;Germán Moltó ,&nbsp;Sandro Fiore ,&nbsp;Isabel Campos ,&nbsp;Donatello Elia ,&nbsp;Francesco Sarandrea ,&nbsp;A. Paul Millar ,&nbsp;Daniele Spiga ,&nbsp;Matteo Bunino ,&nbsp;Gabriele Accarino ,&nbsp;Lorenzo Asprea ,&nbsp;Samuel Bernardo ,&nbsp;Miguel Caballer ,&nbsp;Charis Chatzikyriakou ,&nbsp;Diego Ciangottini ,&nbsp;Michele Claus ,&nbsp;Andrea Cristofori ,&nbsp;Davide Donno ,&nbsp;Juraj Zvolensky","doi":"10.1016/j.future.2025.108312","DOIUrl":"10.1016/j.future.2025.108312","url":null,"abstract":"<div><div>The EU project interTwin, co-designed and implemented the prototype of an interdisciplinary Digital Twin Engine (DTE), an open-source platform that provides generic and domain-specific software components for modelling and simulation to integrate application-specific Digital Twins (DTs). The DTE is built upon a co-designed conceptual model - the DTE blueprint architecture - guided by open standards and interoperability principles. The ambition is to develop a unified approach to the implementation of DTs that is applicable across diverse scientific disciplines to foster collaborations and facilitate developments. Co-design involved DT use cases from high-energy physics, radio astronomy, astroparticle physics, climate research, and environmental monitoring, which drove advancements in modelling and simulation by leveraging heterogeneous distributed digital infrastructures, enabling dynamic workflow composition, real-time data management and processing, quality and uncertainty tracing of models, and multi-source data fusion.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108312"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145785045","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Model partitioning and batch scheduling: Leveraging local resources for cost-efficient device-cloud collaborative serverless inference 模型分区和批调度:利用本地资源实现低成本的设备云协同无服务器推理
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2025-12-23 DOI: 10.1016/j.future.2025.108338
Yong Pan , Chenchen Zhang , Yue Zeng , Haiyu Yue , Ziye Hou , Jiang Xiong
In recent years, serverless computing has become a new paradigm for efficient and low-cost model inference due to its advantages in dynamic scaling and fine-grained resource allocation. Dynamic batching improves resource utilization by dynamically aggregating inference requests to balance latency and throughput, while adaptive model partitioning divides models into lightweight edge modules and cloud-intensive computing units to optimize global resources through collaborative architecture, driving the optimization and performance enhancement of intelligent inference. However, existing model partitioning methods do not consider the resource utilization of low-cost user local devices, which may lead to insufficient utilization of local resources and increased operational costs. This paper explores how to optimize batch processing decisions and model partitioning strategies in a device-cloud collaborative scenario under resource constraints, aiming to maximize local resource utilization, minimize costs, and improve inference efficiency. To characterize the mathematical relationship between batch processing decisions, model partitioning strategies, and delay-cost combination decisions, we formulate the problem as a mixed integer nonlinear programming model and prove its NP-hardness. We introduce the Model Partitioning and Batch Scheduling Algorithm (MPBS), a serverless-accelerated inference mechanism that leverages dynamic Deep Neural Network (DNN) model partitioning. This architecture generates resource utilization efficient partitioning schemes according to heterogeneous hardware characteristics and request priorities, ensuring efficient device-task alignment. Additionally, a dynamic parallel scheduling mechanism, driven by a scheduling engine, enables global resource optimization by leveraging the batch processing capabilities of cloud instances, collectively enhancing local resource utilization and accelerating cloud container performance. Extensive simulation results demonstrate that compared to state-of-the-art solutions, MPBS reduces resource overhead by 20.6% while satisfying the execution time specified by Service Level Objective (SLO).
近年来,无服务器计算以其动态扩展和细粒度资源分配的优势,成为高效、低成本模型推理的新范式。动态批处理通过动态聚合推理请求来平衡延迟和吞吐量,提高资源利用率;自适应模型划分将模型划分为轻量级边缘模块和云密集型计算单元,通过协同架构优化全局资源,推动智能推理的优化和性能提升。然而,现有的模型划分方法没有考虑低成本用户本地设备的资源利用率,可能导致本地资源利用率不足,增加运营成本。本文探讨了在资源约束下,如何优化设备-云协同场景下的批处理决策和模型划分策略,以最大化本地资源利用率,最小化成本,提高推理效率。为了描述批处理决策、模型划分策略和延迟成本组合决策之间的数学关系,我们将问题表述为一个混合整数非线性规划模型,并证明了其np -硬度。我们介绍了模型分区和批调度算法(MPBS),这是一种利用动态深度神经网络(DNN)模型分区的无服务器加速推理机制。该体系结构根据异构硬件特性和请求优先级生成资源利用率高的分区方案,保证了设备-任务的高效对齐。此外,由调度引擎驱动的动态并行调度机制通过利用云实例的批处理能力来实现全局资源优化,从而共同提高本地资源利用率并加速云容器性能。广泛的仿真结果表明,与最先进的解决方案相比,MPBS在满足服务水平目标(SLO)指定的执行时间的同时,减少了20.6%的资源开销。
{"title":"Model partitioning and batch scheduling: Leveraging local resources for cost-efficient device-cloud collaborative serverless inference","authors":"Yong Pan ,&nbsp;Chenchen Zhang ,&nbsp;Yue Zeng ,&nbsp;Haiyu Yue ,&nbsp;Ziye Hou ,&nbsp;Jiang Xiong","doi":"10.1016/j.future.2025.108338","DOIUrl":"10.1016/j.future.2025.108338","url":null,"abstract":"<div><div>In recent years, serverless computing has become a new paradigm for efficient and low-cost model inference due to its advantages in dynamic scaling and fine-grained resource allocation. Dynamic batching improves resource utilization by dynamically aggregating inference requests to balance latency and throughput, while adaptive model partitioning divides models into lightweight edge modules and cloud-intensive computing units to optimize global resources through collaborative architecture, driving the optimization and performance enhancement of intelligent inference. However, existing model partitioning methods do not consider the resource utilization of low-cost user local devices, which may lead to insufficient utilization of local resources and increased operational costs. This paper explores how to optimize batch processing decisions and model partitioning strategies in a device-cloud collaborative scenario under resource constraints, aiming to maximize local resource utilization, minimize costs, and improve inference efficiency. To characterize the mathematical relationship between batch processing decisions, model partitioning strategies, and delay-cost combination decisions, we formulate the problem as a mixed integer nonlinear programming model and prove its NP-hardness. We introduce the Model Partitioning and Batch Scheduling Algorithm (MPBS), a serverless-accelerated inference mechanism that leverages dynamic Deep Neural Network (DNN) model partitioning. This architecture generates resource utilization efficient partitioning schemes according to heterogeneous hardware characteristics and request priorities, ensuring efficient device-task alignment. Additionally, a dynamic parallel scheduling mechanism, driven by a scheduling engine, enables global resource optimization by leveraging the batch processing capabilities of cloud instances, collectively enhancing local resource utilization and accelerating cloud container performance. Extensive simulation results demonstrate that compared to state-of-the-art solutions, MPBS reduces resource overhead by 20.6% while satisfying the execution time specified by Service Level Objective (SLO).</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108338"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145823157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed multi-objective consumer-centric routing for LoRa-based IoT-enabled FANET 基于lora的物联网FANET分布式多目标以消费者为中心的路由
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2026-01-06 DOI: 10.1016/j.future.2026.108368
Omer Chughtai , Muhammad Waqas Rehan , Muhammad Naeem , Ali Hamdan Alenezi , Sajjad Ali Haider
LoRa-enabled Flying Ad Hoc Networks (FANETs) offer long-range and energy-efficient connectivity for next-generation IoT and post-disaster communication infrastructures, yet their performance is fundamentally constrained by limited bandwidth, dynamic topologies, and uneven energy depletion across aerial nodes. This work develops a distributed consumer-centric multi-objective routing framework (Proposed-CCR) that jointly optimizes residual energy, link quality, and flow-level priority through a lightweight utility-driven forwarding mechanism. The design integrates composite cost modeling, two-hop neighbor awareness, adaptive path monitoring, and local repair to ensure scalable, resilient, and delay-aware multi-hop communication. Extensive simulations demonstrate that Proposed-CCR reduces per-packet energy consumption by 28–35%, extends network lifetime by over 35%, and decreases high-priority flow delay by nearly 40% relative to state-of-the-art schemes including MinHop, ACOR, GCCR, and BBCCR. These results confirm the effectiveness of a consumer-centric, LoRa-aware multi-objective heuristic for UAV-IoT integration and emergency communication scenarios, while highlighting practical opportunities for sustainable and resource-efficient airborne networking architectures.
基于lora的飞行自组织网络(fanet)为下一代物联网和灾后通信基础设施提供了远程和节能的连接,但其性能从根本上受到带宽有限、动态拓扑和空中节点间能量消耗不均匀的限制。本研究开发了一个以消费者为中心的分布式多目标路由框架(Proposed-CCR),通过轻量级实用驱动的转发机制共同优化剩余能量、链路质量和流级优先级。该设计集成了复合成本建模、两跳邻居感知、自适应路径监控和本地修复,以确保可扩展、弹性和延迟感知的多跳通信。大量的仿真表明,与MinHop、ACOR、GCCR和BBCCR等最先进的方案相比,提议的ccr将每包能耗降低了28-35%,将网络寿命延长了35%以上,并将高优先级流延迟降低了近40%。这些结果证实了以消费者为中心、lora感知的多目标启发式方法在无人机-物联网集成和应急通信场景中的有效性,同时强调了可持续和资源高效的机载网络架构的实践机会。
{"title":"Distributed multi-objective consumer-centric routing for LoRa-based IoT-enabled FANET","authors":"Omer Chughtai ,&nbsp;Muhammad Waqas Rehan ,&nbsp;Muhammad Naeem ,&nbsp;Ali Hamdan Alenezi ,&nbsp;Sajjad Ali Haider","doi":"10.1016/j.future.2026.108368","DOIUrl":"10.1016/j.future.2026.108368","url":null,"abstract":"<div><div>LoRa-enabled Flying Ad Hoc Networks (FANETs) offer long-range and energy-efficient connectivity for next-generation IoT and post-disaster communication infrastructures, yet their performance is fundamentally constrained by limited bandwidth, dynamic topologies, and uneven energy depletion across aerial nodes. This work develops a distributed consumer-centric multi-objective routing framework (Proposed-CCR) that jointly optimizes residual energy, link quality, and flow-level priority through a lightweight utility-driven forwarding mechanism. The design integrates composite cost modeling, two-hop neighbor awareness, adaptive path monitoring, and local repair to ensure scalable, resilient, and delay-aware multi-hop communication. Extensive simulations demonstrate that Proposed-CCR reduces per-packet energy consumption by 28–35%, extends network lifetime by over 35%, and decreases high-priority flow delay by nearly 40% relative to state-of-the-art schemes including MinHop, ACOR, GCCR, and BBCCR. These results confirm the effectiveness of a consumer-centric, LoRa-aware multi-objective heuristic for UAV-IoT integration and emergency communication scenarios, while highlighting practical opportunities for sustainable and resource-efficient airborne networking architectures.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108368"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145957076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HERCULES: A scalable and elastic ad-hoc file system for large-scale computing systems HERCULES:用于大规模计算系统的可伸缩和弹性的临时文件系统
IF 6.2 2区 计算机科学 Q1 COMPUTER SCIENCE, THEORY & METHODS Pub Date : 2026-06-01 Epub Date: 2025-12-28 DOI: 10.1016/j.future.2025.108350
Genaro Sánchez-Gallegos, Cosmin Petre, Javier Garcia-Blas, Jesus Carretero
The increasing demand for data processing by new, data-intensive applications is placing significant strain on the performance and capacity of HPC storage systems. Advancements in storage technologies, such as NVMe and persistent memory, have been introduced to address these demands. However, relying exclusively on ultra-fast storage devices is not cost-effective, necessitating multi-tier storage hierarchies to manage data based on its usage. In response, ad-hoc file systems have been proposed as a solution. These systems use the storage resources available in compute nodes, including memory and persistent storage, to create temporary file systems that adapt to application behavior in the HPC environment. This work presents the design, implementation, and evaluation of HERCULES, a distributed ad-hoc in-memory storage system, with a focus on its new metadata and elasticity model. HERCULES takes advantage of the Unified Communication X (UCX) framework, leveraging RDMA protocols such as Infiniband, Omnipath, shared-memory, and zero-copy transfers for data transfer. It includes elasticity features at runtime and fault-tolerant facilities. The elasticity features, together with flexible policies for data allocation, allow HERCULES to migrate data so that the available resources can be efficiently used. Our exhaustive evaluation results demonstrate a better performance than Lustre and BeeGFS, two parallel file systems heavily used in High-Performance Computing systems, and GekkoFS, an ad-hoc state-of-the-art solution.
新的数据密集型应用对数据处理的需求日益增长,给高性能计算存储系统的性能和容量带来了巨大的压力。为了满足这些需求,已经引入了NVMe和持久内存等存储技术的进步。然而,完全依赖超快存储设备并不划算,需要多层存储层次结构来根据其使用情况管理数据。为此,特设文件系统被提议作为一种解决方案。这些系统使用计算节点中可用的存储资源(包括内存和持久存储)来创建临时文件系统,以适应HPC环境中的应用程序行为。本文介绍了分布式ad-hoc内存存储系统HERCULES的设计、实现和评估,重点介绍了其新的元数据和弹性模型。HERCULES利用统一通信X (UCX)框架,利用RDMA协议(如Infiniband、Omnipath、共享内存和零拷贝传输)进行数据传输。它包括运行时的弹性特性和容错功能。弹性特性和灵活的数据分配策略使HERCULES能够迁移数据,从而有效地利用可用资源。我们详尽的评估结果表明,它的性能优于Lustre和BeeGFS(高性能计算系统中大量使用的两个并行文件系统)和GekkoFS(一种特别的最先进的解决方案)。
{"title":"HERCULES: A scalable and elastic ad-hoc file system for large-scale computing systems","authors":"Genaro Sánchez-Gallegos,&nbsp;Cosmin Petre,&nbsp;Javier Garcia-Blas,&nbsp;Jesus Carretero","doi":"10.1016/j.future.2025.108350","DOIUrl":"10.1016/j.future.2025.108350","url":null,"abstract":"<div><div>The increasing demand for data processing by new, data-intensive applications is placing significant strain on the performance and capacity of HPC storage systems. Advancements in storage technologies, such as NVMe and persistent memory, have been introduced to address these demands. However, relying exclusively on ultra-fast storage devices is not cost-effective, necessitating multi-tier storage hierarchies to manage data based on its usage. In response, <em>ad-hoc</em> file systems have been proposed as a solution. These systems use the storage resources available in compute nodes, including memory and persistent storage, to create temporary file systems that adapt to application behavior in the HPC environment. This work presents the design, implementation, and evaluation of HERCULES, a distributed <em>ad-hoc</em> in-memory storage system, with a focus on its new metadata and elasticity model. HERCULES takes advantage of the Unified Communication X (UCX) framework, leveraging RDMA protocols such as Infiniband, Omnipath, shared-memory, and zero-copy transfers for data transfer. It includes elasticity features at runtime and fault-tolerant facilities. The elasticity features, together with flexible policies for data allocation, allow HERCULES to migrate data so that the available resources can be efficiently used. Our exhaustive evaluation results demonstrate a better performance than Lustre and BeeGFS, two parallel file systems heavily used in High-Performance Computing systems, and GekkoFS, an <em>ad-hoc</em> state-of-the-art solution.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108350"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145845122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Future Generation Computer Systems-The International Journal of Escience
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1