Pub Date : 2026-07-01Epub Date: 2026-01-22DOI: 10.1016/j.future.2026.108394
Bibrak Qamar Chandio, Maciej Brodowicz, Thomas Sterling
The paper provides a unified co-design of: 1) a non-Von Neumann architecture for fine-grain irregular memory computations, 2) a programming and execution model that allows spawning tasks from within the graph vertex data at runtime, 3) language constructs for actions that send work to where the data resides, combining parallel expressiveness of local control objects (LCOs) to implement asynchronous graph processing primitives, 4) and an innovative vertex-centric data-structure, using the concept of Rhizomes, that parallelizes both the out and in-degree load of vertex objects across many cores and yet provides a single programming abstraction to the vertex objects. The data structure hierarchically parallelizes the out-degree load of vertices and the in-degree load laterally. The rhizomes internally communicate and remain consistent, using event-driven synchronization mechanisms, to provide a unified and correct view of the vertex.
Simulated experimental results show performance gains for BFS, SSSP, and Page Rank on large chip sizes for the tested input graph datasets containing highly skewed degree distributions. The improvements come from the ability to express and create fine-grain dynamic computing task in the form of actions, language constructs that aid the compiler to generate code that the runtime system uses to optimally schedule tasks, and the data structure that shares both in and out-degree compute workload among memory-processing elements.
{"title":"A message-driven system for processing highly skewed graphs","authors":"Bibrak Qamar Chandio, Maciej Brodowicz, Thomas Sterling","doi":"10.1016/j.future.2026.108394","DOIUrl":"10.1016/j.future.2026.108394","url":null,"abstract":"<div><div>The paper provides a unified co-design of: 1) a non-Von Neumann architecture for fine-grain irregular memory computations, 2) a programming and execution model that allows spawning tasks from within the graph vertex data at runtime, 3) language constructs for <em>actions</em> that send work to where the data resides, combining parallel expressiveness of local control objects (LCOs) to implement asynchronous graph processing primitives, 4) and an innovative vertex-centric data-structure, using the concept of Rhizomes, that parallelizes both the out and in-degree load of vertex objects across many cores and yet provides a single programming abstraction to the vertex objects. The data structure hierarchically parallelizes the out-degree load of vertices and the in-degree load laterally. The rhizomes internally communicate and remain consistent, using event-driven synchronization mechanisms, to provide a unified and correct view of the vertex.</div><div>Simulated experimental results show performance gains for BFS, SSSP, and Page Rank on large chip sizes for the tested input graph datasets containing highly skewed degree distributions. The improvements come from the ability to express and create fine-grain dynamic computing task in the form of <em>actions</em>, language constructs that aid the compiler to generate code that the runtime system uses to optimally schedule tasks, and the data structure that shares both in and out-degree compute workload among memory-processing elements.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108394"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146033296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-07-01Epub Date: 2026-01-20DOI: 10.1016/j.future.2026.108382
Carol Lo , Thu Yein Win , Zeinab Rezaeifar , Zaheer Khan , Phil Legg
The integration of smart sensors and actuators in industrial environments has expanded the cyber-physical attack surface, making it increasingly difficult to distinguish anomalies caused by cyberattacks from those due to mechanical or electrical faults. This challenge is exacerbated by stealthy, multi-stage attacks leveraging Living off the Land (LOTL) techniques, which often evade conventional anomaly detection or intrusion detection systems (IDS).
This study presents a Digital Twin-based testbed for safe, repeatable simulation of multi-stage cyber-physical attacks targeting Cyber-Physical Systems (CPS) and Industrial Control Systems (ICS). We propose a two-level decision fusion method that aggregates and aligns anomalies across network, process, and host domains in synchronized 1-minute intervals. The first-level fusion improves OT-layer detection by applying confidence-aware decision logic to outputs combined from (a) a supervised deep learning model (LSTM-FCN) for process anomalies, (b) an unsupervised model (Isolation Forest) for OPC UA network anomalies, and (c) process alarm signals. The second-level fusion integrates these results with host-based anomalies, computed through point-based scoring of Wazuh alerts, to provide comprehensive IT/OT situational awareness. Experimental results demonstrate improved detection of stealthy, multi-stage APT attack behaviours. Additionally, Large Language Models (LLM) provide summarization of the integrated IT/OT anomaly logs into human-readable insights, enhancing interpretability and supporting cyber threat hunting.
{"title":"LOTL-hunter: Detecting multi-stage living-off-the-land attacks in cyber-physical systems using decision fusion techniques with digital twins","authors":"Carol Lo , Thu Yein Win , Zeinab Rezaeifar , Zaheer Khan , Phil Legg","doi":"10.1016/j.future.2026.108382","DOIUrl":"10.1016/j.future.2026.108382","url":null,"abstract":"<div><div>The integration of smart sensors and actuators in industrial environments has expanded the cyber-physical attack surface, making it increasingly difficult to distinguish anomalies caused by cyberattacks from those due to mechanical or electrical faults. This challenge is exacerbated by stealthy, multi-stage attacks leveraging Living off the Land (LOTL) techniques, which often evade conventional anomaly detection or intrusion detection systems (IDS).</div><div>This study presents a Digital Twin-based testbed for safe, repeatable simulation of multi-stage cyber-physical attacks targeting Cyber-Physical Systems (CPS) and Industrial Control Systems (ICS). We propose a two-level decision fusion method that aggregates and aligns anomalies across network, process, and host domains in synchronized 1-minute intervals. The first-level fusion improves OT-layer detection by applying confidence-aware decision logic to outputs combined from (a) a supervised deep learning model (LSTM-FCN) for process anomalies, (b) an unsupervised model (Isolation Forest) for OPC UA network anomalies, and (c) process alarm signals. The second-level fusion integrates these results with host-based anomalies, computed through point-based scoring of Wazuh alerts, to provide comprehensive IT/OT situational awareness. Experimental results demonstrate improved detection of stealthy, multi-stage APT attack behaviours. Additionally, Large Language Models (LLM) provide summarization of the integrated IT/OT anomaly logs into human-readable insights, enhancing interpretability and supporting cyber threat hunting.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108382"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146014879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-07-01Epub Date: 2026-01-20DOI: 10.1016/j.future.2026.108380
Alberto Marfoglia , Christian D’Errico , Sabato Mellone , Antonella Carbonaro
Background: The healthcare sector faces diverse challenges, including poor interoperability and a lack of personalized approaches, which limit patient outcomes. Ineffective data exchange and one-size-fits-all treatments fail to meet individual needs. Emerging technologies like digital twins (DTs), the semantic web, and AI show promise in tackling these obstacles. For this reason, we introduced CONNECTED, a conceptual multi-level framework that combines these techniques to deploy general-purpose patient DTs. Objective: This study assesses CONNECTED’s comprehensiveness, applicability, and utility for developing intelligent, personalized healthcare applications. Specifically, we deliver a preliminary version of the framework to predict future patient states and demonstrate its automation benefits in deploying semantically enriched, AI-powered patient DTs. Methods: We enhanced the CONNECTED architecture by providing a formal definition of DT and modularizing its core functionalities into four microservices (Properties, State, Capabilities, and Manifest). The Manifest service facilitates AI model integration through the Model Interface Manifest Ontology (MIMO), enabling automatic data-to-model binding via a reasoner. Using the HeartBeatKG quality assessment tool, we validated MIMO and tested the internal logic by integrating a well-established stroke-risk model. Results: Our implementation comprehends: (1) deploying a FHIR-compliant, patient-centric API for clinical history access, real-time monitoring, and predictive simulation; (2) publishing MIMO; (3) establishing the Manifest protocol for seamless, general-purpose AI model integration tailored to individual patient profiles; and (4) a proof-of-concept benchmarking application comparing multiple stroke risk classifiers. Conclusion: CONNECTED establishes a flexible, scalable foundation for interoperable semantic patient DTs. Automation reduces technical overhead and enables users to focus on delivering personalized, insight-driven care.
{"title":"A knowledge graph-driven framework for deploying AI-powered patient digital twins","authors":"Alberto Marfoglia , Christian D’Errico , Sabato Mellone , Antonella Carbonaro","doi":"10.1016/j.future.2026.108380","DOIUrl":"10.1016/j.future.2026.108380","url":null,"abstract":"<div><div>Background: The healthcare sector faces diverse challenges, including poor interoperability and a lack of personalized approaches, which limit patient outcomes. Ineffective data exchange and one-size-fits-all treatments fail to meet individual needs. Emerging technologies like digital twins (DTs), the semantic web, and AI show promise in tackling these obstacles. For this reason, we introduced CONNECTED, a conceptual multi-level framework that combines these techniques to deploy general-purpose patient DTs. Objective: This study assesses CONNECTED’s comprehensiveness, applicability, and utility for developing intelligent, personalized healthcare applications. Specifically, we deliver a preliminary version of the framework to predict future patient states and demonstrate its automation benefits in deploying semantically enriched, AI-powered patient DTs. Methods: We enhanced the CONNECTED architecture by providing a formal definition of DT and modularizing its core functionalities into four microservices (Properties, State, Capabilities, and Manifest). The Manifest service facilitates AI model integration through the Model Interface Manifest Ontology (MIMO), enabling automatic data-to-model binding via a reasoner. Using the HeartBeatKG quality assessment tool, we validated MIMO and tested the internal logic by integrating a well-established stroke-risk model. Results: Our implementation comprehends: (1) deploying a FHIR-compliant, patient-centric API for clinical history access, real-time monitoring, and predictive simulation; (2) publishing MIMO; (3) establishing the Manifest protocol for seamless, general-purpose AI model integration tailored to individual patient profiles; and (4) a proof-of-concept benchmarking application comparing multiple stroke risk classifiers. Conclusion: CONNECTED establishes a flexible, scalable foundation for interoperable semantic patient DTs. Automation reduces technical overhead and enables users to focus on delivering personalized, insight-driven care.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108380"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146014895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-07-01Epub Date: 2026-02-03DOI: 10.1016/j.future.2026.108409
Seyed Hossein Ahmadpanah
Cold start latency poses a fundamental challenge to the promise of real-time, serverless computing for the Internet of Things (IoT). while there are container warming techniques, they are frequently created for centralized, trusted clouds and do not take into account the size, resource limitations, and hostile environment of decentralized edge environments. Moreover, they usually base their performance claims on oversimplified models that don’t accurately represent operational dynamics in the real world. In this paper, we present Concord, a scalability-focused framework for resilient container warming that has undergone unprecedented validation. Concord formulates the warming problem as a constrained stochastic game, allowing clusters of IoT devices to collaboratively manage shared container instances. A lightweight Byzantine Fault Tolerance (BFT) protocol ensures robustness against malicious actors and secures this collaboration. The main contribution of the framework is its observable performance in real-world scenarios: we demonstrate that Concord reduces the cold start probability to less than 1.2% across 10,000 devices with low energy and traffic overhead using the Azure Functions 2023 public trace. The entire system, evaluation environment, and analysis scripts are made available as a one-click, publicly accessible, and reproducible artifact to support scientific transparency.
{"title":"Concord: A scalable, trace-driven, and reproducible framework for resilient container warming in serverless IoT","authors":"Seyed Hossein Ahmadpanah","doi":"10.1016/j.future.2026.108409","DOIUrl":"10.1016/j.future.2026.108409","url":null,"abstract":"<div><div>Cold start latency poses a fundamental challenge to the promise of real-time, serverless computing for the Internet of Things (IoT). while there are container warming techniques, they are frequently created for centralized, trusted clouds and do not take into account the size, resource limitations, and hostile environment of decentralized edge environments. Moreover, they usually base their performance claims on oversimplified models that don’t accurately represent operational dynamics in the real world. In this paper, we present Concord, a scalability-focused framework for resilient container warming that has undergone unprecedented validation. Concord formulates the warming problem as a constrained stochastic game, allowing clusters of IoT devices to collaboratively manage shared container instances. A lightweight Byzantine Fault Tolerance (BFT) protocol ensures robustness against malicious actors and secures this collaboration. The main contribution of the framework is its observable performance in real-world scenarios: we demonstrate that Concord reduces the cold start probability to less than 1.2% across 10,000 devices with low energy and traffic overhead using the Azure Functions 2023 public trace. The entire system, evaluation environment, and analysis scripts are made available as a one-click, publicly accessible, and reproducible artifact to support scientific transparency.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108409"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146110186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-07-01Epub Date: 2026-02-03DOI: 10.1016/j.future.2026.108412
Rabi Shaw , Suman Majumder
The increasing complexity of cybercrime and identity manipulation activities demands a secure, verifiable, and quantum-resilient framework for criminal identification. This study introduces ‘DQVeriChain’, a distributed quantum-state-verified and decentralized identity-based large language model (LLM) designed for criminal tracking using blockchain. The system integrates quantum principal component analysis (QPCA), GHZ state-fidelity validation, and ZZ-feature mapping to ensure high-fidelity quantum entanglement and secure feature encoding. The verification process combines both the Quantum Self-Attention Mechanism (QSAM) and the Quantum Self-Correcting Attention Mechanism (QSCAM) with LLM-driven inference for intelligent anomaly detection. In addition, validation was performed using IBM Qiskit with 10-fold cross-validation and fidelity thresholds (F ≥ 0.5), achieving prediction accuracy greater than 99% with minimal quantum circuit complexity (5/5 qubits). These experimental results confirm that ‘DQVeriChain’ offers an efficient, tamper-resistant, and scalable architecture suitable for real-time forensic and cybersecurity applications.
{"title":"DQVeriChain: Distributed quantum-state-verified and DID-based self-attentive large language model for criminal tracking using blockchain","authors":"Rabi Shaw , Suman Majumder","doi":"10.1016/j.future.2026.108412","DOIUrl":"10.1016/j.future.2026.108412","url":null,"abstract":"<div><div>The increasing complexity of cybercrime and identity manipulation activities demands a secure, verifiable, and quantum-resilient framework for criminal identification. This study introduces ‘DQVeriChain’, a distributed quantum-state-verified and decentralized identity-based large language model (LLM) designed for criminal tracking using blockchain. The system integrates quantum principal component analysis (QPCA), GHZ state-fidelity validation, and ZZ-feature mapping to ensure high-fidelity quantum entanglement and secure feature encoding. The verification process combines both the Quantum Self-Attention Mechanism (QSAM) and the Quantum Self-Correcting Attention Mechanism (QSCAM) with LLM-driven inference for intelligent anomaly detection. In addition, validation was performed using IBM Qiskit with 10-fold cross-validation and fidelity thresholds (<em>F</em> ≥ 0.5), achieving prediction accuracy greater than 99% with minimal quantum circuit complexity (5/5 qubits). These experimental results confirm that ‘DQVeriChain’ offers an efficient, tamper-resistant, and scalable architecture suitable for real-time forensic and cybersecurity applications.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"180 ","pages":"Article 108412"},"PeriodicalIF":6.2,"publicationDate":"2026-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146110192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-06-01Epub Date: 2025-12-09DOI: 10.1016/j.future.2025.108298
Wensheng Zhang , Hao Cai , Hongli Shi , Zhenzhen Han
Traffic flow forecasting is central to intelligent transportation systems but remains challenging due to tightly coupled spatial temporal dependencies and high-order interactions. Existing deep models often assume static or single-view spatial structure, emphasize only pairwise relations, and struggle to represent dynamic spatial-temporal interactions, leading to a persistent accuracy-efficiency trade-off. To overcome this challenge, we propose a Spatial-Temporal Dual Interactive Graph Convolutional Network (STDIGCN) built around three coordinated components: (i) an adaptive traffic graph learner with macro-micro branches that infer long- and short-term topologies; (ii) a dynamic hypergraph obtained via dual transformations and embedding-based association learning to capture high-order group interactions; and (iii) a spatial-temporal dual-graph interactive convolution module that exchanges information between the graph and hypergraph streams, aligning pairwise node dependencies with high-order edge patterns while preserving multiscale temporal structure. Extensive experiments across six benchmark traffic datasets and multiple horizons demonstrate that STDIGCN outperforms strong baselines while maintaining computational efficiency.
{"title":"Spatial-temporal dual interactive graph convolutional networks for traffic flow forecasting","authors":"Wensheng Zhang , Hao Cai , Hongli Shi , Zhenzhen Han","doi":"10.1016/j.future.2025.108298","DOIUrl":"10.1016/j.future.2025.108298","url":null,"abstract":"<div><div>Traffic flow forecasting is central to intelligent transportation systems but remains challenging due to tightly coupled spatial temporal dependencies and high-order interactions. Existing deep models often assume static or single-view spatial structure, emphasize only pairwise relations, and struggle to represent dynamic spatial-temporal interactions, leading to a persistent accuracy-efficiency trade-off. To overcome this challenge, we propose a Spatial-Temporal Dual Interactive Graph Convolutional Network (STDIGCN) built around three coordinated components: (i) an adaptive traffic graph learner with macro-micro branches that infer long- and short-term topologies; (ii) a dynamic hypergraph obtained via dual transformations and embedding-based association learning to capture high-order group interactions; and (iii) a spatial-temporal dual-graph interactive convolution module that exchanges information between the graph and hypergraph streams, aligning pairwise node dependencies with high-order edge patterns while preserving multiscale temporal structure. Extensive experiments across six benchmark traffic datasets and multiple horizons demonstrate that STDIGCN outperforms strong baselines while maintaining computational efficiency.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108298"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145731208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-06-01Epub Date: 2025-12-28DOI: 10.1016/j.future.2025.108274
Marcelo Santos , Rubens Matos , Marco Vieira , Jean Araujo
Software Aging and Rejuvenation (SAR) has been extensively studied due to its critical role in ensuring the reliable operation of systems. Although container orchestration is essential for efficiently managing and scaling cloud resources, the impact of SAR is not yet fully understood. This paper presents experiments conducted on two versions of Ubuntu Linux, simulating the operational scenarios of a private cloud. Each cluster includes one Main node and three Worker nodes, utilizing Containerd as the container runtime and Kubernetes as the orchestrator, across four distinct scenarios. The primary experimental conditions were maintained across all scenarios, including configurations, workloads, and test duration. Throughout each experiment, metrics such as CPU utilization, memory usage and disk utilization were monitored, considering system-wide values and observations for the Containerd and Kubelet services. The experiments also included measuring the response time of a web server for external HTTP requests submitted to the clusters. The initial scenario focused on investigating the effects of software aging, while subsequent scenarios explored the adoption of different rejuvenation strategies. Effects of software aging were observed across all scenarios, with resource leaks identified, particularly in memory usage, even when the cluster was under no load. The issues observed can lead to performance degradation and compromise reliability and availability if the system crashes due to memory exhaustion.
{"title":"Software aging issues and rejuvenation strategies for a container orchestration system","authors":"Marcelo Santos , Rubens Matos , Marco Vieira , Jean Araujo","doi":"10.1016/j.future.2025.108274","DOIUrl":"10.1016/j.future.2025.108274","url":null,"abstract":"<div><div>Software Aging and Rejuvenation (SAR) has been extensively studied due to its critical role in ensuring the reliable operation of systems. Although container orchestration is essential for efficiently managing and scaling cloud resources, the impact of SAR is not yet fully understood. This paper presents experiments conducted on two versions of Ubuntu Linux, simulating the operational scenarios of a private cloud. Each cluster includes one Main node and three Worker nodes, utilizing Containerd as the container runtime and Kubernetes as the orchestrator, across four distinct scenarios. The primary experimental conditions were maintained across all scenarios, including configurations, workloads, and test duration. Throughout each experiment, metrics such as CPU utilization, memory usage and disk utilization were monitored, considering system-wide values and observations for the Containerd and Kubelet services. The experiments also included measuring the response time of a web server for external HTTP requests submitted to the clusters. The initial scenario focused on investigating the effects of software aging, while subsequent scenarios explored the adoption of different rejuvenation strategies. Effects of software aging were observed across all scenarios, with resource leaks identified, particularly in memory usage, even when the cluster was under no load. The issues observed can lead to performance degradation and compromise reliability and availability if the system crashes due to memory exhaustion.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108274"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145845123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-06-01Epub Date: 2025-12-25DOI: 10.1016/j.future.2025.108341
Jakub Homola, Ondřej Meca, Lubomír Říha, Tomáš Brzobohatý
FETI methods, which build on the Finite Element Method, are utilized for large-scale engineering simulations. They use domain decomposition techniques to divide a large domain into many smaller subdomains, which can be processed in parallel.
Current trends in HPC focus on GPU-accelerated clusters. To utilize them efficiently, FETI solvers should be able to use these accelerators. Recent developments have demonstrated that the fundamental component of the FETI methods, the dual operator, can be successfully offloaded to the GPU.
In this paper, we focus on GPU acceleration of the Hybrid FETI variant. It reduces the size of the projector by using a two-level decomposition, thus allowing for a significantly higher number of compute nodes to be efficiently utilized. In turn, it allows us to split the problem into a larger number of smaller subdomains, which improves single-process performance.
We demonstrate the performance on a real-world problem of transient nonlinear dynamics that requires reassembling of the dual operator, preconditioner, and projector during each call of the solver. On the MareNostrum 5 supercomputer, using Nvidia H100 GPUs, we achieved a speedup of 2.9 for the whole Hybrid FETI solver compared to a CPU-only run.
{"title":"GPU acceleration of hybrid FETI solver for problems of transient nonlinear dynamics","authors":"Jakub Homola, Ondřej Meca, Lubomír Říha, Tomáš Brzobohatý","doi":"10.1016/j.future.2025.108341","DOIUrl":"10.1016/j.future.2025.108341","url":null,"abstract":"<div><div>FETI methods, which build on the Finite Element Method, are utilized for large-scale engineering simulations. They use domain decomposition techniques to divide a large domain into many smaller subdomains, which can be processed in parallel.</div><div>Current trends in HPC focus on GPU-accelerated clusters. To utilize them efficiently, FETI solvers should be able to use these accelerators. Recent developments have demonstrated that the fundamental component of the FETI methods, the dual operator, can be successfully offloaded to the GPU.</div><div>In this paper, we focus on GPU acceleration of the Hybrid FETI variant. It reduces the size of the projector by using a two-level decomposition, thus allowing for a significantly higher number of compute nodes to be efficiently utilized. In turn, it allows us to split the problem into a larger number of smaller subdomains, which improves single-process performance.</div><div>We demonstrate the performance on a real-world problem of transient nonlinear dynamics that requires reassembling of the dual operator, preconditioner, and projector during each call of the solver. On the MareNostrum 5 supercomputer, using Nvidia H100 GPUs, we achieved a speedup of 2.9 for the whole Hybrid FETI solver compared to a CPU-only run.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108341"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145845128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-06-01Epub Date: 2025-12-15DOI: 10.1016/j.future.2025.108313
Gibran Gomez , Kevin van Liebergen , Davide Sanvito , Giuseppe Siracusano , Roberto Gonzalez , Juan Caballero
Cryptocurrency abuse reporting services are a valuable data source about abusive blockchain addresses, prevalent types of cryptocurrency abuse, and their financial impact on victims. However, they may suffer data pollution due to their crowd-sourced nature. This work analyzes the extent and impact of data pollution in cryptocurrency abuse reporting services and proposes a novel LLM-based defense to address the pollution. We collect 289K abuse reports submitted over 6 years to two popular services and use them to answer three research questions. RQ1 analyzes the extent and impact of pollution. We show that spam reports will eventually flood unchecked abuse reporting services, with BitcoinAbuse receiving 75 % of spam before stopping operations. We build a public dataset of 19,443 abuse reports labeled with 19 popular abuse types and use it to reveal the inaccuracy of user-reported abuse types. We identified 91 (0.1 %) benign addresses reported, responsible for 60 % of all the received funds. RQ2 examines whether we can automate identifying valid reports and their classification into abuse types. We propose an unsupervised LLM-based classifier that achieves an F1 score of 0.95 when classifying reports, an F1 of 0.89 when classifying out-of-distribution data, and an F1 of 0.99 when identifying spam reports. Our unsupervised LLM-based classifier clearly outperforms two baselines: a supervised classifier and a naive usage of the LLM. Finally, RQ3 demonstrates the usefulness of our LLM-based classifier for quantifying the financial impact of different cryptocurrency abuse types. We show that victim-reported losses heavily underestimate cybercriminal revenue by estimating a 29 times higher revenue from deposit transactions. We identified that investment scams have the highest financial impact and that extortions have lower conversion rates but compensate for them with massive email campaigns.
{"title":"Clean up the mess: Addressing data pollution in cryptocurrency abuse reporting services","authors":"Gibran Gomez , Kevin van Liebergen , Davide Sanvito , Giuseppe Siracusano , Roberto Gonzalez , Juan Caballero","doi":"10.1016/j.future.2025.108313","DOIUrl":"10.1016/j.future.2025.108313","url":null,"abstract":"<div><div>Cryptocurrency abuse reporting services are a valuable data source about abusive blockchain addresses, prevalent types of cryptocurrency abuse, and their financial impact on victims. However, they may suffer data pollution due to their crowd-sourced nature. This work analyzes the extent and impact of data pollution in cryptocurrency abuse reporting services and proposes a novel LLM-based defense to address the pollution. We collect 289K abuse reports submitted over 6 years to two popular services and use them to answer three research questions. RQ1 analyzes the extent and impact of pollution. We show that spam reports will eventually flood unchecked abuse reporting services, with BitcoinAbuse receiving 75 % of spam before stopping operations. We build a public dataset of 19,443 abuse reports labeled with 19 popular abuse types and use it to reveal the inaccuracy of user-reported abuse types. We identified 91 (0.1 %) benign addresses reported, responsible for 60 % of all the received funds. RQ2 examines whether we can automate identifying valid reports and their classification into abuse types. We propose an unsupervised LLM-based classifier that achieves an F1 score of 0.95 when classifying reports, an F1 of 0.89 when classifying out-of-distribution data, and an F1 of 0.99 when identifying spam reports. Our unsupervised LLM-based classifier clearly outperforms two baselines: a supervised classifier and a naive usage of the LLM. Finally, RQ3 demonstrates the usefulness of our LLM-based classifier for quantifying the financial impact of different cryptocurrency abuse types. We show that victim-reported losses heavily underestimate cybercriminal revenue by estimating a 29 times higher revenue from deposit transactions. We identified that investment scams have the highest financial impact and that extortions have lower conversion rates but compensate for them with massive email campaigns.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108313"},"PeriodicalIF":6.2,"publicationDate":"2026-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145785052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}