Pub Date : 2025-12-12DOI: 10.1016/j.future.2025.108292
Oğuzhan Akyıldız
Task offloading in Connected Vehicle Networks (CVNs), a key part of the Internet of Vehicles (IoV), requires adaptive decision-making to handle computational heterogeneity and communication volatility. Although traditional fog architectures provide a foundational framework, they frequently exhibit limitations in accommodating dynamic topologies and spatiotemporal variability in resource demand. In this study, we present TMx-TORU, a Transfer Learning (TL)-assisted multi-hop task offloading protocol that operationalizes past experience to reduce selection overhead and enhance offloading precision across dynamic vehicular fog networks. TMx-TORU integrates evolutionary optimization algorithms–namely Genetic Algorithm (GA), Non-dominated Sorting Genetic Algorithm-II (NSGA-II), and our lightweight Resource-Efficient Task Offloading algorithm (RELiOff) strategy–with a TL module that learns from prior task routes and service outcomes to bypass redundant computation. Simulation results under varying CPU capacities and transmission ranges show that TL-enhanced variants consistently outperform their baselines, with up to 40.3 % gains in successfully offloaded task count and noticeable improvements in effective resource utilization. While TL-augmented GA and NSGA-II variants showed superior adaptability in throughput-efficiency balance, RELiOff maintained high offloading volume even when efficiency fluctuated, underscoring its strength in low-latency responsiveness. It seems that TMx-TORU effectively integrates mobility patterns, resource awareness, and experiential inference according to the experimental results.
车联网(CVNs)中的任务卸载是车联网(IoV)的关键部分,需要自适应决策来处理计算异构和通信波动。尽管传统的雾架构提供了一个基础框架,但它们在适应动态拓扑和资源需求的时空变化方面经常表现出局限性。在本研究中,我们提出了TMx-TORU,这是一种迁移学习(TL)辅助的多跳任务卸载协议,它利用过去的经验来减少选择开销并提高动态车辆雾网络的卸载精度。TMx-TORU集成了进化优化算法——即遗传算法(GA)、非支配排序遗传算法- ii (NSGA-II)和我们的轻量级资源高效任务卸载算法(RELiOff)策略——以及一个TL模块,该模块可以从先前的任务路由和服务结果中学习,从而绕过冗余计算。在不同CPU容量和传输范围下的模拟结果表明,tl增强的变体始终优于其基线,成功卸载的任务数最多增加40.3%,有效资源利用率显著提高。虽然tl增强的GA和NSGA-II变体在吞吐量-效率平衡方面表现出优越的适应性,但RELiOff即使在效率波动时也能保持高卸载量,强调其在低延迟响应方面的优势。从实验结果来看,TMx-TORU有效地整合了移动模式、资源感知和经验推理。
{"title":"TMx-TORU: Transfer learning enhanced location-aware multi-hop task offloading protocol for connected vehicle networks","authors":"Oğuzhan Akyıldız","doi":"10.1016/j.future.2025.108292","DOIUrl":"10.1016/j.future.2025.108292","url":null,"abstract":"<div><div>Task offloading in Connected Vehicle Networks (CVNs), a key part of the Internet of Vehicles (IoV), requires adaptive decision-making to handle computational heterogeneity and communication volatility. Although traditional fog architectures provide a foundational framework, they frequently exhibit limitations in accommodating dynamic topologies and spatiotemporal variability in resource demand. In this study, we present TMx-TORU, a Transfer Learning (TL)-assisted multi-hop task offloading protocol that operationalizes past experience to reduce selection overhead and enhance offloading precision across dynamic vehicular fog networks. TMx-TORU integrates evolutionary optimization algorithms–namely Genetic Algorithm (GA), Non-dominated Sorting Genetic Algorithm-II (NSGA-II), and our lightweight Resource-Efficient Task Offloading algorithm (RELiOff) strategy–with a TL module that learns from prior task routes and service outcomes to bypass redundant computation. Simulation results under varying CPU capacities and transmission ranges show that TL-enhanced variants consistently outperform their baselines, with up to 40.3 % gains in successfully offloaded task count and noticeable improvements in effective resource utilization. While TL-augmented GA and NSGA-II variants showed superior adaptability in throughput-efficiency balance, RELiOff maintained high offloading volume even when efficiency fluctuated, underscoring its strength in low-latency responsiveness. It seems that TMx-TORU effectively integrates mobility patterns, resource awareness, and experiential inference according to the experimental results.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108292"},"PeriodicalIF":6.2,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145732462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-10DOI: 10.1016/j.future.2025.108300
Chi Zhang, Peng Jiang, Qi Liu, Liehuang Zhu
Decentralized Internet of Things (IoT) collaborative systems necessitate robust access control mechanisms to coordinate collaborative computing and access management in decentralized environments. While functional encryption, as one of the access control technologies, demonstrates promising potential for collaborative ecosystems, existing centralized architectures suffer from single-point failure vulnerabilities and an absence of time access control - critical limitations in decentralized IoT collaboration frameworks. This paper introduces a novel access control paradigm addressing these decentralized system challenges through two key innovations: a trust authority-free decentralized key generation framework, and the implementation of time-constrained decryption policies that protect computational outputs in encrypted form prior to predetermined disclosure periods. Specifically, our technology enables individual clients to autonomously generate local cryptographic key pairs, encrypt data, and negotiate time parameters for result publication. The decryption phase aggregates encrypted data and partial decryption keys from multiple clients to ultimately enable data accessibility. We present a concrete implementation and evaluate it under both idealized and resource-constrained simulation environments, confirming the system’s practicality even with 100 clients in simulated IoT setups.
{"title":"Timed-release and partially private access control for decentralized IoT collaboration systems","authors":"Chi Zhang, Peng Jiang, Qi Liu, Liehuang Zhu","doi":"10.1016/j.future.2025.108300","DOIUrl":"10.1016/j.future.2025.108300","url":null,"abstract":"<div><div>Decentralized Internet of Things (IoT) collaborative systems necessitate robust access control mechanisms to coordinate collaborative computing and access management in decentralized environments. While functional encryption, as one of the access control technologies, demonstrates promising potential for collaborative ecosystems, existing centralized architectures suffer from single-point failure vulnerabilities and an absence of time access control - critical limitations in decentralized IoT collaboration frameworks. This paper introduces a novel access control paradigm addressing these decentralized system challenges through two key innovations: a trust authority-free decentralized key generation framework, and the implementation of time-constrained decryption policies that protect computational outputs in encrypted form prior to predetermined disclosure periods. Specifically, our technology enables individual clients to autonomously generate local cryptographic key pairs, encrypt data, and negotiate time parameters for result publication. The decryption phase aggregates encrypted data and partial decryption keys from multiple clients to ultimately enable data accessibility. We present a concrete implementation and evaluate it under both idealized and resource-constrained simulation environments, confirming the system’s practicality even with 100 clients in simulated IoT setups.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108300"},"PeriodicalIF":6.2,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145731204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-10DOI: 10.1016/j.future.2025.108302
Chaowei Wu , Wen Xiong , Sasa Duan , Yang Wang
In modern urban public transportation systems, tens of thousands of buses traverse on open road networks, serving millions of residents and generating massive GPS trajectory data. Effectively mining this data is critical for improving safety and efficiency. Co-movement pattern mining is a representative compute-intensive technique which is commonly used for bus bunching detection, but when executed on CPU-based systems, it faces scalability and latency challenges. To address this, we present an accelerated co-movement pattern mining framework based on GPU clusters. It integrates workflow management of PySpark with the high-performance computing capabilities of GPUs, and employs a pipeline to perform spatial projection, hybrid indexing, filter-verification, and memory management. We implement our approach on a Spark cluster with three nodes (equipped with six NVIDIA A40 GPUs) and evaluate it on a large-scale dataset comprising 12,788 vehicles, and over 3.22 billion GPS records collected over 31 days. The experimental results show that, compared to CPU-based approaches, our solution achieves a maximum speedup of 15.69 × . These results demonstrate that our solution can effectively support large-scale GPS trajectory analysis in bus transportation systems.
{"title":"Accelerated co-movement patterns mining: A heterogeneous framework based on GPU clusters","authors":"Chaowei Wu , Wen Xiong , Sasa Duan , Yang Wang","doi":"10.1016/j.future.2025.108302","DOIUrl":"10.1016/j.future.2025.108302","url":null,"abstract":"<div><div>In modern urban public transportation systems, tens of thousands of buses traverse on open road networks, serving millions of residents and generating massive GPS trajectory data. Effectively mining this data is critical for improving safety and efficiency. Co-movement pattern mining is a representative compute-intensive technique which is commonly used for bus bunching detection, but when executed on CPU-based systems, it faces scalability and latency challenges. To address this, we present an accelerated co-movement pattern mining framework based on GPU clusters. It integrates workflow management of PySpark with the high-performance computing capabilities of GPUs, and employs a pipeline to perform spatial projection, hybrid indexing, filter-verification, and memory management. We implement our approach on a Spark cluster with three nodes (equipped with six NVIDIA A40 GPUs) and evaluate it on a large-scale dataset comprising 12,788 vehicles, and over 3.22 billion GPS records collected over 31 days. The experimental results show that, compared to CPU-based approaches, our solution achieves a maximum speedup of 15.69 × . These results demonstrate that our solution can effectively support large-scale GPS trajectory analysis in bus transportation systems.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108302"},"PeriodicalIF":6.2,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145732465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-09DOI: 10.1016/j.future.2025.108298
Wensheng Zhang , Hao Cai , Hongli Shi , Zhenzhen Han
Traffic flow forecasting is central to intelligent transportation systems but remains challenging due to tightly coupled spatial temporal dependencies and high-order interactions. Existing deep models often assume static or single-view spatial structure, emphasize only pairwise relations, and struggle to represent dynamic spatial-temporal interactions, leading to a persistent accuracy-efficiency trade-off. To overcome this challenge, we propose a Spatial-Temporal Dual Interactive Graph Convolutional Network (STDIGCN) built around three coordinated components: (i) an adaptive traffic graph learner with macro-micro branches that infer long- and short-term topologies; (ii) a dynamic hypergraph obtained via dual transformations and embedding-based association learning to capture high-order group interactions; and (iii) a spatial-temporal dual-graph interactive convolution module that exchanges information between the graph and hypergraph streams, aligning pairwise node dependencies with high-order edge patterns while preserving multiscale temporal structure. Extensive experiments across six benchmark traffic datasets and multiple horizons demonstrate that STDIGCN outperforms strong baselines while maintaining computational efficiency.
{"title":"Spatial-temporal dual interactive graph convolutional networks for traffic flow forecasting","authors":"Wensheng Zhang , Hao Cai , Hongli Shi , Zhenzhen Han","doi":"10.1016/j.future.2025.108298","DOIUrl":"10.1016/j.future.2025.108298","url":null,"abstract":"<div><div>Traffic flow forecasting is central to intelligent transportation systems but remains challenging due to tightly coupled spatial temporal dependencies and high-order interactions. Existing deep models often assume static or single-view spatial structure, emphasize only pairwise relations, and struggle to represent dynamic spatial-temporal interactions, leading to a persistent accuracy-efficiency trade-off. To overcome this challenge, we propose a Spatial-Temporal Dual Interactive Graph Convolutional Network (STDIGCN) built around three coordinated components: (i) an adaptive traffic graph learner with macro-micro branches that infer long- and short-term topologies; (ii) a dynamic hypergraph obtained via dual transformations and embedding-based association learning to capture high-order group interactions; and (iii) a spatial-temporal dual-graph interactive convolution module that exchanges information between the graph and hypergraph streams, aligning pairwise node dependencies with high-order edge patterns while preserving multiscale temporal structure. Extensive experiments across six benchmark traffic datasets and multiple horizons demonstrate that STDIGCN outperforms strong baselines while maintaining computational efficiency.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"179 ","pages":"Article 108298"},"PeriodicalIF":6.2,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145731208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-09DOI: 10.1016/j.future.2025.108297
Yili Chen, Le Luo, Tao Jiang, Lu Fang, Sheng Xu
Graph processing on shared-memory systems fully utilizes memory bandwidth and avoids communication overhead, yet it is not as energy-efficient as expected. Since memory bandwidth becomes a major bottleneck, using more cores does not always lead to better performance. To address this limitation, we propose two predictive thread-throttling models that infer the optimal number of threads from graph characteristics such as sparsity and skewness, aiming to reduce energy consumption with minimal performance loss. The weighted model is implemented on four representative frameworks, including GreGraphMat, GrePolymer, GreGrazelle and GreLigra, and evaluated on two CPU architectures, Intel Xeon Gold 6230R and Loongson 3A6000. Experimental results show up to beyound 30 % improvement in Energy-Delay Product (EDP) on Intel and consistent 15.8 % reduction with 1.16 × speedup on Loongson. These results confirm that the proposed models achieve robust energy efficiency, strong scalability, and cross-architecture generality in shared-memory graph processing.
{"title":"Structure-aware thread throttling for energy-efficient graph processing on shared-memory systems","authors":"Yili Chen, Le Luo, Tao Jiang, Lu Fang, Sheng Xu","doi":"10.1016/j.future.2025.108297","DOIUrl":"10.1016/j.future.2025.108297","url":null,"abstract":"<div><div>Graph processing on shared-memory systems fully utilizes memory bandwidth and avoids communication overhead, yet it is not as energy-efficient as expected. Since memory bandwidth becomes a major bottleneck, using more cores does not always lead to better performance. To address this limitation, we propose two predictive thread-throttling models that infer the optimal number of threads from graph characteristics such as sparsity and skewness, aiming to reduce energy consumption with minimal performance loss. The weighted model is implemented on four representative frameworks, including GreGraphMat, GrePolymer, GreGrazelle and GreLigra, and evaluated on two CPU architectures, Intel Xeon Gold 6230R and Loongson 3A6000. Experimental results show up to beyound 30 % improvement in Energy-Delay Product (EDP) on Intel and consistent 15.8 % reduction with 1.16 × speedup on Loongson. These results confirm that the proposed models achieve robust energy efficiency, strong scalability, and cross-architecture generality in shared-memory graph processing.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108297"},"PeriodicalIF":6.2,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145731209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-09DOI: 10.1016/j.future.2025.108299
U.E. Abiha , A. Rehman , A. Abbas , M.A. Haider , F.A.M. Al-Yarimi , M.U. Gul , S.R. Hassan
With the continuous expansion and increasing complexity of the Internet of Things (IoT), anomaly detection systems have become prime targets for sophisticated adversarial attacks. These attacks often exploit weaknesses in existing detection frameworks, particularly under conditions of class imbalance and dynamic, heterogeneous data streams. To address this challenge, we propose a robust and scalable ensemble deep learning framework that integrates Conditional Generative Adversarial Networks (cGANs), Denoising Autoencoders (DAEs), and Long Short-Term Memory (LSTM) networks for anomaly detection in IoT environments. Specifically, the framework leverages cGANs to synthesize minority-class samples and alleviate data imbalance, employs DAEs for robust and noise-resilient feature extraction, and utilizes LSTM networks to capture temporal dependencies inherent in sequential IoT data. To further enhance resilience against evasion attacks, we incorporate a tailored multi-layer adversarial training strategy using both clean and dynamically generated adversarial samples along with partial gradient masking. In addition, we introduce a lightweight knowledge distillation framework, enabling a compressed student model to achieve comparable accuracy with reduced inference delay, thereby improving deployment feasibility on edge devices. Our contributions are fivefold: (i) we develop a novel ensemble architecture designed for robust and resilient anomaly detection in heterogeneous IoT systems; (ii) we introduce a customized adversarial training approach optimized for real-time constraints in IoT settings; (iii) we implement a lightweight feature selection and distillation pipeline for complexity reduction; (iv) we conduct comprehensive evaluations using the Distributed Smart Space Orchestration System (DS2OS) and Bot-IoT datasets, achieving strong performance across domains (F1: 96.26 % on DS2OS, 95.94 % on Bot-IoT).; and (v) we demonstrate that the proposed framework consistently outperforms state-of-the-art standalone and hybrid methods across a range of attack scenarios. Overall, the proposed system offers a practical and scalable defense mechanism against emerging threats in future IoT infrastructures.
{"title":"Improving adversarial resilience for anomaly detection in the heterogeneous internet of things through ensemble models","authors":"U.E. Abiha , A. Rehman , A. Abbas , M.A. Haider , F.A.M. Al-Yarimi , M.U. Gul , S.R. Hassan","doi":"10.1016/j.future.2025.108299","DOIUrl":"10.1016/j.future.2025.108299","url":null,"abstract":"<div><div>With the continuous expansion and increasing complexity of the Internet of Things (IoT), anomaly detection systems have become prime targets for sophisticated adversarial attacks. These attacks often exploit weaknesses in existing detection frameworks, particularly under conditions of class imbalance and dynamic, heterogeneous data streams. To address this challenge, we propose a robust and scalable ensemble deep learning framework that integrates Conditional Generative Adversarial Networks (cGANs), Denoising Autoencoders (DAEs), and Long Short-Term Memory (LSTM) networks for anomaly detection in IoT environments. Specifically, the framework leverages cGANs to synthesize minority-class samples and alleviate data imbalance, employs DAEs for robust and noise-resilient feature extraction, and utilizes LSTM networks to capture temporal dependencies inherent in sequential IoT data. To further enhance resilience against evasion attacks, we incorporate a tailored multi-layer adversarial training strategy using both clean and dynamically generated adversarial samples along with partial gradient masking. In addition, we introduce a lightweight knowledge distillation framework, enabling a compressed student model to achieve comparable accuracy with reduced inference delay, thereby improving deployment feasibility on edge devices. Our contributions are fivefold: (i) we develop a novel ensemble architecture designed for robust and resilient anomaly detection in heterogeneous IoT systems; (ii) we introduce a customized adversarial training approach optimized for real-time constraints in IoT settings; (iii) we implement a lightweight feature selection and distillation pipeline for complexity reduction; (iv) we conduct comprehensive evaluations using the Distributed Smart Space Orchestration System (DS2OS) and Bot-IoT datasets, achieving strong performance across domains (F1: 96.26 % on DS2OS, 95.94 % on Bot-IoT).; and (v) we demonstrate that the proposed framework consistently outperforms state-of-the-art standalone and hybrid methods across a range of attack scenarios. Overall, the proposed system offers a practical and scalable defense mechanism against emerging threats in future IoT infrastructures.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108299"},"PeriodicalIF":6.2,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145731210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-09DOI: 10.1016/j.future.2025.108295
Anwesha Mukherjee , Rajkumar Buyya
Computation offloading at lower time and lower energy consumption is crucial for resource-constrained mobile devices. This paper proposes an offloading decision-making model using federated learning. Based on the device configuration, task type, and input, the proposed decision-making model predicts whether the task is computationally intensive or not. If the predicted result is computationally intensive, then based on the network parameters the proposed decision-making model predicts whether to offload or locally execute the task. The experimental results show that the proposed method achieves above 90 % prediction accuracy in offloading decision-making, and reduces the response time and energy consumption of the user device by ∼ 11-31 %. A secure partial computation offloading method for federated learning is also proposed to deal with the Straggler effect of federated learning. The results present that the proposed partial computation offloading method for federated learning has achieved a prediction accuracy of above 98 % for the global model.
{"title":"A joint time and energy-efficient federated learning-based computation offloading method for mobile edge computing","authors":"Anwesha Mukherjee , Rajkumar Buyya","doi":"10.1016/j.future.2025.108295","DOIUrl":"10.1016/j.future.2025.108295","url":null,"abstract":"<div><div>Computation offloading at lower time and lower energy consumption is crucial for resource-constrained mobile devices. This paper proposes an offloading decision-making model using federated learning. Based on the device configuration, task type, and input, the proposed decision-making model predicts whether the task is computationally intensive or not. If the predicted result is <em>computationally intensive</em>, then based on the network parameters the proposed decision-making model predicts <em>whether to offload or locally execute</em> the task. The experimental results show that the proposed method achieves above 90 % prediction accuracy in offloading decision-making, and reduces the response time and energy consumption of the user device by ∼ 11-31 %. A secure partial computation offloading method for federated learning is also proposed to deal with the Straggler effect of federated learning. The results present that the proposed partial computation offloading method for federated learning has achieved a prediction accuracy of above 98 % for the global model.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108295"},"PeriodicalIF":6.2,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145731207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Fog computing offers a decentralized paradigm designed to process time-critical Internet of Things (IoT) tasks with minimal latency. However, due to the resource-constrained nature of fog nodes, offloading tasks to the cloud is often necessary, resulting in increased delays. To mitigate this limitation, the integration of volunteer computing with fog environments is proposed, utilizing idle computational resources from nearby devices to support latency-sensitive workloads. Furthermore, energy efficiency is a critical concern in fog computing, as it influences both operational expenditure and environmental impact. This study introduces a twofold contribution to enhance workflow scheduling. First, a volunteer selection algorithm is developed to optimally match urgent workflow tasks with suitable volunteer devices. Second, a hybrid scheduling algorithm, Sobol-FDO-SC, combines the Sobol sequence for population initialization with the Fitness Dependent Optimizer (FDO) and the Sine Cosine Algorithm (SCA). The Sobol sequence improves global search capability by avoiding premature convergence, while SCA enhances convergence speed and balances exploration-exploitation dynamics. Additionally, Dynamic Voltage and Frequency Scaling (DVFS) is applied to optimize energy consumption. Experimental evaluations demonstrate that the proposed method outperforms existing techniques in terms of makespan, energy efficiency, cost, and SLA violations.
{"title":"A volunteer-supported fog computing environment for DVFS based workflow scheduling","authors":"Anahita Dehshid, Reihaneh Khorsand, Keyvan Mohebbi","doi":"10.1016/j.future.2025.108301","DOIUrl":"10.1016/j.future.2025.108301","url":null,"abstract":"<div><div>Fog computing offers a decentralized paradigm designed to process time-critical Internet of Things (IoT) tasks with minimal latency. However, due to the resource-constrained nature of fog nodes, offloading tasks to the cloud is often necessary, resulting in increased delays. To mitigate this limitation, the integration of volunteer computing with fog environments is proposed, utilizing idle computational resources from nearby devices to support latency-sensitive workloads. Furthermore, energy efficiency is a critical concern in fog computing, as it influences both operational expenditure and environmental impact. This study introduces a twofold contribution to enhance workflow scheduling. First, a volunteer selection algorithm is developed to optimally match urgent workflow tasks with suitable volunteer devices. Second, a hybrid scheduling algorithm, Sobol-FDO-SC, combines the Sobol sequence for population initialization with the Fitness Dependent Optimizer (FDO) and the Sine Cosine Algorithm (SCA). The Sobol sequence improves global search capability by avoiding premature convergence, while SCA enhances convergence speed and balances exploration-exploitation dynamics. Additionally, Dynamic Voltage and Frequency Scaling (DVFS) is applied to optimize energy consumption. Experimental evaluations demonstrate that the proposed method outperforms existing techniques in terms of makespan, energy efficiency, cost, and SLA violations.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108301"},"PeriodicalIF":6.2,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145732466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-09DOI: 10.1016/j.future.2025.108255
Jian Xu , Bing Guo , Yan Shen , Fei Chen
With the rapid development of Internet of Things (IoT) technology, edge devices such as smartphones and sensors generate large volumes of data. Although traditional synchronous federated learning frameworks can perform distributed model training while ensuring data privacy, they often face training bottlenecks and delays in IoT environments due to device heterogeneity and computational capacity differences. These issues significantly affect training efficiency and model performance. To address these challenges, we propose a two-layer asynchronous federated learning algorithm. The algorithm uses singular value decomposition for quantization and feature extraction of edge node data, and constructs a two-layer training architecture consisting of a central server, cluster leader nodes, and regular nodes through clustering methods. We design a two-stage asynchronous training process, where model parameters are first asynchronously submitted and aggregated within the cluster, and then the aggregation of the global model is improved by distinguishing the local convergence states of the nodes, thereby reducing communication overhead and mitigating model drift. Moreover, the algorithm implements inter-cluster synchronous training by quantifying the similarity of data features across clusters, improving the model’s generalization ability and accuracy. The experimental results on the Fashion-MNIST, CIFAR-10, Sentiment140, and Blue Gene/L datasets validate the effectiveness of our method. Compared with existing approaches, our algorithm demonstrates significant improvements in prediction accuracy while considerably reducing communication requirements.
{"title":"A two-Layer asynchronous federated learning for heterogeneous IoT devices","authors":"Jian Xu , Bing Guo , Yan Shen , Fei Chen","doi":"10.1016/j.future.2025.108255","DOIUrl":"10.1016/j.future.2025.108255","url":null,"abstract":"<div><div>With the rapid development of Internet of Things (IoT) technology, edge devices such as smartphones and sensors generate large volumes of data. Although traditional synchronous federated learning frameworks can perform distributed model training while ensuring data privacy, they often face training bottlenecks and delays in IoT environments due to device heterogeneity and computational capacity differences. These issues significantly affect training efficiency and model performance. To address these challenges, we propose a two-layer asynchronous federated learning algorithm. The algorithm uses singular value decomposition for quantization and feature extraction of edge node data, and constructs a two-layer training architecture consisting of a central server, cluster leader nodes, and regular nodes through clustering methods. We design a two-stage asynchronous training process, where model parameters are first asynchronously submitted and aggregated within the cluster, and then the aggregation of the global model is improved by distinguishing the local convergence states of the nodes, thereby reducing communication overhead and mitigating model drift. Moreover, the algorithm implements inter-cluster synchronous training by quantifying the similarity of data features across clusters, improving the model’s generalization ability and accuracy. The experimental results on the Fashion-MNIST, CIFAR-10, Sentiment140, and Blue Gene/L datasets validate the effectiveness of our method. Compared with existing approaches, our algorithm demonstrates significant improvements in prediction accuracy while considerably reducing communication requirements.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108255"},"PeriodicalIF":6.2,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145732467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-07DOI: 10.1016/j.future.2025.108294
Mirko Nardi , Lorenzo Valerio , Andrea Passarella
Federated Learning (FL) enables decentralized machine learning while preserving data privacy, making it ideal for sensitive applications where data cannot be shared. While FL has been widely studied in supervised contexts, its application to unsupervised learning remains underdeveloped. This work introduces FedCRef, a novel unsupervised federated learning method designed to uncover all underlying data distributions across decentralized clients without requiring labels. This task, known as Federated Clustering, presents challenges due to heterogeneous, non-uniform data distributions and the lack of centralized coordination. Unlike previous methods that assume a one-cluster-per-client setup or require prior knowledge of the number of clusters, FedCRef generalizes to multi-cluster-per-client scenarios. Clients iteratively refine their data partitions while discovering all distinct distributions in the system. The process combines local clustering, model exchange and evaluation via reconstruction error analysis, and collaborative refinement within federated groups of similar distributions to enhance clustering accuracy. Extensive evaluations on four public datasets (EMNIST, KMNIST, Fashion-MNIST and KMNIST49) show that FedCRef successfully identifies true global data distributions, achieving an average local accuracy of up to 95 %. The method is also robust to noisy conditions, scalable, and lightweight, making it suitable for resource-constrained edge devices.
{"title":"Federated clustering: An unsupervised cluster-wise training for decentralized data distributions","authors":"Mirko Nardi , Lorenzo Valerio , Andrea Passarella","doi":"10.1016/j.future.2025.108294","DOIUrl":"10.1016/j.future.2025.108294","url":null,"abstract":"<div><div>Federated Learning (FL) enables decentralized machine learning while preserving data privacy, making it ideal for sensitive applications where data cannot be shared. While FL has been widely studied in supervised contexts, its application to unsupervised learning remains underdeveloped. This work introduces FedCRef, a novel unsupervised federated learning method designed to uncover all underlying data distributions across decentralized clients without requiring labels. This task, known as Federated Clustering, presents challenges due to heterogeneous, non-uniform data distributions and the lack of centralized coordination. Unlike previous methods that assume a one-cluster-per-client setup or require prior knowledge of the number of clusters, FedCRef generalizes to multi-cluster-per-client scenarios. Clients iteratively refine their data partitions while discovering all distinct distributions in the system. The process combines local clustering, model exchange and evaluation via reconstruction error analysis, and collaborative refinement within federated groups of similar distributions to enhance clustering accuracy. Extensive evaluations on four public datasets (EMNIST, KMNIST, Fashion-MNIST and KMNIST49) show that FedCRef successfully identifies true global data distributions, achieving an average local accuracy of up to 95 %. The method is also robust to noisy conditions, scalable, and lightweight, making it suitable for resource-constrained edge devices.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"178 ","pages":"Article 108294"},"PeriodicalIF":6.2,"publicationDate":"2025-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145704948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}