In the ever-evolving landscape of smart city transportation, effective traffic management remains a critical challenge. To address this, we propose a novel Smart Traffic Management System (STMS) Architecture algorithm that combines cutting-edge technologies, including Blockchain, IoT, edge computing, and reinforcement learning. STMS aims to optimize traffic flow, minimize congestion, and enhance transportation efficiency while ensuring data integrity, security, and decentralized decision-making. STMS integrates the Twin Delayed Deep Deterministic Policy Gradient (TD3) reinforcement learning algorithm with Blockchain technology to enable secure and transparent data sharing among traffic-related entities. Smart contracts are deployed on the Blockchain to automate the execution of predefined traffic rules, ensuring compliance and accountability. Integrating IoT sensors on vehicles, roadways, and traffic signals provides real-time traffic data, while edge nodes perform local traffic analysis and contribute to optimization. The algorithm’s decentralized decision-making empowers edge devices, traffic signals, and vehicles to interact autonomously, making informed decisions based on local data and predefined rules stored on the Blockchain. TD3 optimizes traffic signal timings, route suggestions, and traffic flow control, ensuring smooth transportation operations. STMSs holistic approach addresses traffic management challenges in smart cities by combining advanced technologies. By leveraging Blockchain’s immutability, IoT’s real-time insights, edge computing’s local intelligence, and TD3’s reinforcement learning capabilities, STMS presents a robust solution for achieving efficient and secure transportation systems. This research underscores the potential for innovative algorithms to revolutionize urban mobility, ushering in a new era of smart and sustainable transportation networks.
{"title":"Exploring the Synergy of Blockchain, IoT, and Edge Computing in Smart Traffic Management across Urban Landscapes","authors":"Yu Chen, Yilun Qiu, Zhenyu Tang, Shuling Long, Lingfeng Zhao, Zhong Tang","doi":"10.1007/s10723-024-09762-6","DOIUrl":"https://doi.org/10.1007/s10723-024-09762-6","url":null,"abstract":"<p>In the ever-evolving landscape of smart city transportation, effective traffic management remains a critical challenge. To address this, we propose a novel Smart Traffic Management System (STMS) Architecture algorithm that combines cutting-edge technologies, including Blockchain, IoT, edge computing, and reinforcement learning. STMS aims to optimize traffic flow, minimize congestion, and enhance transportation efficiency while ensuring data integrity, security, and decentralized decision-making. STMS integrates the Twin Delayed Deep Deterministic Policy Gradient (TD3) reinforcement learning algorithm with Blockchain technology to enable secure and transparent data sharing among traffic-related entities. Smart contracts are deployed on the Blockchain to automate the execution of predefined traffic rules, ensuring compliance and accountability. Integrating IoT sensors on vehicles, roadways, and traffic signals provides real-time traffic data, while edge nodes perform local traffic analysis and contribute to optimization. The algorithm’s decentralized decision-making empowers edge devices, traffic signals, and vehicles to interact autonomously, making informed decisions based on local data and predefined rules stored on the Blockchain. TD3 optimizes traffic signal timings, route suggestions, and traffic flow control, ensuring smooth transportation operations. STMSs holistic approach addresses traffic management challenges in smart cities by combining advanced technologies. By leveraging Blockchain’s immutability, IoT’s real-time insights, edge computing’s local intelligence, and TD3’s reinforcement learning capabilities, STMS presents a robust solution for achieving efficient and secure transportation systems. This research underscores the potential for innovative algorithms to revolutionize urban mobility, ushering in a new era of smart and sustainable transportation networks.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-04-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140615758","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-16DOI: 10.1007/s10723-024-09760-8
Neha Kaushik, Harish Kumar, Vinay Raj
Microservices has become a buzzword in industry as many large IT giants such as Amazon, Twitter, Uber, etc have started migrating their existing applications to this new style and few of them have started building their new applications with this style. Due to increasing user requirements and the need to add more business functionalities to the existing applications, the web applications designed using the microservices style also face a few performance challenges. Though this style has been successfully adopted in the design of large enterprise applications, still the applications face performance related issues. It is clear from the literature that most of the articles focus only on the backend microservices. To the best of our knowledge, there has been no solution proposed considering micro frontends along with the backend microservices. To improve the performance of the microservices based web applications, in this paper, a new framework for the design of web applications with micro frontends for frontend and microservices in the backend of the application is presented. To assess the proposed framework, an empirical investigation is performed to analyze the performance and it is found that the applications designed with micro frontends with microservices have performed better than the applications with monolithic frontends. Additionally, to predict the performance of microservices based applications, a machine learning model is proposed as machine learning has wide applications in software engineering related activities. The accuracy of the proposed model using different metrics is also presented.
随着亚马逊、Twitter、Uber 等许多大型 IT 巨头开始将其现有应用程序迁移到这种新风格,微服务已成为业界的热门词汇,其中少数公司已开始使用这种风格构建新的应用程序。由于用户需求不断增加,而且需要在现有应用程序中添加更多业务功能,使用微服务样式设计的网络应用程序也面临着一些性能挑战。虽然这种风格已成功应用于大型企业应用程序的设计中,但这些应用程序仍然面临着与性能相关的问题。从文献中可以明显看出,大多数文章只关注后端微服务。据我们所知,还没有人提出过将微前端与后端微服务一起考虑的解决方案。为了提高基于微服务的网络应用程序的性能,本文提出了一种新的网络应用程序设计框架,前端采用微前端,后端采用微服务。为了评估所提出的框架,我们进行了一项实证调查来分析其性能,结果发现,使用微前端和微服务设计的应用程序比使用单体前端的应用程序性能更好。此外,为了预测基于微服务的应用程序的性能,还提出了一个机器学习模型,因为机器学习在软件工程相关活动中有着广泛的应用。此外,还介绍了所提模型使用不同指标的准确性。
{"title":"Micro Frontend Based Performance Improvement and Prediction for Microservices Using Machine Learning","authors":"Neha Kaushik, Harish Kumar, Vinay Raj","doi":"10.1007/s10723-024-09760-8","DOIUrl":"https://doi.org/10.1007/s10723-024-09760-8","url":null,"abstract":"<p>Microservices has become a buzzword in industry as many large IT giants such as Amazon, Twitter, Uber, etc have started migrating their existing applications to this new style and few of them have started building their new applications with this style. Due to increasing user requirements and the need to add more business functionalities to the existing applications, the web applications designed using the microservices style also face a few performance challenges. Though this style has been successfully adopted in the design of large enterprise applications, still the applications face performance related issues. It is clear from the literature that most of the articles focus only on the backend microservices. To the best of our knowledge, there has been no solution proposed considering micro frontends along with the backend microservices. To improve the performance of the microservices based web applications, in this paper, a new framework for the design of web applications with micro frontends for frontend and microservices in the backend of the application is presented. To assess the proposed framework, an empirical investigation is performed to analyze the performance and it is found that the applications designed with micro frontends with microservices have performed better than the applications with monolithic frontends. Additionally, to predict the performance of microservices based applications, a machine learning model is proposed as machine learning has wide applications in software engineering related activities. The accuracy of the proposed model using different metrics is also presented.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-04-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140583383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-02DOI: 10.1007/s10723-024-09757-3
Tao Hai, Muammer Aksoy, Celestine Iwendi, Ebuka Ibeke, Senthilkumar Mohan
The lack of data security and the hazardous nature of the Internet of Vehicles (IoV), in the absence of networking settings, have prevented the openness and self-organization of the vehicle networks of IoV cars. The lapses originating in the areas of Confidentiality, Integrity, and Authenticity (CIA) have also increased the possibility of malicious attacks. To overcome these challenges, this paper proposes an updated Games-based CIA security mechanism to secure IoVs using Blockchain and Artificial Intelligence (AI) technology. The proposed framework consists of a trustworthy authorization solution three layers, including the authentication of vehicles using Physical Unclonable Functions (PUFs), a flexible Proof-of-Work (dPOW) consensus framework, and AI-enhanced duel gaming. The credibility of the framework is validated by different security analyses, showcasing its superiority over existing systems in terms of security, functionality, computation, and transaction overhead. Additionally, the proposed solution effectively handles challenges like side channel and physical cloning attacks, which many existing frameworks fail to address. The implementation of this mechanism involves the use of a reduced encumbered blockchain, coupled with AI-based authentication through duel gaming, showcasing its efficiency and physical-level support, a feature not present in most existing blockchain-based IoV verification frameworks.
由于缺乏联网设置,车联网(IoV)的数据安全性和危险性不足,阻碍了车联网汽车网络的开放性和自组织性。源于保密性、完整性和真实性(CIA)领域的漏洞也增加了恶意攻击的可能性。为了克服这些挑战,本文提出了一种更新的基于游戏的 CIA 安全机制,利用区块链和人工智能(AI)技术确保物联网汽车的安全。所提出的框架由三层可信授权解决方案组成,包括使用物理不可克隆函数(PUF)对车辆进行身份验证、灵活的工作证明(dPOW)共识框架和人工智能增强型对决游戏。不同的安全分析验证了该框架的可信度,表明其在安全性、功能性、计算量和交易开销方面优于现有系统。此外,所提出的解决方案还能有效处理侧信道和物理克隆攻击等挑战,而许多现有框架都无法解决这些问题。该机制的实施涉及使用减少了加密的区块链,并通过决斗游戏与基于人工智能的身份验证相结合,从而展示了其效率和物理层支持,这是大多数现有基于区块链的物联网验证框架所不具备的功能。
{"title":"CIA Security for Internet of Vehicles and Blockchain-AI Integration","authors":"Tao Hai, Muammer Aksoy, Celestine Iwendi, Ebuka Ibeke, Senthilkumar Mohan","doi":"10.1007/s10723-024-09757-3","DOIUrl":"https://doi.org/10.1007/s10723-024-09757-3","url":null,"abstract":"<p>The lack of data security and the hazardous nature of the Internet of Vehicles (IoV), in the absence of networking settings, have prevented the openness and self-organization of the vehicle networks of IoV cars. The lapses originating in the areas of Confidentiality, Integrity, and Authenticity (CIA) have also increased the possibility of malicious attacks. To overcome these challenges, this paper proposes an updated Games-based CIA security mechanism to secure IoVs using Blockchain and Artificial Intelligence (AI) technology. The proposed framework consists of a trustworthy authorization solution three layers, including the authentication of vehicles using Physical Unclonable Functions (PUFs), a flexible Proof-of-Work (dPOW) consensus framework, and AI-enhanced duel gaming. The credibility of the framework is validated by different security analyses, showcasing its superiority over existing systems in terms of security, functionality, computation, and transaction overhead. Additionally, the proposed solution effectively handles challenges like side channel and physical cloning attacks, which many existing frameworks fail to address. The implementation of this mechanism involves the use of a reduced encumbered blockchain, coupled with AI-based authentication through duel gaming, showcasing its efficiency and physical-level support, a feature not present in most existing blockchain-based IoV verification frameworks.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140583649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-26DOI: 10.1007/s10723-024-09759-1
Abstract
In recent years, internet enterprises have transitioned from traditional monolithic service to microservice architecture to better meet evolving business requirements. However, it also brings great challenges to the resource management of service providers. Existing research has not fully considered the request characteristics of internet application scenarios. Some studies apply traditional task scheduling models and strategies to microservice scheduling scenarios, while others optimize microservice deployment and request routing separately. In this paper, we propose a microservice instance deployment algorithm based on genetic and local search, and a request routing algorithm based on probabilistic forwarding. The service graph with complex dependencies is decomposed into multiple service chains, and the open Jackson queuing network is applied to analyze the performance of the microservice system. Data evaluation results demonstrate that our scheme significantly outperforms the benchmark strategy. Our algorithm has reduced the average response latency by 37%-67% and enhanced request success rate by 8%-115% compared to other baseline algorithms.
{"title":"On the Joint Design of Microservice Deployment and Routing in Cloud Data Centers","authors":"","doi":"10.1007/s10723-024-09759-1","DOIUrl":"https://doi.org/10.1007/s10723-024-09759-1","url":null,"abstract":"<h3>Abstract</h3> <p>In recent years, internet enterprises have transitioned from traditional monolithic service to microservice architecture to better meet evolving business requirements. However, it also brings great challenges to the resource management of service providers. Existing research has not fully considered the request characteristics of internet application scenarios. Some studies apply traditional task scheduling models and strategies to microservice scheduling scenarios, while others optimize microservice deployment and request routing separately. In this paper, we propose a microservice instance deployment algorithm based on genetic and local search, and a request routing algorithm based on probabilistic forwarding. The service graph with complex dependencies is decomposed into multiple service chains, and the open Jackson queuing network is applied to analyze the performance of the microservice system. Data evaluation results demonstrate that our scheme significantly outperforms the benchmark strategy. Our algorithm has reduced the average response latency by 37%-67% and enhanced request success rate by 8%-115% compared to other baseline algorithms.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140302885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-14DOI: 10.1007/s10723-024-09755-5
Shujie Qiu
Educational institutions today are embracing technology to enhance education quality through intelligent systems. This study introduces an innovative strategy to boost the performance of such procedures by seamlessly integrating machine learning on edge devices and cloud infrastructure. The proposed framework harnesses the capabilities of a Hybrid 1D Convolutional Neural Network (CNN) and Long Short-Term Memory Network (LSTM) architecture, offering profound insights into intelligent education. Operating at the crossroads of localised and centralised analyses, the Hybrid 1D CNN-LSTM architecture signifies a significant advancement. It directly engages edge devices used by students and educators, laying the groundwork for personalised learning experiences. This architecture adeptly captures the intricacies of various modalities, including text, images, and videos, by harmonising 1D CNN layers and LSTM modules. This approach facilitates the extraction of tailored features from each modality and the exploration of temporal intricacies. Consequently, the architecture provides a holistic comprehension of student engagement and comprehension dynamics, unveiling individual learning preferences. Moreover, the framework seamlessly integrates data from edge devices into the cloud infrastructure, allowing insights from both domains to merge. Educators benefit from attention-enhanced feature maps that encapsulate personalised insights, empowering them to customise content and strategies according to student learning preferences. The approach bridges real-time, localised analysis with comprehensive cloud-mediated insights, paving the path for transformative educational experiences. Empirical validation reinforces the effectiveness of the Hybrid 1D CNN-LSTM architecture, cementing its potential to revolutionise intelligent education within academic institutions. This fusion of machine learning across edge devices and cloud architecture can reshape the educational landscape, ushering in a more innovative and more responsive learning environment that caters to the diverse needs of students and educators alike.
{"title":"Improving Performance of Smart Education Systems by Integrating Machine Learning on Edge Devices and Cloud in Educational Institutions","authors":"Shujie Qiu","doi":"10.1007/s10723-024-09755-5","DOIUrl":"https://doi.org/10.1007/s10723-024-09755-5","url":null,"abstract":"<p>Educational institutions today are embracing technology to enhance education quality through intelligent systems. This study introduces an innovative strategy to boost the performance of such procedures by seamlessly integrating machine learning on edge devices and cloud infrastructure. The proposed framework harnesses the capabilities of a Hybrid 1D Convolutional Neural Network (CNN) and Long Short-Term Memory Network (LSTM) architecture, offering profound insights into intelligent education. Operating at the crossroads of localised and centralised analyses, the Hybrid 1D CNN-LSTM architecture signifies a significant advancement. It directly engages edge devices used by students and educators, laying the groundwork for personalised learning experiences. This architecture adeptly captures the intricacies of various modalities, including text, images, and videos, by harmonising 1D CNN layers and LSTM modules. This approach facilitates the extraction of tailored features from each modality and the exploration of temporal intricacies. Consequently, the architecture provides a holistic comprehension of student engagement and comprehension dynamics, unveiling individual learning preferences. Moreover, the framework seamlessly integrates data from edge devices into the cloud infrastructure, allowing insights from both domains to merge. Educators benefit from attention-enhanced feature maps that encapsulate personalised insights, empowering them to customise content and strategies according to student learning preferences. The approach bridges real-time, localised analysis with comprehensive cloud-mediated insights, paving the path for transformative educational experiences. Empirical validation reinforces the effectiveness of the Hybrid 1D CNN-LSTM architecture, cementing its potential to revolutionise intelligent education within academic institutions. This fusion of machine learning across edge devices and cloud architecture can reshape the educational landscape, ushering in a more innovative and more responsive learning environment that caters to the diverse needs of students and educators alike.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140147452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-11DOI: 10.1007/s10723-024-09745-7
Kamalesh Karmakar, Anurina Tarafdar, Rajib K. Das, Sunirmal Khatua
Workflows are special applications used to solve complex scientific problems. The emerging Workflow as a Service (WaaS) model provides scientists with an effective way of deploying their workflow applications in Cloud environments. The WaaS model can execute multiple workflows in a multi-tenant Cloud environment. Scheduling the tasks of the workflows in the WaaS model has several challenges. The scheduling approach must properly utilize the underlying Cloud resources and satisfy the users’ Quality of Service (QoS) requirements for all the workflows. In this work, we have proposed a heurisine-sensitive workflows in a containerized Cloud environment for the WaaS model. We formulated the problem of minimizing the MIPS (million instructions per second) requirement of tasks while satisfying the deadline of the workflows as a non-linear optimization problem and applied the Lagranges multiplier method to solve it. It allows us to configure/scale the containers’ resources and reduce costs. We also ensure maximum utilization of VM’s resources while allocating containers to VMs. Furthermore, we have proposed an approach to effectively scale containers and VMs to improve the schedulability of the workflows at runtime to deal with the dynamic arrival of the workflows. Extensive experiments and comparisons with other state-of-the-art works show that the proposed approach can significantly improve resource utilization, prevent deadline violation, and reduce the cost of renting Cloud resources for the WaaS model.
{"title":"Cost-efficient Workflow as a Service using Containers","authors":"Kamalesh Karmakar, Anurina Tarafdar, Rajib K. Das, Sunirmal Khatua","doi":"10.1007/s10723-024-09745-7","DOIUrl":"https://doi.org/10.1007/s10723-024-09745-7","url":null,"abstract":"<p>Workflows are special applications used to solve complex scientific problems. The emerging Workflow as a Service (WaaS) model provides scientists with an effective way of deploying their workflow applications in Cloud environments. The WaaS model can execute multiple workflows in a multi-tenant Cloud environment. Scheduling the tasks of the workflows in the WaaS model has several challenges. The scheduling approach must properly utilize the underlying Cloud resources and satisfy the users’ Quality of Service (QoS) requirements for all the workflows. In this work, we have proposed a heurisine-sensitive workflows in a containerized Cloud environment for the WaaS model. We formulated the problem of minimizing the MIPS (million instructions per second) requirement of tasks while satisfying the deadline of the workflows as a non-linear optimization problem and applied the Lagranges multiplier method to solve it. It allows us to configure/scale the containers’ resources and reduce costs. We also ensure maximum utilization of VM’s resources while allocating containers to VMs. Furthermore, we have proposed an approach to effectively scale containers and VMs to improve the schedulability of the workflows at runtime to deal with the dynamic arrival of the workflows. Extensive experiments and comparisons with other state-of-the-art works show that the proposed approach can significantly improve resource utilization, prevent deadline violation, and reduce the cost of renting Cloud resources for the WaaS model.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140097911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-03-09DOI: 10.1007/s10723-024-09756-4
Hongjian Li, Wei Luo, Wenbin Xie, Huaqing Ye, Xiaolin Duan
{"title":"Adaptive Scheduling Framework of Streaming Applications based on Resource Demand Prediction with Hybrid Algorithms","authors":"Hongjian Li, Wei Luo, Wenbin Xie, Huaqing Ye, Xiaolin Duan","doi":"10.1007/s10723-024-09756-4","DOIUrl":"https://doi.org/10.1007/s10723-024-09756-4","url":null,"abstract":"","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140077034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-29DOI: 10.1007/s10723-024-09750-w
Shangshang Wang, Yuqin Jing, Kezhu Wang, Xue Wang
This study tackles the problem of increasing efficiency and scalability in deep neural network (DNN) systems by employing collaborative inference, an approach that is gaining popularity because to its ability to maximize computational resources. It involves splitting a pre-trained DNN model into two parts and running them separately on user equipment (UE) and edge servers. This approach is advantageous because it results in faster and more energy-efficient inference, as computation can be offloaded to edge servers rather than relying solely on UEs. However, a significant challenge of collaborative belief is the dynamic coupling of DNN layers, which makes it difficult to separate and run the layers independently. To address this challenge, we proposed a novel approach to optimize collaborative inference in a multi-agent scenario where a single-edge server coordinates the assumption of multiple UEs. Our proposed method suggests using an autoencoder-based technique to reduce the size of intermediary features and constructing tasks using the deep policy inference Q-inference network’s overhead (DPIQN). To optimize the collaborative inference, employ the Deep Recurrent Policy Inference Q-Network (DRPIQN) technique, which allows for a hybrid action space. The results of the tests demonstrate that this approach can significantly reduce inference latency by up to 56% and energy usage by up to 72% on various networks. Overall, this proposed approach provides an efficient and effective method for implementing collaborative inference in multi-agent scenarios, which could have significant implications for developing DNN systems.
{"title":"Multi-Agent Systems for Collaborative Inference Based on Deep Policy Q-Inference Network","authors":"Shangshang Wang, Yuqin Jing, Kezhu Wang, Xue Wang","doi":"10.1007/s10723-024-09750-w","DOIUrl":"https://doi.org/10.1007/s10723-024-09750-w","url":null,"abstract":"<p>This study tackles the problem of increasing efficiency and scalability in deep neural network (DNN) systems by employing collaborative inference, an approach that is gaining popularity because to its ability to maximize computational resources. It involves splitting a pre-trained DNN model into two parts and running them separately on user equipment (UE) and edge servers. This approach is advantageous because it results in faster and more energy-efficient inference, as computation can be offloaded to edge servers rather than relying solely on UEs. However, a significant challenge of collaborative belief is the dynamic coupling of DNN layers, which makes it difficult to separate and run the layers independently. To address this challenge, we proposed a novel approach to optimize collaborative inference in a multi-agent scenario where a single-edge server coordinates the assumption of multiple UEs. Our proposed method suggests using an autoencoder-based technique to reduce the size of intermediary features and constructing tasks using the deep policy inference Q-inference network’s overhead (DPIQN). To optimize the collaborative inference, employ the Deep Recurrent Policy Inference Q-Network (DRPIQN) technique, which allows for a hybrid action space. The results of the tests demonstrate that this approach can significantly reduce inference latency by up to 56% and energy usage by up to 72% on various networks. Overall, this proposed approach provides an efficient and effective method for implementing collaborative inference in multi-agent scenarios, which could have significant implications for developing DNN systems.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140003991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-29DOI: 10.1007/s10723-024-09752-8
Haotian Pang, Zhanwei Wang
Advancing in communication systems requires nearby devices to act as networks when devices are not in use. Such technology is mobile edge computing, which provides enormous communication services in the network. In this research, we explore a multiuser smart Internet of Vehicles (IoV) network with mobile edge computing (MEC) assistance, where the first edge server can assist in completing the intense computing jobs from the vehicular users. Many currently available works for MEC networks primarily concentrate on minimising system latency to ensure the quality of service (QoS) for users by designing some offloading strategies. Still, they need to account for the retail prices from the server and, as a result, the budgetary constraints of the users. To solve this problem, we present a Dueling Double Deep Q Network (D3QN) with an Optimal Stopping Theory (OST) strategy that helps to solve the multi-task joint edge problems and minimises the offloading problems in MEC-based IoV networks. The multi-task-offloading model aims to increase the likelihood of offloading to the ideal servers by utilising the OST characteristics. Lastly, simulators show how the proposed methods perform better than the traditional ones. The findings demonstrate that the suggested offloading techniques may be successfully applied in mobile nodes and significantly cut the anticipated time required to process the workloads.
{"title":"Dueling Double Deep Q Network Strategy in MEC for Smart Internet of Vehicles Edge Computing Networks","authors":"Haotian Pang, Zhanwei Wang","doi":"10.1007/s10723-024-09752-8","DOIUrl":"https://doi.org/10.1007/s10723-024-09752-8","url":null,"abstract":"<p>Advancing in communication systems requires nearby devices to act as networks when devices are not in use. Such technology is mobile edge computing, which provides enormous communication services in the network. In this research, we explore a multiuser smart Internet of Vehicles (IoV) network with mobile edge computing (MEC) assistance, where the first edge server can assist in completing the intense computing jobs from the vehicular users. Many currently available works for MEC networks primarily concentrate on minimising system latency to ensure the quality of service (QoS) for users by designing some offloading strategies. Still, they need to account for the retail prices from the server and, as a result, the budgetary constraints of the users. To solve this problem, we present a Dueling Double Deep Q Network (D3QN) with an Optimal Stopping Theory (OST) strategy that helps to solve the multi-task joint edge problems and minimises the offloading problems in MEC-based IoV networks. The multi-task-offloading model aims to increase the likelihood of offloading to the ideal servers by utilising the OST characteristics. Lastly, simulators show how the proposed methods perform better than the traditional ones. The findings demonstrate that the suggested offloading techniques may be successfully applied in mobile nodes and significantly cut the anticipated time required to process the workloads.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140004046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-02-28DOI: 10.1007/s10723-024-09746-6
Yanli Xing
Edge computing has emerged as an innovative paradigm, bringing cloud service resources closer to mobile consumers at the network's edge. This proximity enables efficient processing of computationally demanding and time-sensitive tasks. However, the dynamic nature of the edge network, characterized by a high density of devices, diverse mobile usage patterns, a wide range of applications, and sporadic traffic, often leads to uneven resource distribution. This imbalance hampers system efficiency and contributes to task failures. To overcome these challenges, we propose a novel approach known as the DRL-LSTM approach, which combines Deep Reinforcement Learning (DRL) with Long Short-Term Memory (LSTM) architecture. The primary objective of the DRL-LSTM approach is to optimize workload planning in edge computing environments. Leveraging the capabilities of DRL, this approach effectively handles complex and multidimensional workload planning problems. By incorporating LSTM as a recurrent neural network, it captures and models temporal dependencies in sequential data, enabling efficient workload management, reduced service time, and enhanced task completion rates. Additionally, the DRL-LSTM approach integrates Deep-Q-Network (DQN) algorithms to address the complexity and high dimensionality of workload scheduling problems. Through simulations, we demonstrate that the DRL-LSTM approach outperforms alternative approaches regarding service time, virtual machine (VM) utilization, and the rate of failed tasks. The integration of DRL and LSTM enables the process to effectively tackle the challenges associated with workload planning in edge computing, leading to improved system performance. The proposed DRL-LSTM approach offers a promising solution for optimizing workload planning in edge computing environments. Combining the power of Deep Reinforcement Learning, Long Short-Term Memory architecture, and Deep-Q-Network algorithms facilitates efficient resource allocation, reduces service time, and increases task completion rates. It holds significant potential for enhancing the overall performance and effectiveness of edge computing systems.
{"title":"Work Scheduling in Cloud Network Based on Deep Q-LSTM Models for Efficient Resource Utilization","authors":"Yanli Xing","doi":"10.1007/s10723-024-09746-6","DOIUrl":"https://doi.org/10.1007/s10723-024-09746-6","url":null,"abstract":"<p>Edge computing has emerged as an innovative paradigm, bringing cloud service resources closer to mobile consumers at the network's edge. This proximity enables efficient processing of computationally demanding and time-sensitive tasks. However, the dynamic nature of the edge network, characterized by a high density of devices, diverse mobile usage patterns, a wide range of applications, and sporadic traffic, often leads to uneven resource distribution. This imbalance hampers system efficiency and contributes to task failures. To overcome these challenges, we propose a novel approach known as the DRL-LSTM approach, which combines Deep Reinforcement Learning (DRL) with Long Short-Term Memory (LSTM) architecture. The primary objective of the DRL-LSTM approach is to optimize workload planning in edge computing environments. Leveraging the capabilities of DRL, this approach effectively handles complex and multidimensional workload planning problems. By incorporating LSTM as a recurrent neural network, it captures and models temporal dependencies in sequential data, enabling efficient workload management, reduced service time, and enhanced task completion rates. Additionally, the DRL-LSTM approach integrates Deep-Q-Network (DQN) algorithms to address the complexity and high dimensionality of workload scheduling problems. Through simulations, we demonstrate that the DRL-LSTM approach outperforms alternative approaches regarding service time, virtual machine (VM) utilization, and the rate of failed tasks. The integration of DRL and LSTM enables the process to effectively tackle the challenges associated with workload planning in edge computing, leading to improved system performance. The proposed DRL-LSTM approach offers a promising solution for optimizing workload planning in edge computing environments. Combining the power of Deep Reinforcement Learning, Long Short-Term Memory architecture, and Deep-Q-Network algorithms facilitates efficient resource allocation, reduces service time, and increases task completion rates. It holds significant potential for enhancing the overall performance and effectiveness of edge computing systems.</p>","PeriodicalId":54817,"journal":{"name":"Journal of Grid Computing","volume":null,"pages":null},"PeriodicalIF":5.5,"publicationDate":"2024-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140004255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}