Xu Zhou;Jing Yang;Yijun Li;Shaobo Li;Zhidong Su;Jialin Lu
{"title":"EC-TRL:用于边缘云环境动态资源调度的进化加权聚类和变压器增强强化学习","authors":"Xu Zhou;Jing Yang;Yijun Li;Shaobo Li;Zhidong Su;Jialin Lu","doi":"10.1109/JIOT.2024.3496200","DOIUrl":null,"url":null,"abstract":"With the rapid development of edge computing, devices now offer powerful computing capabilities and diverse applications. However, the surge in smart devices accessing the Internet overwhelms edge servers, which have limited and unevenly distributed resources. This results in challenges like energy management, load balancing (LB), real-time performance, and system complexity. Existing research fails to comprehensively consider these challenges’ combined impact, making it difficult to maximize performance when facing real complex scenarios. To address the above issues, this article proposes an edge cloud resource scheduling scheme based on evolutionary-weighted clustering and transformer-augmented reinforcement learning (EC-TRL). First, server nodes are deployed at the center of user clusters, based on user device locations, to optimize communication delay and evenly distribute resources. Second, the multiobjective scheduling optimization problem under delay constraints is converted into a Markov decision problem, and a deep reinforcement learning method based on soft actor-critic (SAC) is proposed. Finally, actor transformer (AT) and critic transformer (CT) are proposed to improve the network structure of SAC, capture long-term dependencies and complex patterns in long task scheduling sequences, and improve the model’s adaptability and generalization performance in complex dynamic environments. Through comparison experiments with round robin, random, proximal policy optimization, dueling double deep Q-learning network, SAC-L, and SAC-M, the results show that the proposed method improves the optimization performance of energy consumption, LB, and rejection rate of edge cloud resource scheduling by at least 9.57%, 10.90%, and 5.05%.","PeriodicalId":54347,"journal":{"name":"IEEE Internet of Things Journal","volume":"12 6","pages":"7503-7517"},"PeriodicalIF":8.2000,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EC-TRL: Evolutionary-Weighted Clustering and Transformer-Augmented Reinforcement Learning for Dynamic Resource Scheduling in Edge Cloud Environments\",\"authors\":\"Xu Zhou;Jing Yang;Yijun Li;Shaobo Li;Zhidong Su;Jialin Lu\",\"doi\":\"10.1109/JIOT.2024.3496200\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the rapid development of edge computing, devices now offer powerful computing capabilities and diverse applications. However, the surge in smart devices accessing the Internet overwhelms edge servers, which have limited and unevenly distributed resources. This results in challenges like energy management, load balancing (LB), real-time performance, and system complexity. Existing research fails to comprehensively consider these challenges’ combined impact, making it difficult to maximize performance when facing real complex scenarios. To address the above issues, this article proposes an edge cloud resource scheduling scheme based on evolutionary-weighted clustering and transformer-augmented reinforcement learning (EC-TRL). First, server nodes are deployed at the center of user clusters, based on user device locations, to optimize communication delay and evenly distribute resources. Second, the multiobjective scheduling optimization problem under delay constraints is converted into a Markov decision problem, and a deep reinforcement learning method based on soft actor-critic (SAC) is proposed. Finally, actor transformer (AT) and critic transformer (CT) are proposed to improve the network structure of SAC, capture long-term dependencies and complex patterns in long task scheduling sequences, and improve the model’s adaptability and generalization performance in complex dynamic environments. Through comparison experiments with round robin, random, proximal policy optimization, dueling double deep Q-learning network, SAC-L, and SAC-M, the results show that the proposed method improves the optimization performance of energy consumption, LB, and rejection rate of edge cloud resource scheduling by at least 9.57%, 10.90%, and 5.05%.\",\"PeriodicalId\":54347,\"journal\":{\"name\":\"IEEE Internet of Things Journal\",\"volume\":\"12 6\",\"pages\":\"7503-7517\"},\"PeriodicalIF\":8.2000,\"publicationDate\":\"2024-11-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Internet of Things Journal\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10750292/\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Internet of Things Journal","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10750292/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
EC-TRL: Evolutionary-Weighted Clustering and Transformer-Augmented Reinforcement Learning for Dynamic Resource Scheduling in Edge Cloud Environments
With the rapid development of edge computing, devices now offer powerful computing capabilities and diverse applications. However, the surge in smart devices accessing the Internet overwhelms edge servers, which have limited and unevenly distributed resources. This results in challenges like energy management, load balancing (LB), real-time performance, and system complexity. Existing research fails to comprehensively consider these challenges’ combined impact, making it difficult to maximize performance when facing real complex scenarios. To address the above issues, this article proposes an edge cloud resource scheduling scheme based on evolutionary-weighted clustering and transformer-augmented reinforcement learning (EC-TRL). First, server nodes are deployed at the center of user clusters, based on user device locations, to optimize communication delay and evenly distribute resources. Second, the multiobjective scheduling optimization problem under delay constraints is converted into a Markov decision problem, and a deep reinforcement learning method based on soft actor-critic (SAC) is proposed. Finally, actor transformer (AT) and critic transformer (CT) are proposed to improve the network structure of SAC, capture long-term dependencies and complex patterns in long task scheduling sequences, and improve the model’s adaptability and generalization performance in complex dynamic environments. Through comparison experiments with round robin, random, proximal policy optimization, dueling double deep Q-learning network, SAC-L, and SAC-M, the results show that the proposed method improves the optimization performance of energy consumption, LB, and rejection rate of edge cloud resource scheduling by at least 9.57%, 10.90%, and 5.05%.
期刊介绍:
The EEE Internet of Things (IoT) Journal publishes articles and review articles covering various aspects of IoT, including IoT system architecture, IoT enabling technologies, IoT communication and networking protocols such as network coding, and IoT services and applications. Topics encompass IoT's impacts on sensor technologies, big data management, and future internet design for applications like smart cities and smart homes. Fields of interest include IoT architecture such as things-centric, data-centric, service-oriented IoT architecture; IoT enabling technologies and systematic integration such as sensor technologies, big sensor data management, and future Internet design for IoT; IoT services, applications, and test-beds such as IoT service middleware, IoT application programming interface (API), IoT application design, and IoT trials/experiments; IoT standardization activities and technology development in different standard development organizations (SDO) such as IEEE, IETF, ITU, 3GPP, ETSI, etc.