{"title":"Multilayer multivariate forecasting network for precise resource utilization prediction in edge data centers","authors":"Shivani Tripathi , Priyadarshni , Rajiv Misra , T.N. Singh","doi":"10.1016/j.future.2024.107692","DOIUrl":null,"url":null,"abstract":"<div><div>Efficient resource management and accurate prediction of cloud workloads are vital in modern cloud computing environments, where dynamic and volatile workloads present significant challenges. Traditional forecasting models often fail to fully capture the intricate temporal dependencies and non-linear patterns inherent in cloud data, leading to inefficiencies in resource utilization. To overcome these limitations, this research introduces the MultiLayer Multivariate Resource Predictor (MMRP), a novel deep learning architecture that seamlessly integrates a Multi-Head Attention Transformer model with Convolutional Neural Networks and Bidirectional Long Short-Term Memory units. The proposed model is designed to excel in capturing long-range dependencies and complex patterns, thereby significantly enhancing the accuracy of workload predictions. Extensive, rigorous experimentation using real-world Alibaba and Google cluster traces reveals that the proposed model consistently outperforms existing state-of-the-art models and related cloud resource utilization prediction in both univariate and multivariate time series forecasting tasks. The model demonstrates a remarkable improvement in prediction performance, with an average R squared increase of 5.76% and a Mean Absolute Percentage Error reduction of 84.9% compared to the best-performing baseline models. Furthermore, our model achieves a significant reduction in Root Mean Square Error by approximately 35.34% and decreases Mean Absolute Error by about 39.49% on average. Its scalability and adaptability across various cloud environments underscore the proposed model’s potential to optimize resource allocation, paving the way for more efficient and reliable cloud-based systems.</div></div>","PeriodicalId":55132,"journal":{"name":"Future Generation Computer Systems-The International Journal of Escience","volume":"166 ","pages":"Article 107692"},"PeriodicalIF":6.2000,"publicationDate":"2025-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Future Generation Computer Systems-The International Journal of Escience","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167739X24006563","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, THEORY & METHODS","Score":null,"Total":0}
引用次数: 0
Abstract
Efficient resource management and accurate prediction of cloud workloads are vital in modern cloud computing environments, where dynamic and volatile workloads present significant challenges. Traditional forecasting models often fail to fully capture the intricate temporal dependencies and non-linear patterns inherent in cloud data, leading to inefficiencies in resource utilization. To overcome these limitations, this research introduces the MultiLayer Multivariate Resource Predictor (MMRP), a novel deep learning architecture that seamlessly integrates a Multi-Head Attention Transformer model with Convolutional Neural Networks and Bidirectional Long Short-Term Memory units. The proposed model is designed to excel in capturing long-range dependencies and complex patterns, thereby significantly enhancing the accuracy of workload predictions. Extensive, rigorous experimentation using real-world Alibaba and Google cluster traces reveals that the proposed model consistently outperforms existing state-of-the-art models and related cloud resource utilization prediction in both univariate and multivariate time series forecasting tasks. The model demonstrates a remarkable improvement in prediction performance, with an average R squared increase of 5.76% and a Mean Absolute Percentage Error reduction of 84.9% compared to the best-performing baseline models. Furthermore, our model achieves a significant reduction in Root Mean Square Error by approximately 35.34% and decreases Mean Absolute Error by about 39.49% on average. Its scalability and adaptability across various cloud environments underscore the proposed model’s potential to optimize resource allocation, paving the way for more efficient and reliable cloud-based systems.
期刊介绍:
Computing infrastructures and systems are constantly evolving, resulting in increasingly complex and collaborative scientific applications. To cope with these advancements, there is a growing need for collaborative tools that can effectively map, control, and execute these applications.
Furthermore, with the explosion of Big Data, there is a requirement for innovative methods and infrastructures to collect, analyze, and derive meaningful insights from the vast amount of data generated. This necessitates the integration of computational and storage capabilities, databases, sensors, and human collaboration.
Future Generation Computer Systems aims to pioneer advancements in distributed systems, collaborative environments, high-performance computing, and Big Data analytics. It strives to stay at the forefront of developments in grids, clouds, and the Internet of Things (IoT) to effectively address the challenges posed by these wide-area, fully distributed sensing and computing systems.