{"title":"用于预测数据中心用电效果的基于注意力的集成深度神经网络架构","authors":"Yang-Cheng Shih , Sathesh Tamilarasan , Chin-Sheng Chen , Omid Ali Zargar , Yean-Der Kuan","doi":"10.1016/j.ijft.2024.100866","DOIUrl":null,"url":null,"abstract":"<div><div>Addressing the critical need for enhanced power usage effectiveness in data centers (DCs), this study pioneers an improved convolutional long short-term memory with deep neural network (CLDNN) model, enriched with attention mechanisms for precise DC performance prediction. We rigorously evaluate our model against leading architectures – long short-term memory (LSTM), attention-based (att-LSTM), convolutional LSTM (CNN-LSTM), gated recurrent unit (GRU), and CNN-GRU – to affirm its superiority in predictive accuracy and robustness. The integration of convolutional layers processes hourly data inputs efficiently, reducing complexity and improving pattern detection. A subsequent flattening layer optimizes accuracy, while a dual-layered LSTM and a deep neural network delve into frequency, temporal dynamics, and complex data relationships. Incorporating an attention mechanism into the att-CLDNN model has revolutionized predictive analytics in DC energy management, significantly enhancing accuracy by highlighting crucial data interdependencies. This model's unparalleled precision, evidenced by achieving the lowest Mean Squared Error (MSE) of 0.000179, the minimum Mean Absolute Error (MAE) of 0.01048, and the highest R2 Score of 0.977031, underscores its effectiveness. Crucially, this breakthrough fosters sustainability in energy management, promoting greener DC operations through precise energy use predictions, leading to substantial energy savings and reduced carbon emissions, in alignment with global sustainability objectives.</div></div>","PeriodicalId":36341,"journal":{"name":"International Journal of Thermofluids","volume":"24 ","pages":"Article 100866"},"PeriodicalIF":0.0000,"publicationDate":"2024-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2666202724003070/pdfft?md5=02114f51200ad55f98552ab236201d02&pid=1-s2.0-S2666202724003070-main.pdf","citationCount":"0","resultStr":"{\"title\":\"Attention-based integrated deep neural network architecture for predicting the effectiveness of data center power usage\",\"authors\":\"Yang-Cheng Shih , Sathesh Tamilarasan , Chin-Sheng Chen , Omid Ali Zargar , Yean-Der Kuan\",\"doi\":\"10.1016/j.ijft.2024.100866\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Addressing the critical need for enhanced power usage effectiveness in data centers (DCs), this study pioneers an improved convolutional long short-term memory with deep neural network (CLDNN) model, enriched with attention mechanisms for precise DC performance prediction. We rigorously evaluate our model against leading architectures – long short-term memory (LSTM), attention-based (att-LSTM), convolutional LSTM (CNN-LSTM), gated recurrent unit (GRU), and CNN-GRU – to affirm its superiority in predictive accuracy and robustness. The integration of convolutional layers processes hourly data inputs efficiently, reducing complexity and improving pattern detection. A subsequent flattening layer optimizes accuracy, while a dual-layered LSTM and a deep neural network delve into frequency, temporal dynamics, and complex data relationships. Incorporating an attention mechanism into the att-CLDNN model has revolutionized predictive analytics in DC energy management, significantly enhancing accuracy by highlighting crucial data interdependencies. This model's unparalleled precision, evidenced by achieving the lowest Mean Squared Error (MSE) of 0.000179, the minimum Mean Absolute Error (MAE) of 0.01048, and the highest R2 Score of 0.977031, underscores its effectiveness. Crucially, this breakthrough fosters sustainability in energy management, promoting greener DC operations through precise energy use predictions, leading to substantial energy savings and reduced carbon emissions, in alignment with global sustainability objectives.</div></div>\",\"PeriodicalId\":36341,\"journal\":{\"name\":\"International Journal of Thermofluids\",\"volume\":\"24 \",\"pages\":\"Article 100866\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S2666202724003070/pdfft?md5=02114f51200ad55f98552ab236201d02&pid=1-s2.0-S2666202724003070-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Thermofluids\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2666202724003070\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"Chemical Engineering\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Thermofluids","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666202724003070","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"Chemical Engineering","Score":null,"Total":0}
Attention-based integrated deep neural network architecture for predicting the effectiveness of data center power usage
Addressing the critical need for enhanced power usage effectiveness in data centers (DCs), this study pioneers an improved convolutional long short-term memory with deep neural network (CLDNN) model, enriched with attention mechanisms for precise DC performance prediction. We rigorously evaluate our model against leading architectures – long short-term memory (LSTM), attention-based (att-LSTM), convolutional LSTM (CNN-LSTM), gated recurrent unit (GRU), and CNN-GRU – to affirm its superiority in predictive accuracy and robustness. The integration of convolutional layers processes hourly data inputs efficiently, reducing complexity and improving pattern detection. A subsequent flattening layer optimizes accuracy, while a dual-layered LSTM and a deep neural network delve into frequency, temporal dynamics, and complex data relationships. Incorporating an attention mechanism into the att-CLDNN model has revolutionized predictive analytics in DC energy management, significantly enhancing accuracy by highlighting crucial data interdependencies. This model's unparalleled precision, evidenced by achieving the lowest Mean Squared Error (MSE) of 0.000179, the minimum Mean Absolute Error (MAE) of 0.01048, and the highest R2 Score of 0.977031, underscores its effectiveness. Crucially, this breakthrough fosters sustainability in energy management, promoting greener DC operations through precise energy use predictions, leading to substantial energy savings and reduced carbon emissions, in alignment with global sustainability objectives.