{"title":"深度学习模型训练在嵌入式GPU设备上的适用性:实证研究","authors":"Po-Hsuan Chou, Chao Wang, Chih-Shuo Mei","doi":"10.1109/MECO58584.2023.10155048","DOIUrl":null,"url":null,"abstract":"The wide applications of deep learning techniques have motivated the inclusion of both embedded GPU devices and workstation GPU cards into contemporary Industrial Internet-of-Things (IIoT) systems. Due to substantial differences between the two types of GPUs, deep-learning model training in its current practice is run on GPU cards, and embedded GPU devices are used for inferences or partial model training at best. To supply with empirical evidence and aid the decision of deep learning workload placement, this paper reports a set of experiments on the timeliness and energy efficiency of each GPU type, running both Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) model training. The results suggest that embedded GPUs did save the total energy cost despite the longer response time, but the amount of energy saving might not be significant in a practical sense. Further in this paper we report a case study for prognostics applications using LSTM. The results suggest that, by comparison, an embedded GPU may save about 90 percent of energy consumption at the cost of doubling the application response time. But neither the save in energy cost nor the increase in response time is significant enough to impact the application. These findings suggest that it may be feasible to place model training workload on either workstation GPU or embedded GPU.","PeriodicalId":187825,"journal":{"name":"2023 12th Mediterranean Conference on Embedded Computing (MECO)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Applicability of Deep Learning Model Trainings on Embedded GPU Devices: An Empirical Study\",\"authors\":\"Po-Hsuan Chou, Chao Wang, Chih-Shuo Mei\",\"doi\":\"10.1109/MECO58584.2023.10155048\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The wide applications of deep learning techniques have motivated the inclusion of both embedded GPU devices and workstation GPU cards into contemporary Industrial Internet-of-Things (IIoT) systems. Due to substantial differences between the two types of GPUs, deep-learning model training in its current practice is run on GPU cards, and embedded GPU devices are used for inferences or partial model training at best. To supply with empirical evidence and aid the decision of deep learning workload placement, this paper reports a set of experiments on the timeliness and energy efficiency of each GPU type, running both Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) model training. The results suggest that embedded GPUs did save the total energy cost despite the longer response time, but the amount of energy saving might not be significant in a practical sense. Further in this paper we report a case study for prognostics applications using LSTM. The results suggest that, by comparison, an embedded GPU may save about 90 percent of energy consumption at the cost of doubling the application response time. But neither the save in energy cost nor the increase in response time is significant enough to impact the application. These findings suggest that it may be feasible to place model training workload on either workstation GPU or embedded GPU.\",\"PeriodicalId\":187825,\"journal\":{\"name\":\"2023 12th Mediterranean Conference on Embedded Computing (MECO)\",\"volume\":\"10 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 12th Mediterranean Conference on Embedded Computing (MECO)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MECO58584.2023.10155048\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 12th Mediterranean Conference on Embedded Computing (MECO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MECO58584.2023.10155048","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Applicability of Deep Learning Model Trainings on Embedded GPU Devices: An Empirical Study
The wide applications of deep learning techniques have motivated the inclusion of both embedded GPU devices and workstation GPU cards into contemporary Industrial Internet-of-Things (IIoT) systems. Due to substantial differences between the two types of GPUs, deep-learning model training in its current practice is run on GPU cards, and embedded GPU devices are used for inferences or partial model training at best. To supply with empirical evidence and aid the decision of deep learning workload placement, this paper reports a set of experiments on the timeliness and energy efficiency of each GPU type, running both Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) model training. The results suggest that embedded GPUs did save the total energy cost despite the longer response time, but the amount of energy saving might not be significant in a practical sense. Further in this paper we report a case study for prognostics applications using LSTM. The results suggest that, by comparison, an embedded GPU may save about 90 percent of energy consumption at the cost of doubling the application response time. But neither the save in energy cost nor the increase in response time is significant enough to impact the application. These findings suggest that it may be feasible to place model training workload on either workstation GPU or embedded GPU.