Yihao Wang, Ling Gao, J. Ren, Rui Cao, Hai Wang, Jie Zheng, Quanli Gao
{"title":"ATO-EDGE: Adaptive Task Offloading for Deep Learning in Resource-Constrained Edge Computing Systems","authors":"Yihao Wang, Ling Gao, J. Ren, Rui Cao, Hai Wang, Jie Zheng, Quanli Gao","doi":"10.1109/ICPADS53394.2021.00025","DOIUrl":null,"url":null,"abstract":"On-device deep learning enables mobile devices to perform complex tasks, such as object detection and voice translation, regardless of the network condition. The advanced deep learning model gives an excellent performance, also leads to a heavy burden on resource-limited devices (i.e., mobile devices). To speed up the on-device deep learning. Prior studies focus on developing lightweight network architecture for real-time inference by sacrificing model accuracy. This paper presents ATO-EDGE: adaptive task offloading for deep learning based on edge computing. Considering three optimization goals, energy consumption, accuracy, and latency, ATO-EDGE leverages an offline pre-trained model to select a suitable deep learning model on a specific device to process the given task. We apply our approach to object detection and evaluate it on Jetson TX2, Xilinx ZYNQ 7020, and Raspberry 3B+. The deep learning model candidates contain ten typical object detection models trained on Microsoft COCO 2017 dataset. We obtain, on average, 28.25%, 35.44%, and 0.9 improvements respectively for latency, energy consumption, and mAP (mean average precision) when compared to the SOTA DETR model on the Raspberry Pi.","PeriodicalId":309508,"journal":{"name":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 27th International Conference on Parallel and Distributed Systems (ICPADS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICPADS53394.2021.00025","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
On-device deep learning enables mobile devices to perform complex tasks, such as object detection and voice translation, regardless of the network condition. The advanced deep learning model gives an excellent performance, also leads to a heavy burden on resource-limited devices (i.e., mobile devices). To speed up the on-device deep learning. Prior studies focus on developing lightweight network architecture for real-time inference by sacrificing model accuracy. This paper presents ATO-EDGE: adaptive task offloading for deep learning based on edge computing. Considering three optimization goals, energy consumption, accuracy, and latency, ATO-EDGE leverages an offline pre-trained model to select a suitable deep learning model on a specific device to process the given task. We apply our approach to object detection and evaluate it on Jetson TX2, Xilinx ZYNQ 7020, and Raspberry 3B+. The deep learning model candidates contain ten typical object detection models trained on Microsoft COCO 2017 dataset. We obtain, on average, 28.25%, 35.44%, and 0.9 improvements respectively for latency, energy consumption, and mAP (mean average precision) when compared to the SOTA DETR model on the Raspberry Pi.