{"title":"Hybrid metaheuristics for selective inference task offloading under time and energy constraints for real-time IoT sensing systems","authors":"Abdelkarim Ben Sada, Amar Khelloufi, Abdenacer Naouri, Huansheng Ning, Sahraoui Dhelim","doi":"10.1007/s10586-024-04578-1","DOIUrl":null,"url":null,"abstract":"<p>The recent widespread of AI-powered real-time applications necessitates the use of edge computing for inference task offloading. Power constrained edge devices are required to balance between processing inference tasks locally or offload to edge servers. This decision is determined according to the time constraint demanded by the real-time nature of applications, and the energy constraint dictated by the device’s power budget. This problem is further exacerbated in the case of systems leveraging multiple local inference models varying in size and accuracy. In this work, we tackle the problem of assigning inference models to inference tasks either using local inference models or by offloading to edge servers under time and energy constraints while maximizing the overall accuracy of the system. This problem is shown to be strongly NP-hard and therefore, we propose a hybrid genetic algorithm (HGSTO) to solve this problem. We leverage the speed of simulated annealing (SA) with the accuracy of genetic algorithms (GA) to develop a hybrid, fast and accurate algorithm compared with classic GA, SA and Particle Swarm Optimization (PSO). Experiment results show that HGSTO achieved on-par or higher accuracy than GA while resulting in significantly lower scheduling times compared to other schemes.</p>","PeriodicalId":501576,"journal":{"name":"Cluster Computing","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Cluster Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s10586-024-04578-1","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
The recent widespread of AI-powered real-time applications necessitates the use of edge computing for inference task offloading. Power constrained edge devices are required to balance between processing inference tasks locally or offload to edge servers. This decision is determined according to the time constraint demanded by the real-time nature of applications, and the energy constraint dictated by the device’s power budget. This problem is further exacerbated in the case of systems leveraging multiple local inference models varying in size and accuracy. In this work, we tackle the problem of assigning inference models to inference tasks either using local inference models or by offloading to edge servers under time and energy constraints while maximizing the overall accuracy of the system. This problem is shown to be strongly NP-hard and therefore, we propose a hybrid genetic algorithm (HGSTO) to solve this problem. We leverage the speed of simulated annealing (SA) with the accuracy of genetic algorithms (GA) to develop a hybrid, fast and accurate algorithm compared with classic GA, SA and Particle Swarm Optimization (PSO). Experiment results show that HGSTO achieved on-par or higher accuracy than GA while resulting in significantly lower scheduling times compared to other schemes.