Koustabh Dolui, Sam Michiels, D. Hughes, H. Hallez
{"title":"Context Aware Adaptive ML Inference in Mobile-Cloud Applications","authors":"Koustabh Dolui, Sam Michiels, D. Hughes, H. Hallez","doi":"10.1109/WoWMoM54355.2022.00031","DOIUrl":null,"url":null,"abstract":"With the emergence of mobile devices having enough resources to execute real-time ML inference, deployment opportunities arise on mobile devices while keeping privacy-sensitive data close to the source and reducing server load. Moreover, offloading inference to a cloud server facilitates deployment of neural network-based applications on resource-constrained devices. Depending on the application goals and execution context of the application, the optimal deployment on either cloud server or mobile device varies during the lifetime of an application. In this paper, we propose a context-aware middleware that enables optimization of deployed application software to satisfy the application’s functional goals in accordance with changing execution context and environmental conditions. We facilitate system design through the abstraction of deployed software components as states and make use of finite state machines and contextual triggers to model the reconfiguration of the system. We evaluate our framework using a real-world nutritional monitoring application via food image recognition deployed in a two-tier mobile and cloud architecture. We compare the proposed solution with various static deployments of the application and show that our approach can react to changing application goals at run-time in order to reduce server load and thereby increase scalability.","PeriodicalId":275324,"journal":{"name":"2022 IEEE 23rd International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 23rd International Symposium on a World of Wireless, Mobile and Multimedia Networks (WoWMoM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WoWMoM54355.2022.00031","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
With the emergence of mobile devices having enough resources to execute real-time ML inference, deployment opportunities arise on mobile devices while keeping privacy-sensitive data close to the source and reducing server load. Moreover, offloading inference to a cloud server facilitates deployment of neural network-based applications on resource-constrained devices. Depending on the application goals and execution context of the application, the optimal deployment on either cloud server or mobile device varies during the lifetime of an application. In this paper, we propose a context-aware middleware that enables optimization of deployed application software to satisfy the application’s functional goals in accordance with changing execution context and environmental conditions. We facilitate system design through the abstraction of deployed software components as states and make use of finite state machines and contextual triggers to model the reconfiguration of the system. We evaluate our framework using a real-world nutritional monitoring application via food image recognition deployed in a two-tier mobile and cloud architecture. We compare the proposed solution with various static deployments of the application and show that our approach can react to changing application goals at run-time in order to reduce server load and thereby increase scalability.