Nishan Gunawardena, Gough Yumu Lui, Jeewani Anupama Ginige, Bahman Javadi
{"title":"利用边缘智能和模型优化的智能手机眼球跟踪系统","authors":"Nishan Gunawardena, Gough Yumu Lui, Jeewani Anupama Ginige, Bahman Javadi","doi":"arxiv-2408.12463","DOIUrl":null,"url":null,"abstract":"A significant limitation of current smartphone-based eye-tracking algorithms\nis their low accuracy when applied to video-type visual stimuli, as they are\ntypically trained on static images. Also, the increasing demand for real-time\ninteractive applications like games, VR, and AR on smartphones requires\novercoming the limitations posed by resource constraints such as limited\ncomputational power, battery life, and network bandwidth. Therefore, we\ndeveloped two new smartphone eye-tracking techniques for video-type visuals by\ncombining Convolutional Neural Networks (CNN) with two different Recurrent\nNeural Networks (RNN), namely Long Short Term Memory (LSTM) and Gated Recurrent\nUnit (GRU). Our CNN+LSTM and CNN+GRU models achieved an average Root Mean\nSquare Error of 0.955cm and 1.091cm, respectively. To address the computational\nconstraints of smartphones, we developed an edge intelligence architecture to\nenhance the performance of smartphone-based eye tracking. We applied various\noptimisation methods like quantisation and pruning to deep learning models for\nbetter energy, CPU, and memory usage on edge devices, focusing on real-time\nprocessing. Using model quantisation, the model inference time in the CNN+LSTM\nand CNN+GRU models was reduced by 21.72% and 19.50%, respectively, on edge\ndevices.","PeriodicalId":501291,"journal":{"name":"arXiv - CS - Performance","volume":"88 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Smartphone-based Eye Tracking System using Edge Intelligence and Model Optimisation\",\"authors\":\"Nishan Gunawardena, Gough Yumu Lui, Jeewani Anupama Ginige, Bahman Javadi\",\"doi\":\"arxiv-2408.12463\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A significant limitation of current smartphone-based eye-tracking algorithms\\nis their low accuracy when applied to video-type visual stimuli, as they are\\ntypically trained on static images. Also, the increasing demand for real-time\\ninteractive applications like games, VR, and AR on smartphones requires\\novercoming the limitations posed by resource constraints such as limited\\ncomputational power, battery life, and network bandwidth. Therefore, we\\ndeveloped two new smartphone eye-tracking techniques for video-type visuals by\\ncombining Convolutional Neural Networks (CNN) with two different Recurrent\\nNeural Networks (RNN), namely Long Short Term Memory (LSTM) and Gated Recurrent\\nUnit (GRU). Our CNN+LSTM and CNN+GRU models achieved an average Root Mean\\nSquare Error of 0.955cm and 1.091cm, respectively. To address the computational\\nconstraints of smartphones, we developed an edge intelligence architecture to\\nenhance the performance of smartphone-based eye tracking. We applied various\\noptimisation methods like quantisation and pruning to deep learning models for\\nbetter energy, CPU, and memory usage on edge devices, focusing on real-time\\nprocessing. Using model quantisation, the model inference time in the CNN+LSTM\\nand CNN+GRU models was reduced by 21.72% and 19.50%, respectively, on edge\\ndevices.\",\"PeriodicalId\":501291,\"journal\":{\"name\":\"arXiv - CS - Performance\",\"volume\":\"88 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Performance\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.12463\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Performance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.12463","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Smartphone-based Eye Tracking System using Edge Intelligence and Model Optimisation
A significant limitation of current smartphone-based eye-tracking algorithms
is their low accuracy when applied to video-type visual stimuli, as they are
typically trained on static images. Also, the increasing demand for real-time
interactive applications like games, VR, and AR on smartphones requires
overcoming the limitations posed by resource constraints such as limited
computational power, battery life, and network bandwidth. Therefore, we
developed two new smartphone eye-tracking techniques for video-type visuals by
combining Convolutional Neural Networks (CNN) with two different Recurrent
Neural Networks (RNN), namely Long Short Term Memory (LSTM) and Gated Recurrent
Unit (GRU). Our CNN+LSTM and CNN+GRU models achieved an average Root Mean
Square Error of 0.955cm and 1.091cm, respectively. To address the computational
constraints of smartphones, we developed an edge intelligence architecture to
enhance the performance of smartphone-based eye tracking. We applied various
optimisation methods like quantisation and pruning to deep learning models for
better energy, CPU, and memory usage on edge devices, focusing on real-time
processing. Using model quantisation, the model inference time in the CNN+LSTM
and CNN+GRU models was reduced by 21.72% and 19.50%, respectively, on edge
devices.