iShadow: design of a wearable, real-time mobile gaze tracker

A. Mayberry, Pan Hu, Benjamin M Marlin, C. Salthouse, Deepak Ganesan
{"title":"iShadow: design of a wearable, real-time mobile gaze tracker","authors":"A. Mayberry, Pan Hu, Benjamin M Marlin, C. Salthouse, Deepak Ganesan","doi":"10.1145/2594368.2594388","DOIUrl":null,"url":null,"abstract":"Continuous, real-time tracking of eye gaze is valuable in a variety of scenarios including hands-free interaction with the physical world, detection of unsafe behaviors, leveraging visual context for advertising, life logging, and others. While eye tracking is commonly used in clinical trials and user studies, it has not bridged the gap to everyday consumer use. The challenge is that a real-time eye tracker is a power-hungry and computation-intensive device which requires continuous sensing of the eye using an imager running at many tens of frames per second, and continuous processing of the image stream using sophisticated gaze estimation algorithms. Our key contribution is the design of an eye tracker that dramatically reduces the sensing and computation needs for eye tracking, thereby achieving orders of magnitude reductions in power consumption and form-factor. The key idea is that eye images are extremely redundant, therefore we can estimate gaze by using a small subset of carefully chosen pixels per frame. We instantiate this idea in a prototype hardware platform equipped with a low-power image sensor that provides random access to pixel values, a low-power ARM Cortex M3 microcontroller, and a bluetooth radio to communicate with a mobile phone. The sparse pixel-based gaze estimation algorithm is a multi-layer neural network learned using a state-of-the-art sparsity-inducing regularization function that minimizes the gaze prediction error while simultaneously minimizing the number of pixels used. Our results show that we can operate at roughly 70mW of power, while continuously estimating eye gaze at the rate of 30 Hz with errors of roughly 3 degrees.","PeriodicalId":131209,"journal":{"name":"Proceedings of the 12th annual international conference on Mobile systems, applications, and services","volume":"30 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"79","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 12th annual international conference on Mobile systems, applications, and services","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/2594368.2594388","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 79

Abstract

Continuous, real-time tracking of eye gaze is valuable in a variety of scenarios including hands-free interaction with the physical world, detection of unsafe behaviors, leveraging visual context for advertising, life logging, and others. While eye tracking is commonly used in clinical trials and user studies, it has not bridged the gap to everyday consumer use. The challenge is that a real-time eye tracker is a power-hungry and computation-intensive device which requires continuous sensing of the eye using an imager running at many tens of frames per second, and continuous processing of the image stream using sophisticated gaze estimation algorithms. Our key contribution is the design of an eye tracker that dramatically reduces the sensing and computation needs for eye tracking, thereby achieving orders of magnitude reductions in power consumption and form-factor. The key idea is that eye images are extremely redundant, therefore we can estimate gaze by using a small subset of carefully chosen pixels per frame. We instantiate this idea in a prototype hardware platform equipped with a low-power image sensor that provides random access to pixel values, a low-power ARM Cortex M3 microcontroller, and a bluetooth radio to communicate with a mobile phone. The sparse pixel-based gaze estimation algorithm is a multi-layer neural network learned using a state-of-the-art sparsity-inducing regularization function that minimizes the gaze prediction error while simultaneously minimizing the number of pixels used. Our results show that we can operate at roughly 70mW of power, while continuously estimating eye gaze at the rate of 30 Hz with errors of roughly 3 degrees.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
isshadow:一款可穿戴、实时移动凝视追踪器的设计
持续的、实时的眼睛注视跟踪在各种场景中都很有价值,包括与物理世界的免手交互、不安全行为的检测、利用视觉环境进行广告、生活日志等。虽然眼动追踪通常用于临床试验和用户研究,但它还没有在日常消费者使用中弥合差距。挑战在于,实时眼动仪是一种耗电和计算密集型的设备,它需要使用每秒运行数十帧的成像仪连续感知眼睛,并使用复杂的注视估计算法连续处理图像流。我们的主要贡献是设计了一款眼动仪,大大减少了眼动追踪的传感和计算需求,从而实现了功耗和外形因素的数量级降低。关键思想是眼睛图像是非常冗余的,因此我们可以通过使用每帧精心选择的像素的小子集来估计凝视。我们在一个原型硬件平台中实例化了这个想法,该平台配备了一个低功耗图像传感器,可以随机访问像素值,一个低功耗ARM Cortex M3微控制器,以及一个与移动电话通信的蓝牙无线电。基于稀疏像素的凝视估计算法是一种多层神经网络,使用最先进的稀疏性诱导正则化函数来学习,最小化凝视预测误差,同时最小化所使用的像素数量。我们的研究结果表明,我们可以在大约70mW的功率下工作,同时以30 Hz的速率连续估计眼睛的凝视,误差大约为3度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Characterizing resource usage for mobile web browsing Demo: Yalut -- user-centric social networking overlay Demo: Mapping global mobile performance trends with mobilyzer and mobiPerf Poster: DriveBlue: can bluetooth enhance your driving experience? Balancing design and technology to tackle global grand challenges
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1