真实环境中的ArUco/凝视跟踪

IF 4.6 2区 教育学 Q1 EDUCATION & EDUCATIONAL RESEARCH Educational Technology & Society Pub Date : 2018-01-14 DOI:10.3929/ETHZ-B-000222486
Vsevolod Peysakhovich, F. Dehais, A. Duchowski
{"title":"真实环境中的ArUco/凝视跟踪","authors":"Vsevolod Peysakhovich, F. Dehais, A. Duchowski","doi":"10.3929/ETHZ-B-000222486","DOIUrl":null,"url":null,"abstract":"The emergence of affordable mobile eye-trackers has allowed to study gaze behavior in real-world environments. However, the gaze mapping from recorded video to a static reference image is a complex and open problem. Finding a reference image within the video frames, i.e., image matching, can give satisfying results, but occluded or overlapped objects are almost impossible to locate using this technique. We suggest using ArUco fiducial markers (and their associated software library available in OpenCV) to map gaze to dynamic Areas Of Interest (AOIs) within a reference image. Although such markers have been used previously, technical details of marker detection and mapping have been sparse. The current approach consists of three steps: (1) define an AOI using markers, then (2) resolve any conflict among overlapping AOIs, and (3) map the gaze point to the \nreference image. A dynamic AOI can be defined using one or more corner markers. When camera rotations are limited and the object is relatively orthogonal to the camera, it is possible to define an AOI using only one corner marker. When the camera rotates, its pose estimation is required to project corner points to the camera image plane. An AOI can also be defined with four corner markers, which has the advantage of robustness with respect to camera rotations, and no a priori required knowledge of the physical dimensions of the object. The two approaches can be combined, e.g., when using four corner markers and one of the markers is most (due to occlusion or view angle), the basis vectors can be used to interpolate the position of the lost marker. When two or more AOIs overlap and all the markers are tracked, gaze should be marked on the AOI closer to the camera. The distance to an object can be defined knowing the length of the object, the number of pixels spanned on the image and the pre-computed camera focal parameter. Once the AOIs are defined and marker overlaps are resolved, the gaze point can be mapped to the coordinates of the reference image using homography.","PeriodicalId":51381,"journal":{"name":"Educational Technology & Society","volume":"14 1","pages":"70-71"},"PeriodicalIF":4.6000,"publicationDate":"2018-01-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"ArUco/Gaze Tracking in Real Environments\",\"authors\":\"Vsevolod Peysakhovich, F. Dehais, A. Duchowski\",\"doi\":\"10.3929/ETHZ-B-000222486\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The emergence of affordable mobile eye-trackers has allowed to study gaze behavior in real-world environments. However, the gaze mapping from recorded video to a static reference image is a complex and open problem. Finding a reference image within the video frames, i.e., image matching, can give satisfying results, but occluded or overlapped objects are almost impossible to locate using this technique. We suggest using ArUco fiducial markers (and their associated software library available in OpenCV) to map gaze to dynamic Areas Of Interest (AOIs) within a reference image. Although such markers have been used previously, technical details of marker detection and mapping have been sparse. The current approach consists of three steps: (1) define an AOI using markers, then (2) resolve any conflict among overlapping AOIs, and (3) map the gaze point to the \\nreference image. A dynamic AOI can be defined using one or more corner markers. When camera rotations are limited and the object is relatively orthogonal to the camera, it is possible to define an AOI using only one corner marker. When the camera rotates, its pose estimation is required to project corner points to the camera image plane. An AOI can also be defined with four corner markers, which has the advantage of robustness with respect to camera rotations, and no a priori required knowledge of the physical dimensions of the object. The two approaches can be combined, e.g., when using four corner markers and one of the markers is most (due to occlusion or view angle), the basis vectors can be used to interpolate the position of the lost marker. When two or more AOIs overlap and all the markers are tracked, gaze should be marked on the AOI closer to the camera. The distance to an object can be defined knowing the length of the object, the number of pixels spanned on the image and the pre-computed camera focal parameter. Once the AOIs are defined and marker overlaps are resolved, the gaze point can be mapped to the coordinates of the reference image using homography.\",\"PeriodicalId\":51381,\"journal\":{\"name\":\"Educational Technology & Society\",\"volume\":\"14 1\",\"pages\":\"70-71\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2018-01-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Educational Technology & Society\",\"FirstCategoryId\":\"95\",\"ListUrlMain\":\"https://doi.org/10.3929/ETHZ-B-000222486\",\"RegionNum\":2,\"RegionCategory\":\"教育学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"EDUCATION & EDUCATIONAL RESEARCH\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Educational Technology & Society","FirstCategoryId":"95","ListUrlMain":"https://doi.org/10.3929/ETHZ-B-000222486","RegionNum":2,"RegionCategory":"教育学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"EDUCATION & EDUCATIONAL RESEARCH","Score":null,"Total":0}
引用次数: 2

摘要

价格合理的移动眼动仪的出现使人们能够研究现实世界环境中的凝视行为。然而,从录制的视频到静态参考图像的凝视映射是一个复杂而开放的问题。在视频帧中找到参考图像,即图像匹配,可以给出令人满意的结果,但使用该技术几乎不可能定位被遮挡或重叠的对象。我们建议使用ArUco基准标记(以及OpenCV中提供的相关软件库)将视线映射到参考图像内的动态感兴趣区域(AOI)。尽管这种标记以前已经使用过,但标记检测和映射的技术细节很少。当前的方法包括三个步骤:(1)使用标记定义AOI,然后(2)解决重叠AOI之间的任何冲突,以及(3)将注视点映射到参考图像。可以使用一个或多个角标记来定义动态AOI。当摄影机旋转受到限制并且对象与摄影机相对正交时,可以仅使用一个角标记来定义AOI。当相机旋转时,需要对其姿态进行估计,以将角点投影到相机图像平面。AOI也可以用四个角标记来定义,这具有相对于相机旋转的鲁棒性的优点,并且不需要先验地了解对象的物理尺寸。这两种方法可以组合,例如,当使用四个角标记并且其中一个标记最多(由于遮挡或视角)时,可以使用基向量来插值丢失标记的位置。当两个或多个AOI重叠并且所有标记都被跟踪时,视线应该标记在离相机更近的AOI上。可以在知道对象的长度、图像上的像素数和预先计算的相机焦点参数的情况下定义到对象的距离。一旦定义了AOI并解决了标记重叠,就可以使用单应性将注视点映射到参考图像的坐标。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
ArUco/Gaze Tracking in Real Environments
The emergence of affordable mobile eye-trackers has allowed to study gaze behavior in real-world environments. However, the gaze mapping from recorded video to a static reference image is a complex and open problem. Finding a reference image within the video frames, i.e., image matching, can give satisfying results, but occluded or overlapped objects are almost impossible to locate using this technique. We suggest using ArUco fiducial markers (and their associated software library available in OpenCV) to map gaze to dynamic Areas Of Interest (AOIs) within a reference image. Although such markers have been used previously, technical details of marker detection and mapping have been sparse. The current approach consists of three steps: (1) define an AOI using markers, then (2) resolve any conflict among overlapping AOIs, and (3) map the gaze point to the reference image. A dynamic AOI can be defined using one or more corner markers. When camera rotations are limited and the object is relatively orthogonal to the camera, it is possible to define an AOI using only one corner marker. When the camera rotates, its pose estimation is required to project corner points to the camera image plane. An AOI can also be defined with four corner markers, which has the advantage of robustness with respect to camera rotations, and no a priori required knowledge of the physical dimensions of the object. The two approaches can be combined, e.g., when using four corner markers and one of the markers is most (due to occlusion or view angle), the basis vectors can be used to interpolate the position of the lost marker. When two or more AOIs overlap and all the markers are tracked, gaze should be marked on the AOI closer to the camera. The distance to an object can be defined knowing the length of the object, the number of pixels spanned on the image and the pre-computed camera focal parameter. Once the AOIs are defined and marker overlaps are resolved, the gaze point can be mapped to the coordinates of the reference image using homography.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Educational Technology & Society
Educational Technology & Society EDUCATION & EDUCATIONAL RESEARCH-
CiteScore
9.10
自引率
2.50%
发文量
1
审稿时长
20 weeks
期刊最新文献
Associating Insulin Instillation with Negative Pressure Wound Therapy Reduces Insulin Resistance in Diabetic Patients with Infected Foot Wounds Solid Pseudopapillary Neoplasm of the Pancreas: Case Report Mass testing and containment of Covid-19 Groin Abscess: Assessing The Diagnostic Accuracy Of Emergency Department Clinicians And Imaging Modalities Fulminant Necrotizing Fasciitis And Myositis With Streptococcal Toxic Shock Syndrome In A Patient With Rheumatoid Arthritis On Tocilizumab: A Case Report
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1