基于头部姿态估计的物体自动视觉跟踪定性分析

Ayeshka Abeysinghe, Isuri Devlini Arachchige, Pradeepa Samarasinghe, Vidushani Dhanawansa, Menan Velayuthan
{"title":"基于头部姿态估计的物体自动视觉跟踪定性分析","authors":"Ayeshka Abeysinghe, Isuri Devlini Arachchige, Pradeepa Samarasinghe, Vidushani Dhanawansa, Menan Velayuthan","doi":"10.1109/ICAC57685.2022.10025053","DOIUrl":null,"url":null,"abstract":"An automated approach for object tracking and gaze estimation via head pose estimation is crucial, to facilitate a range of applications in the domain of -human-computer interfacing, this includes the analysis of head movement with respect to a stimulus in assessing one’s level of attention. While varied approaches for gaze estimation and object tracking exist, their suitability within such applications have not been justified. In order to address this gap, this paper conducts a quantitative comparison of existing models for gaze estimation including Mediapipe and standalone models of Openface and custom head pose estimation with MTCNN face detection; and object detection including models from CSRT object tracker, YOLO object detector, and a custom object detector. The accuracy of the aforementioned models were compared against the annotations of the EYEDIAP dataset, to evaluate their accuracy both relative and non-relative to each other. The analysis revealed that the custom object detector and the Openface models are relatively more accurate than the others when comparing the number of annotations, absolute mean error, and the relationship between x displacement-yaw, and y displacement-pitch, and thereby can be used in combination for gaze tracking tasks.","PeriodicalId":292397,"journal":{"name":"2022 4th International Conference on Advancements in Computing (ICAC)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Qualitative Analysis of Automated Visual Tracking of Objects Through Head Pose Estimation\",\"authors\":\"Ayeshka Abeysinghe, Isuri Devlini Arachchige, Pradeepa Samarasinghe, Vidushani Dhanawansa, Menan Velayuthan\",\"doi\":\"10.1109/ICAC57685.2022.10025053\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"An automated approach for object tracking and gaze estimation via head pose estimation is crucial, to facilitate a range of applications in the domain of -human-computer interfacing, this includes the analysis of head movement with respect to a stimulus in assessing one’s level of attention. While varied approaches for gaze estimation and object tracking exist, their suitability within such applications have not been justified. In order to address this gap, this paper conducts a quantitative comparison of existing models for gaze estimation including Mediapipe and standalone models of Openface and custom head pose estimation with MTCNN face detection; and object detection including models from CSRT object tracker, YOLO object detector, and a custom object detector. The accuracy of the aforementioned models were compared against the annotations of the EYEDIAP dataset, to evaluate their accuracy both relative and non-relative to each other. The analysis revealed that the custom object detector and the Openface models are relatively more accurate than the others when comparing the number of annotations, absolute mean error, and the relationship between x displacement-yaw, and y displacement-pitch, and thereby can be used in combination for gaze tracking tasks.\",\"PeriodicalId\":292397,\"journal\":{\"name\":\"2022 4th International Conference on Advancements in Computing (ICAC)\",\"volume\":\"3 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 4th International Conference on Advancements in Computing (ICAC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAC57685.2022.10025053\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 4th International Conference on Advancements in Computing (ICAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAC57685.2022.10025053","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

一种通过头部姿势估计来进行目标跟踪和凝视估计的自动化方法至关重要,这有助于在人机界面领域的一系列应用,这包括在评估一个人的注意力水平时分析与刺激有关的头部运动。虽然存在各种凝视估计和目标跟踪方法,但它们在此类应用中的适用性尚未得到证明。为了解决这一问题,本文对现有的凝视估计模型(包括Mediapipe和Openface的独立模型)和基于MTCNN人脸检测的自定义头姿估计模型进行了定量比较;以及对象检测,包括来自CSRT对象跟踪器、YOLO对象检测器和自定义对象检测器的模型。将上述模型的准确性与EYEDIAP数据集的注释进行比较,以评估它们之间的相对和非相对准确性。分析表明,在对比标注数量、绝对平均误差、x位移-偏航和y位移-俯仰关系等方面,自定义目标检测器和Openface模型相对更准确,可以组合用于注视跟踪任务。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Qualitative Analysis of Automated Visual Tracking of Objects Through Head Pose Estimation
An automated approach for object tracking and gaze estimation via head pose estimation is crucial, to facilitate a range of applications in the domain of -human-computer interfacing, this includes the analysis of head movement with respect to a stimulus in assessing one’s level of attention. While varied approaches for gaze estimation and object tracking exist, their suitability within such applications have not been justified. In order to address this gap, this paper conducts a quantitative comparison of existing models for gaze estimation including Mediapipe and standalone models of Openface and custom head pose estimation with MTCNN face detection; and object detection including models from CSRT object tracker, YOLO object detector, and a custom object detector. The accuracy of the aforementioned models were compared against the annotations of the EYEDIAP dataset, to evaluate their accuracy both relative and non-relative to each other. The analysis revealed that the custom object detector and the Openface models are relatively more accurate than the others when comparing the number of annotations, absolute mean error, and the relationship between x displacement-yaw, and y displacement-pitch, and thereby can be used in combination for gaze tracking tasks.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Emission Activity Parts Extraction using custom Named Entity Recognition Solid-Waste Management System for Urban Sri Lanka Using IOT and Machine Learning SMART DIARY: Autonomous System for Daily Diary Creation and Prioritization of Daily Activities for Improved Well-Being Using Neural Networks and Machine Learning Assistant Zone – Homeschooling Assistance System based on Natural Language Processing DevFlair: A Framework to Automate the Pre-screening Process of Software Engineering Job Candidates
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1