How much training data for facial action unit detection?

Jeffrey M Girard, Jeffrey F Cohn, László A Jeni, Simon Lucey, Fernando De la Torre
{"title":"How much training data for facial action unit detection?","authors":"Jeffrey M Girard,&nbsp;Jeffrey F Cohn,&nbsp;László A Jeni,&nbsp;Simon Lucey,&nbsp;Fernando De la Torre","doi":"10.1109/FG.2015.7163106","DOIUrl":null,"url":null,"abstract":"<p><p>By systematically varying the number of subjects and the number of frames per subject, we explored the influence of training set size on appearance and shape-based approaches to facial action unit (AU) detection. Digital video and expert coding of spontaneous facial activity from 80 subjects (over 350,000 frames) were used to train and test support vector machine classifiers. Appearance features were shape-normalized SIFT descriptors and shape features were 66 facial landmarks. Ten-fold cross-validation was used in all evaluations. Number of subjects and number of frames per subject differentially affected appearance and shape-based classifiers. For appearance features, which are high-dimensional, increasing the number of training subjects from 8 to 64 incrementally improved performance, regardless of the number of frames taken from each subject (ranging from 450 through 3600). In contrast, for shape features, increases in the number of training subjects and frames were associated with mixed results. In summary, maximal performance was attained using appearance features from large numbers of subjects with as few as 450 frames per subject. These findings suggest that variation in the number of subjects rather than number of frames per subject yields most efficient performance.</p>","PeriodicalId":91494,"journal":{"name":"IEEE International Conference on Automatic Face & Gesture Recognition and Workshops","volume":"1 ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2015-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/FG.2015.7163106","citationCount":"35","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE International Conference on Automatic Face & Gesture Recognition and Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FG.2015.7163106","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 35

Abstract

By systematically varying the number of subjects and the number of frames per subject, we explored the influence of training set size on appearance and shape-based approaches to facial action unit (AU) detection. Digital video and expert coding of spontaneous facial activity from 80 subjects (over 350,000 frames) were used to train and test support vector machine classifiers. Appearance features were shape-normalized SIFT descriptors and shape features were 66 facial landmarks. Ten-fold cross-validation was used in all evaluations. Number of subjects and number of frames per subject differentially affected appearance and shape-based classifiers. For appearance features, which are high-dimensional, increasing the number of training subjects from 8 to 64 incrementally improved performance, regardless of the number of frames taken from each subject (ranging from 450 through 3600). In contrast, for shape features, increases in the number of training subjects and frames were associated with mixed results. In summary, maximal performance was attained using appearance features from large numbers of subjects with as few as 450 frames per subject. These findings suggest that variation in the number of subjects rather than number of frames per subject yields most efficient performance.

Abstract Image

Abstract Image

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
面部动作单元检测的训练数据有多少?
通过系统地改变受试者的数量和每个受试者的帧数,我们探索了训练集大小对外观和基于形状的面部动作单元(AU)检测方法的影响。使用80个被试(超过35万帧)的自发面部活动的数字视频和专家编码来训练和测试支持向量机分类器。外观特征是形状归一化SIFT描述子,形状特征是66个面部标志。所有评价均采用十倍交叉验证。受试者的数量和每个受试者的帧数对外观和基于形状的分类器有不同的影响。对于高维的外观特征,将训练对象的数量从8个增加到64个,无论从每个对象获取的帧数(从450到3600不等)如何,都可以逐步提高性能。相比之下,对于形状特征,训练对象和框架数量的增加与混合结果相关。总而言之,使用大量受试者的外观特征获得最大性能,每个受试者只有450帧。这些发现表明,受试者数量的变化而不是每个受试者的帧数产生最有效的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021, Jodhpur, India, December 15-18, 2021 Message from the General and Program Chairs FG 2020 Proof-Theoretic Aspects of Hybrid Type-Logical Grammars Undecidability of a Newly Proposed Calculus for CatLog3 On the Computational Complexity of Head Movement and Affix Hopping
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1