首页 > 最新文献

Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition最新文献

英文 中文
3D Face Reconstruction Error Decomposed: A Modular Benchmark for Fair and Fast Method Evaluation. 三维人脸重构误差分解:一种公平快速评价方法的模块化基准。
Evangelos Sariyanidi, Claudio Ferrari, Federico Nocentini, Stefano Berretti, Andrea Cavallaro, Birkan Tunc

Computing the standard benchmark metric for 3D face reconstruction, namely geometric error, requires a number of steps, such as mesh cropping, rigid alignment, or point correspondence. Current benchmark tools are monolithic (they implement a specific combination of these steps), even though there is no consensus on the best way to measure error. We present a toolkit for a Modularized 3D Face reconstruction Benchmark (M3DFB), where the fundamental components of error computation are segregated and interchangeable, allowing one to quantify the effect of each. Furthermore, we propose a new component, namely correction, and present a computationally efficient approach that penalizes for mesh topology inconsistency. Using this toolkit, we test 16 error estimators with 10 reconstruction methods on two real and two synthetic datasets. Critically, the widely used ICP-based estimator provides the worst benchmarking performance, as it significantly alters the true ranking of the top-5 reconstruction methods. Notably, the correlation of ICP with the true error can be as low as 0.41. Moreover, non-rigid alignment leads to significant improvement (correlation larger than 0.90), highlighting the importance of annotating 3D landmarks on datasets. Finally, the proposed correction scheme, together with non-rigid warping, leads to an accuracy on a par with the best non-rigid ICP-based estimators, but runs an order of magnitude faster. Our open-source codebase is designed for researchers to easily compare alternatives for each component, thus helping accelerating progress in benchmarking for 3D face reconstruction and, furthermore, supporting the improvement of learned reconstruction methods, which depend on accurate error estimation for effective training.

计算三维人脸重建的标准基准度量,即几何误差,需要许多步骤,如网格裁剪,刚性对齐或点对应。当前的基准测试工具是单一的(它们实现了这些步骤的特定组合),尽管对于测量误差的最佳方法还没有达成共识。我们提出了一个模块化3D人脸重建基准(M3DFB)的工具包,其中误差计算的基本组件是分离的和可互换的,允许量化每个组件的影响。此外,我们提出了一个新的组件,即校正,并提出了一种计算效率高的方法来惩罚网格拓扑不一致。使用该工具包,我们在两个真实数据集和两个合成数据集上测试了16个误差估计器和10种重构方法。关键的是,广泛使用的基于icp的估计器提供了最差的基准测试性能,因为它显着改变了前5种重建方法的真实排名。值得注意的是,ICP与真实误差的相关性可以低至0.41。此外,非刚性对齐导致显著改善(相关性大于0.90),突出了在数据集上注释3D地标的重要性。最后,提出的校正方案,加上非刚性翘曲,导致精度与最佳的非刚性icp估计器相当,但运行速度快了一个数量级。我们的开源代码库是为研究人员设计的,可以方便地比较每个组件的替代方案,从而有助于加快3D人脸重建基准测试的进展,此外,支持学习重建方法的改进,这些方法依赖于准确的误差估计来进行有效的训练。
{"title":"3D Face Reconstruction Error Decomposed: A Modular Benchmark for Fair and Fast Method Evaluation.","authors":"Evangelos Sariyanidi, Claudio Ferrari, Federico Nocentini, Stefano Berretti, Andrea Cavallaro, Birkan Tunc","doi":"10.1109/fg61629.2025.11099357","DOIUrl":"https://doi.org/10.1109/fg61629.2025.11099357","url":null,"abstract":"<p><p>Computing the standard benchmark metric for 3D face reconstruction, namely geometric error, requires a number of steps, such as mesh cropping, rigid alignment, or point correspondence. Current benchmark tools are monolithic (they implement a specific combination of these steps), even though there is no consensus on the best way to measure error. We present a toolkit for a Modularized 3D Face reconstruction Benchmark (M3DFB), where the fundamental components of error computation are segregated and interchangeable, allowing one to quantify the effect of each. Furthermore, we propose a new component, namely <i>correction</i>, and present a computationally efficient approach that penalizes for mesh topology inconsistency. Using this toolkit, we test 16 error estimators with 10 reconstruction methods on two real and two synthetic datasets. Critically, the widely used ICP-based estimator provides the worst benchmarking performance, as it significantly alters the true ranking of the top-5 reconstruction methods. Notably, the correlation of ICP with the true error can be as low as 0.41. Moreover, non-rigid alignment leads to significant improvement (correlation larger than 0.90), highlighting the importance of annotating 3D landmarks on datasets. Finally, the proposed correction scheme, together with non-rigid warping, leads to an accuracy on a par with the best non-rigid ICP-based estimators, but runs an order of magnitude faster. Our open-source codebase is designed for researchers to easily compare alternatives for each component, thus helping accelerating progress in benchmarking for 3D face reconstruction and, furthermore, supporting the improvement of learned reconstruction methods, which depend on accurate error estimation for effective training.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2025 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12380054/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144982056","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beyond FACS: Data-driven Facial Expression Dictionaries, with Application to Predicting Autism. 超越FACS:数据驱动的面部表情词典,用于预测自闭症。
Evangelos Sariyanidi, Lisa Yankowitz, Robert T Schultz, John D Herrington, Birkan Tunc, Jeffrey Cohn

The Facial Action Coding System (FACS) has been used by numerous studies to investigate the links between facial behavior and mental health. The laborious and costly process of FACS coding has motivated the development of machine learning frameworks for Action Unit (AU) detection. Despite intense efforts spanning three decades, the detection accuracy for many AUs is considered to be below the threshold needed for behavioral research. Also, many AUs are excluded altogether, making it impossible to fulfill the ultimate goal of FACS-the representation of any facial expression in its entirety. This paper considers an alternative approach. Instead of creating automated tools that mimic FACS experts, we propose to use a new coding system that mimics the key properties of FACS. Specifically, we construct a data-driven coding system called the Facial Basis, which contains units that correspond to localized and interpretable 3D facial movements, and overcomes three structural limitations of automated FACS coding. First, the proposed method is completely unsupervised, bypassing costly, laborious and variable manual annotation. Second, Facial Basis reconstructs all observable movement, rather than relying on a limited repertoire of recognizable movements (as in automated FACS). Finally, the Facial Basis units are additive, whereas AUs may fail detection when they appear in a non-additive combination. The proposed method outperforms the most frequently used AU detector in predicting autism diagnosis from in-person and remote conversations, highlighting the importance of encoding facial behavior comprehensively. To our knowledge, Facial Basis is the first alternative to FACS for deconstructing facial expressions in videos into localized movements. We provide an open source implementation of the method at github.com/sariyanidi/FacialBasis.

面部动作编码系统(FACS)已被许多研究用于调查面部行为与心理健康之间的联系。费力和昂贵的FACS编码过程推动了行动单元(AU)检测机器学习框架的发展。尽管经过了三十年的努力,许多AUs的检测精度被认为低于行为研究所需的阈值。此外,许多人工智能被完全排除在外,这使得facs的最终目标——任何面部表情的完整表示——无法实现。本文考虑了另一种方法。我们建议使用模仿FACS关键属性的新编码系统,而不是创建模仿FACS专家的自动化工具。具体来说,我们构建了一个数据驱动的编码系统,称为面部基础,其中包含对应于本地化和可解释的3D面部运动的单元,并克服了自动化FACS编码的三个结构限制。首先,该方法是完全无监督的,绕过了昂贵、费力和可变的人工注释。其次,面部基础重建所有可观察到的运动,而不是依赖于有限的可识别运动(如自动FACS)。最后,面部基础单元是相加的,而当它们出现在非相加的组合中时,au可能无法被检测到。该方法在预测面对面和远程对话的自闭症诊断方面优于最常用的AU检测器,突出了全面编码面部行为的重要性。据我们所知,面部基础是FACS的第一个替代方案,用于将视频中的面部表情解构为局部运动。我们在github.com/sariyanidi/FacialBasis上提供了该方法的开源实现。
{"title":"Beyond FACS: Data-driven Facial Expression Dictionaries, with Application to Predicting Autism.","authors":"Evangelos Sariyanidi, Lisa Yankowitz, Robert T Schultz, John D Herrington, Birkan Tunc, Jeffrey Cohn","doi":"10.1109/fg61629.2025.11099288","DOIUrl":"https://doi.org/10.1109/fg61629.2025.11099288","url":null,"abstract":"<p><p>The Facial Action Coding System (FACS) has been used by numerous studies to investigate the links between facial behavior and mental health. The laborious and costly process of FACS coding has motivated the development of machine learning frameworks for Action Unit (AU) detection. Despite intense efforts spanning three decades, the detection accuracy for many AUs is considered to be below the threshold needed for behavioral research. Also, many AUs are excluded altogether, making it impossible to fulfill the ultimate goal of FACS-the representation of <i>any</i> facial expression in its entirety. This paper considers an alternative approach. Instead of creating automated tools that mimic FACS experts, we propose to use a new coding system that mimics the key properties of FACS. Specifically, we construct a data-driven coding system called the Facial Basis, which contains units that correspond to localized and interpretable 3D facial movements, and overcomes three structural limitations of automated FACS coding. First, the proposed method is completely unsupervised, bypassing costly, laborious and variable manual annotation. Second, Facial Basis reconstructs all observable movement, rather than relying on a limited repertoire of recognizable movements (as in automated FACS). Finally, the Facial Basis units are additive, whereas AUs may fail detection when they appear in a non-additive combination. The proposed method outperforms the most frequently used AU detector in predicting autism diagnosis from in-person and remote conversations, highlighting the importance of encoding facial behavior comprehensively. To our knowledge, Facial Basis is the first alternative to FACS for deconstructing facial expressions in videos into localized movements. We provide an open source implementation of the method at github.com/sariyanidi/FacialBasis.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2025 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12369895/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144981960","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Goals, Tasks, and Bonds: Toward the Computational Assessment of Therapist Versus Client Perception of Working Alliance. 目标、任务和纽带:治疗师与客户对工作联盟认知的计算评估》。
Alexandria K Vail, Jeffrey Girard, Lauren Bylsma, Jeffrey Cohn, Jay Fournier, Holly Swartz, Louis-Philippe Morency

Early client dropout is one of the most significant challenges facing psychotherapy: recent studies suggest that at least one in five clients will leave treatment prematurely. Clients may terminate therapy for various reasons, but one of the most common causes is the lack of a strong working alliance. The concept of working alliance captures the collaborative relationship between a client and their therapist when working toward the progress and recovery of the client seeking treatment. Unfortunately, clients are often unwilling to directly express dissatisfaction in care until they have already decided to terminate therapy. On the other side, therapists may miss subtle signs of client discontent during treatment before it is too late. In this work, we demonstrate that nonverbal behavior analysis may aid in bridging this gap. The present study focuses primarily on the head gestures of both the client and therapist, contextualized within conversational turn-taking actions between the pair during psychotherapy sessions. We identify multiple behavior patterns suggestive of an individual's perspective on the working alliance; interestingly, these patterns also differ between the client and the therapist. These patterns inform the development of predictive models for self-reported ratings of working alliance, which demonstrate significant predictive power for both client and therapist ratings. Future applications of such models may stimulate preemptive intervention to strengthen a weak working alliance, whether explicitly attempting to repair the existing alliance or establishing a more suitable client-therapist pairing, to ensure that clients encounter fewer barriers to receiving the treatment they need.

客户过早退出治疗是心理治疗面临的最大挑战之一:最近的研究表明,至少有五分之一的客户会过早离开治疗。客户终止治疗的原因多种多样,但最常见的原因之一是缺乏强有力的工作联盟。工作联盟的概念体现了客户与治疗师之间的合作关系,这种合作关系旨在帮助寻求治疗的客户取得进步和康复。不幸的是,客户往往不愿意直接表达对治疗的不满,直到他们已经决定终止治疗。另一方面,治疗师可能会在治疗过程中错过客户不满的细微迹象,否则就为时已晚。在这项研究中,我们证明了非语言行为分析可以帮助弥合这一差距。本研究主要关注求助者和治疗师的头部手势,并将其与心理治疗过程中双方的轮流对话行为结合起来。我们发现了暗示个人对工作联盟看法的多种行为模式;有趣的是,这些模式在求助者和治疗师之间也存在差异。这些模式为工作联盟自我报告评级预测模型的开发提供了信息,该模型对求助者和治疗师的评级都具有显著的预测能力。这种模型在未来的应用可能会刺激先发制人的干预,以加强薄弱的工作联盟,无论是明确尝试修复现有的联盟,还是建立更合适的客户-治疗师配对,以确保客户在接受他们需要的治疗时遇到更少的障碍。
{"title":"Goals, Tasks, and Bonds: Toward the Computational Assessment of Therapist Versus Client Perception of Working Alliance.","authors":"Alexandria K Vail, Jeffrey Girard, Lauren Bylsma, Jeffrey Cohn, Jay Fournier, Holly Swartz, Louis-Philippe Morency","doi":"10.1109/fg52635.2021.9667021","DOIUrl":"10.1109/fg52635.2021.9667021","url":null,"abstract":"<p><p>Early client dropout is one of the most significant challenges facing psychotherapy: recent studies suggest that at least one in five clients will leave treatment prematurely. Clients may terminate therapy for various reasons, but one of the most common causes is the lack of a strong <i>working alliance</i>. The concept of working alliance captures the collaborative relationship between a client and their therapist when working toward the progress and recovery of the client seeking treatment. Unfortunately, clients are often unwilling to directly express dissatisfaction in care until they have already decided to terminate therapy. On the other side, therapists may miss subtle signs of client discontent during treatment before it is too late. In this work, we demonstrate that nonverbal behavior analysis may aid in bridging this gap. The present study focuses primarily on the head gestures of both the client and therapist, contextualized within conversational turn-taking actions between the pair during psychotherapy sessions. We identify multiple behavior patterns suggestive of an individual's perspective on the working alliance; interestingly, these patterns also differ between the client and the therapist. These patterns inform the development of predictive models for self-reported ratings of working alliance, which demonstrate significant predictive power for both client and therapist ratings. Future applications of such models may stimulate preemptive intervention to strengthen a weak working alliance, whether explicitly attempting to repair the existing alliance or establishing a more suitable client-therapist pairing, to ensure that clients encounter fewer barriers to receiving the treatment they need.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2021 ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2021-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9355426/pdf/nihms-1771359.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"40700885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simple and Effective Approaches for Uncertainty Prediction in Facial Action Unit Intensity Regression. 面部动作单元强度回归中不确定性预测的简单有效方法。
Torsten Wörtwein, Louis-Philippe Morency

Knowing how much to trust a prediction is important for many critical applications. We describe two simple approaches to estimate uncertainty in regression prediction tasks and compare their performance and complexity against popular approaches. We operationalize uncertainty in regression as the absolute error between a model's prediction and the ground truth. Our two proposed approaches use a secondary model to predict the uncertainty of a primary predictive model. Our first approach leverages the assumption that similar observations are likely to have similar uncertainty and predicts uncertainty with a non-parametric method. Our second approach trains a secondary model to directly predict the uncertainty of the primary predictive model. Both approaches outperform other established uncertainty estimation approaches on the MNIST, DISFA, and BP4D+ datasets. Furthermore, we observe that approaches that directly predict the uncertainty generally perform better than approaches that indirectly estimate uncertainty.

知道在多大程度上信任预测对于许多关键应用程序都很重要。我们描述了两种简单的方法来估计回归预测任务中的不确定性,并将它们的性能和复杂性与流行的方法进行了比较。我们将回归中的不确定性作为模型预测与基本事实之间的绝对误差来操作。我们提出的两种方法使用二级模型来预测初级预测模型的不确定性。我们的第一种方法利用了类似观测可能具有类似不确定性的假设,并使用非参数方法预测不确定性。我们的第二种方法训练一个辅助模型来直接预测主要预测模型的不确定性。在MNIST、DISFA和BP4D+数据集上,这两种方法都优于其他已建立的不确定性估计方法。此外,我们观察到直接预测不确定性的方法通常比间接估计不确定性的方法表现得更好。
{"title":"Simple and Effective Approaches for Uncertainty Prediction in Facial Action Unit Intensity Regression.","authors":"Torsten Wörtwein,&nbsp;Louis-Philippe Morency","doi":"10.1109/fg47880.2020.00045","DOIUrl":"https://doi.org/10.1109/fg47880.2020.00045","url":null,"abstract":"<p><p>Knowing how much to trust a prediction is important for many critical applications. We describe two simple approaches to estimate uncertainty in regression prediction tasks and compare their performance and complexity against popular approaches. We operationalize uncertainty in regression as the absolute error between a model's prediction and the ground truth. Our two proposed approaches use a secondary model to predict the uncertainty of a primary predictive model. Our first approach leverages the assumption that similar observations are likely to have similar uncertainty and predicts uncertainty with a non-parametric method. Our second approach trains a secondary model to directly predict the uncertainty of the primary predictive model. Both approaches outperform other established uncertainty estimation approaches on the MNIST, DISFA, and BP4D+ datasets. Furthermore, we observe that approaches that directly predict the uncertainty generally perform better than approaches that indirectly estimate uncertainty.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2020 ","pages":"452-456"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/fg47880.2020.00045","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"25453101","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Nonverbal Behavioral Patterns Predict Social Rejection Elicited Aggression. 非语言行为模式预测社会排斥引发的攻击。
Megan Quarmley, Zhibo Yang, Shahrukh Athar, Gregory Zelinksy, Dimitris Samaras, Johanna M Jarcho

Peer-based aggression following social rejection is a costly and prevalent problem for which existing treatments have had little success. This may be because aggression is a complex process influenced by current states of attention and arousal, which are difficult to measure on a moment to moment basis via self report. It is therefore crucial to identify nonverbal behavioral indices of attention and arousal that predict subsequent aggression. We used Support Vector Machines (SVMs) and eye gaze duration and pupillary response features, measured during positive and negative peer-based social interactions, to predict subsequent aggressive behavior towards those same peers. We found that eye gaze and pupillary reactivity not only predicted aggressive behavior, but performed better than models that included information about the participant's exposure to harsh parenting or trait aggression. Eye gaze and pupillary reactivity models also performed equally as well as those that included information about peer reputation (e.g. whether the peer was rejecting or accepting). This is the first study to decode nonverbal eye behavior during social interaction to predict social rejection-elicited aggression.

社会排斥后的同伴攻击是一种昂贵而普遍的问题,现有的治疗方法收效甚微。这可能是因为攻击是一个复杂的过程,受到当前注意力和觉醒状态的影响,这很难通过自我报告来衡量。因此,识别非语言行为的注意和唤醒指标,预测随后的攻击行为是至关重要的。我们使用支持向量机(svm)和眼睛注视时间和瞳孔反应特征,在积极和消极的基于同伴的社会互动中测量,来预测随后对相同同伴的攻击行为。我们发现,眼睛注视和瞳孔反应不仅能预测攻击行为,而且比包含参与者受到严厉父母教育或特质攻击信息的模型表现得更好。眼睛凝视和瞳孔反应模型的表现也与那些包含同伴声誉信息(例如同伴是拒绝还是接受)的模型一样好。这是第一个在社会互动中解码非语言的眼睛行为来预测社会排斥引发的攻击的研究。
{"title":"Nonverbal Behavioral Patterns Predict Social Rejection Elicited Aggression.","authors":"Megan Quarmley,&nbsp;Zhibo Yang,&nbsp;Shahrukh Athar,&nbsp;Gregory Zelinksy,&nbsp;Dimitris Samaras,&nbsp;Johanna M Jarcho","doi":"10.1109/fg47880.2020.00111","DOIUrl":"https://doi.org/10.1109/fg47880.2020.00111","url":null,"abstract":"<p><p>Peer-based aggression following social rejection is a costly and prevalent problem for which existing treatments have had little success. This may be because aggression is a complex process influenced by current states of attention and arousal, which are difficult to measure on a moment to moment basis via self report. It is therefore crucial to identify nonverbal behavioral indices of attention and arousal that predict subsequent aggression. We used Support Vector Machines (SVMs) and eye gaze duration and pupillary response features, measured during positive and negative peer-based social interactions, to predict subsequent aggressive behavior towards those same peers. We found that eye gaze and pupillary reactivity not only predicted aggressive behavior, but performed better than models that included information about the participant's exposure to harsh parenting or trait aggression. Eye gaze and pupillary reactivity models also performed equally as well as those that included information about peer reputation (e.g. whether the peer was rejecting or accepting). This is the first study to decode nonverbal eye behavior during social interaction to predict social rejection-elicited aggression.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2020 ","pages":"557-561"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/fg47880.2020.00111","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"39774870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detecting Depression Severity by Interpretable Representations of Motion Dynamics. 通过可解释的运动动态表示检测抑郁症严重程度
Anis Kacem, Zakia Hammal, Mohamed Daoudi, Jeffrey Cohn

Recent breakthroughs in deep learning using automated measurement of face and head motion have made possible the first objective measurement of depression severity. While powerful, deep learning approaches lack interpretability. We developed an interpretable method of automatically measuring depression severity that uses barycentric coordinates of facial landmarks and a Lie-algebra based rotation matrix of 3D head motion. Using these representations, kinematic features are extracted, preprocessed, and encoded using Gaussian Mixture Models (GMM) and Fisher vector encoding. A multi-class SVM is used to classify the encoded facial and head movement dynamics into three levels of depression severity. The proposed approach was evaluated in adults with history of chronic depression. The method approached the classification accuracy of state-of-the-art deep learning while enabling clinically and theoretically relevant findings. The velocity and acceleration of facial movement strongly mapped onto depression severity symptoms consistent with clinical data and theory.

最近,利用自动测量面部和头部运动的深度学习技术取得了突破性进展,首次实现了对抑郁症严重程度的客观测量。深度学习方法虽然功能强大,但缺乏可解释性。我们开发了一种可解释的自动测量抑郁症严重程度的方法,该方法使用面部地标的巴里中心坐标和基于李代数的三维头部运动旋转矩阵。利用这些表征,可以提取运动学特征,进行预处理,并使用高斯混合模型(GMM)和费雪向量编码进行编码。使用多类 SVM 将编码后的面部和头部运动动态分为三个抑郁严重程度等级。所提出的方法在有慢性抑郁症病史的成年人中进行了评估。该方法接近最先进的深度学习的分类准确性,同时还能得出临床和理论相关的结论。面部运动的速度和加速度与抑郁症的严重程度症状密切相关,这与临床数据和理论一致。
{"title":"Detecting Depression Severity by Interpretable Representations of Motion Dynamics.","authors":"Anis Kacem, Zakia Hammal, Mohamed Daoudi, Jeffrey Cohn","doi":"10.1109/FG.2018.00116","DOIUrl":"10.1109/FG.2018.00116","url":null,"abstract":"<p><p>Recent breakthroughs in deep learning using automated measurement of face and head motion have made possible the first objective measurement of depression severity. While powerful, deep learning approaches lack interpretability. We developed an interpretable method of automatically measuring depression severity that uses barycentric coordinates of facial landmarks and a Lie-algebra based rotation matrix of 3D head motion. Using these representations, kinematic features are extracted, preprocessed, and encoded using Gaussian Mixture Models (GMM) and Fisher vector encoding. A multi-class SVM is used to classify the encoded facial and head movement dynamics into three levels of depression severity. The proposed approach was evaluated in adults with history of chronic depression. The method approached the classification accuracy of state-of-the-art deep learning while enabling clinically and theoretically relevant findings. The velocity and acceleration of facial movement strongly mapped onto depression severity symptoms consistent with clinical data and theory.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2018 ","pages":"739-745"},"PeriodicalIF":0.0,"publicationDate":"2018-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6157749/pdf/nihms950419.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"36538326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning for Action and Gesture Recognition in Image Sequences: A Survey 图像序列中动作和手势识别的深度学习:综述
Maryam Asadi-Aghbolaghi, Albert Clapés, M. Bellantonio, H. Escalante, V. Ponce-López, Xavier Baró, Isabelle M Guyon, S. Kasaei, Sergio Escalera
{"title":"Deep Learning for Action and Gesture Recognition in Image Sequences: A Survey","authors":"Maryam Asadi-Aghbolaghi, Albert Clapés, M. Bellantonio, H. Escalante, V. Ponce-López, Xavier Baró, Isabelle M Guyon, S. Kasaei, Sergio Escalera","doi":"10.1007/978-3-319-57021-1_19","DOIUrl":"https://doi.org/10.1007/978-3-319-57021-1_19","url":null,"abstract":"","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"1 1","pages":"539-578"},"PeriodicalIF":0.0,"publicationDate":"2017-07-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80009798","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 48
FERA 2017 - Addressing Head Pose in the Third Facial Expression Recognition and Analysis Challenge. FERA 2017 -在第三次面部表情识别和分析挑战中解决头部姿势。
Michel F Valstar, Enrique Sánchez-Lozano, Jeffrey F Cohn, László A Jeni, Jeffrey M Girard, Zheng Zhang, Lijun Yin, Maja Pantic

The field of Automatic Facial Expression Analysis has grown rapidly in recent years. However, despite progress in new approaches as well as benchmarking efforts, most evaluations still focus on either posed expressions, near-frontal recordings, or both. This makes it hard to tell how existing expression recognition approaches perform under conditions where faces appear in a wide range of poses (or camera views), displaying ecologically valid expressions. The main obstacle for assessing this is the availability of suitable data, and the challenge proposed here addresses this limitation. The FG 2017 Facial Expression Recognition and Analysis challenge (FERA 2017) extends FERA 2015 to the estimation of Action Units occurrence and intensity under different camera views. In this paper we present the third challenge in automatic recognition of facial expressions, to be held in conjunction with the 12th IEEE conference on Face and Gesture Recognition, May 2017, in Washington, United States. Two sub-challenges are defined: the detection of AU occurrence, and the estimation of AU intensity. In this work we outline the evaluation protocol, the data used, and the results of a baseline method for both sub-challenges.

面部表情自动分析是近年来发展迅速的研究领域。然而,尽管在新方法和基准测试方面取得了进展,大多数评估仍然集中在摆姿势的表情,近额部记录,或两者兼而有之。这使得现有的表情识别方法很难在面部出现在各种姿势(或相机视图)的情况下表现如何,显示生态有效的表情。评估这一点的主要障碍是适当数据的可用性,这里提出的挑战解决了这一限制。FG 2017面部表情识别和分析挑战(FERA 2017)将FERA 2015扩展到不同相机视图下动作单元发生和强度的估计。在本文中,我们提出了面部表情自动识别的第三个挑战,该挑战将于2017年5月在美国华盛顿举行的第12届IEEE面部和手势识别会议上举行。定义了两个子挑战:AU发生的检测和AU强度的估计。在这项工作中,我们概述了评估方案,使用的数据,以及两个子挑战的基线方法的结果。
{"title":"FERA 2017 - Addressing Head Pose in the Third Facial Expression Recognition and Analysis Challenge.","authors":"Michel F Valstar, Enrique Sánchez-Lozano, Jeffrey F Cohn, László A Jeni, Jeffrey M Girard, Zheng Zhang, Lijun Yin, Maja Pantic","doi":"10.1109/FG.2017.107","DOIUrl":"10.1109/FG.2017.107","url":null,"abstract":"<p><p>The field of Automatic Facial Expression Analysis has grown rapidly in recent years. However, despite progress in new approaches as well as benchmarking efforts, most evaluations still focus on either posed expressions, near-frontal recordings, or both. This makes it hard to tell how existing expression recognition approaches perform under conditions where faces appear in a wide range of poses (or camera views), displaying ecologically valid expressions. The main obstacle for assessing this is the availability of suitable data, and the challenge proposed here addresses this limitation. The FG 2017 Facial Expression Recognition and Analysis challenge (FERA 2017) extends FERA 2015 to the estimation of Action Units occurrence and intensity under different camera views. In this paper we present the third challenge in automatic recognition of facial expressions, to be held in conjunction with the 12th IEEE conference on Face and Gesture Recognition, May 2017, in Washington, United States. Two sub-challenges are defined: the detection of AU occurrence, and the estimation of AU intensity. In this work we outline the evaluation protocol, the data used, and the results of a baseline method for both sub-challenges.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2017 ","pages":"839-847"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/FG.2017.107","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35967120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 123
Sayette Group Formation Task (GFT) Spontaneous Facial Expression Database. Sayette Group Formation Task (GFT)自发面部表情数据库。
Jeffrey M Girard, Wen-Sheng Chu, László A Jeni, Jeffrey F Cohn, Fernando De la Torre, Michael A Sayette

Despite the important role that facial expressions play in interpersonal communication and our knowledge that interpersonal behavior is influenced by social context, no currently available facial expression database includes multiple interacting participants. The Sayette Group Formation Task (GFT) database addresses the need for well-annotated video of multiple participants during unscripted interactions. The database includes 172,800 video frames from 96 participants in 32 three-person groups. To aid in the development of automated facial expression analysis systems, GFT includes expert annotations of FACS occurrence and intensity, facial landmark tracking, and baseline results for linear SVM, deep learning, active patch learning, and personalized classification. Baseline performance is quantified and compared using identical partitioning and a variety of metrics (including means and confidence intervals). The highest performance scores were found for the deep learning and active patch learning methods. Learn more at http://osf.io/7wcyz.

尽管面部表情在人际交往中发挥着重要作用,我们也知道人际行为受社会环境的影响,但目前还没有包含多个互动参与者的面部表情数据库。Sayette Group Formation Task (GFT)数据库解决了在无脚本交互过程中对多个参与者的良好注释视频的需求。该数据库包括来自32个三人小组的96名参与者的172,800个视频帧。为了帮助开发自动面部表情分析系统,GFT包括FACS发生和强度的专家注释、面部地标跟踪、线性支持向量机的基线结果、深度学习、主动补丁学习和个性化分类。使用相同的划分和各种度量(包括平均值和置信区间)对基线性能进行量化和比较。深度学习和主动补丁学习方法的性能得分最高。更多信息请访问http://osf.io/7wcyz。
{"title":"Sayette Group Formation Task (GFT) Spontaneous Facial Expression Database.","authors":"Jeffrey M Girard, Wen-Sheng Chu, László A Jeni, Jeffrey F Cohn, Fernando De la Torre, Michael A Sayette","doi":"10.1109/FG.2017.144","DOIUrl":"10.1109/FG.2017.144","url":null,"abstract":"<p><p>Despite the important role that facial expressions play in interpersonal communication and our knowledge that interpersonal behavior is influenced by social context, no currently available facial expression database includes multiple interacting participants. The Sayette Group Formation Task (GFT) database addresses the need for well-annotated video of multiple participants during unscripted interactions. The database includes 172,800 video frames from 96 participants in 32 three-person groups. To aid in the development of automated facial expression analysis systems, GFT includes expert annotations of FACS occurrence and intensity, facial landmark tracking, and baseline results for linear SVM, deep learning, active patch learning, and personalized classification. Baseline performance is quantified and compared using identical partitioning and a variety of metrics (including means and confidence intervals). The highest performance scores were found for the deep learning and active patch learning methods. Learn more at http://osf.io/7wcyz.</p>","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"2017 ","pages":"581-588"},"PeriodicalIF":0.0,"publicationDate":"2017-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1109/FG.2017.144","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"35966631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
Challenges in Multi-modal Gesture Recognition 多模态手势识别中的挑战
Sergio Escalera, V. Athitsos, Isabelle M Guyon
{"title":"Challenges in Multi-modal Gesture Recognition","authors":"Sergio Escalera, V. Athitsos, Isabelle M Guyon","doi":"10.1007/978-3-319-57021-1_1","DOIUrl":"https://doi.org/10.1007/978-3-319-57021-1_1","url":null,"abstract":"","PeriodicalId":87341,"journal":{"name":"Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition","volume":"83 1","pages":"72:1-72:54"},"PeriodicalIF":0.0,"publicationDate":"2016-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88215947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
期刊
Proceedings of the ... International Conference on Automatic Face and Gesture Recognition. IEEE International Conference on Automatic Face & Gesture Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1