首页 > 最新文献

2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)最新文献

英文 中文
Emotion-Aware Transformer Encoder for Empathetic Dialogue Generation 情感感知变压器编码器共情对话的产生
Raman Goel, Seba Susan, Sachin Vashisht, Armaan Dhanda
Modern day conversational agents are trained to emulate the manner in which humans communicate. To emotionally bond with the user, these virtual agents need to be aware of the affective state of the user. Transformers are the recent state of the art in sequence-to-sequence learning that involves training an encoder-decoder model with word embeddings from utterance-response pairs. We propose an emotion-aware transformer encoder for capturing the emotional quotient in the user utterance in order to generate human-like empathetic responses. The contributions of our paper are as follows: 1) An emotion detector module trained on the input utterances determines the affective state of the user in the initial phase 2) A novel transformer encoder is proposed that adds and normalizes the word embedding with emotion embedding thereby integrating the semantic and affective aspects of the input utterance 3) The encoder and decoder stacks belong to the Transformer-XL architecture which is the recent state of the art in language modeling. Experimentation on the benchmark Facebook AI empathetic dialogue dataset confirms the efficacy of our model from the higher BLEU-4 scores achieved for the generated responses as compared to existing methods. Emotionally intelligent virtual agents are now a reality and inclusion of affect as a modality in all human-machine interfaces is foreseen in the immediate future.
现代的会话代理被训练成模仿人类交流的方式。为了与用户建立情感联系,这些虚拟代理需要了解用户的情感状态。变形金刚是序列到序列学习的最新技术,它包括用话语-响应对中的词嵌入来训练编码器-解码器模型。我们提出了一种情感感知转换器编码器,用于捕获用户话语中的情商,以产生类似人类的移情反应。本文的贡献如下:1)在输入话语上训练的情感检测器模块确定了用户在初始阶段的情感状态2)提出了一种新的变压器编码器,该编码器将情感嵌入与单词嵌入添加并规范化,从而集成了输入话语的语义和情感方面3)编码器和解码器堆栈属于transformer - xl架构,这是语言建模中最新的技术。在基准Facebook AI移情对话数据集上的实验证实了我们模型的有效性,与现有方法相比,生成的响应获得了更高的BLEU-4分数。情感智能虚拟代理现在已经成为现实,在不久的将来,可以预见所有人机界面都将情感作为一种形式。
{"title":"Emotion-Aware Transformer Encoder for Empathetic Dialogue Generation","authors":"Raman Goel, Seba Susan, Sachin Vashisht, Armaan Dhanda","doi":"10.1109/ACIIW52867.2021.9666315","DOIUrl":"https://doi.org/10.1109/ACIIW52867.2021.9666315","url":null,"abstract":"Modern day conversational agents are trained to emulate the manner in which humans communicate. To emotionally bond with the user, these virtual agents need to be aware of the affective state of the user. Transformers are the recent state of the art in sequence-to-sequence learning that involves training an encoder-decoder model with word embeddings from utterance-response pairs. We propose an emotion-aware transformer encoder for capturing the emotional quotient in the user utterance in order to generate human-like empathetic responses. The contributions of our paper are as follows: 1) An emotion detector module trained on the input utterances determines the affective state of the user in the initial phase 2) A novel transformer encoder is proposed that adds and normalizes the word embedding with emotion embedding thereby integrating the semantic and affective aspects of the input utterance 3) The encoder and decoder stacks belong to the Transformer-XL architecture which is the recent state of the art in language modeling. Experimentation on the benchmark Facebook AI empathetic dialogue dataset confirms the efficacy of our model from the higher BLEU-4 scores achieved for the generated responses as compared to existing methods. Emotionally intelligent virtual agents are now a reality and inclusion of affect as a modality in all human-machine interfaces is foreseen in the immediate future.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126862580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Dream Net: a privacy preserving continual leaming model for face emotion recognition 梦境网:一种保护隐私的面部情绪识别持续学习模型
M. Mainsant, M. Solinas, M. Reyboz, C. Godin, M. Mermillod
Continual learning is a growing challenge of artificial intelligence. Among algorithms alleviating catastrophic forgetting that have been developed in the past years, only few studies were focused on face emotion recognition. In parallel, the field of emotion recognition raised the ethical issue of privacy preserving. This paper presents Dream Net, a privacy preserving continual learning model for face emotion recognition. Using a pseudo-rehearsal approach, this model alleviates catastrophic forgetting by capturing the mapping function of a trained network without storing examples of the learned knowledge. We evaluated Dream Net on the Fer-2013 database and obtained an average accuracy of 45% ± 2 at the end of incremental learning of all classes compare to 16% ± 0 without any continual learning model.
持续学习是人工智能面临的一个越来越大的挑战。在过去几年开发的减轻灾难性遗忘的算法中,针对面部情绪识别的研究很少。与此同时,情感识别领域提出了隐私保护的伦理问题。提出了一种保护隐私的人脸情感识别连续学习模型——梦网。该模型采用伪排练方法,通过捕获训练网络的映射函数而不存储所学知识的示例来减轻灾难性遗忘。我们在2013年4月的数据库中评估了Dream Net,在所有类别的增量学习结束时,平均准确率为45%±2,而没有任何持续学习模型的平均准确率为16%±0。
{"title":"Dream Net: a privacy preserving continual leaming model for face emotion recognition","authors":"M. Mainsant, M. Solinas, M. Reyboz, C. Godin, M. Mermillod","doi":"10.1109/aciiw52867.2021.9666338","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666338","url":null,"abstract":"Continual learning is a growing challenge of artificial intelligence. Among algorithms alleviating catastrophic forgetting that have been developed in the past years, only few studies were focused on face emotion recognition. In parallel, the field of emotion recognition raised the ethical issue of privacy preserving. This paper presents Dream Net, a privacy preserving continual learning model for face emotion recognition. Using a pseudo-rehearsal approach, this model alleviates catastrophic forgetting by capturing the mapping function of a trained network without storing examples of the learned knowledge. We evaluated Dream Net on the Fer-2013 database and obtained an average accuracy of 45% ± 2 at the end of incremental learning of all classes compare to 16% ± 0 without any continual learning model.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121695869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Emotion Recognition In Emergency Call Centers: The challenge of real-life emotions 紧急呼叫中心的情绪识别:现实生活中情绪的挑战
Théo Deschamps-Berger
Detected emotional states of speakers are a key component of constructive social relationships but also of efficiency for capturing the degree of emergency. This paper provides an overview of my doctoral project that focuses on bimodal emotion recognition in an emergency call center with deep end-to-end learning techniques using the most advanced approaches such as transformer and zero-shot learning. In this work, we will first propose a supervised classification system for bimodal emotion recognition (paralinguistic and linguistic). Then, we will investigate an unsupervised system as a complement to the previous one in order to deal with “unseen” emotions and mixtures of real-life emotions. Our previous studies mainly explored the acoustic modality of speech emotion recognition (SER), we achieved close to the state-of-the-art results on the improvised part of the well-known database IEMOCAP and we applied our approach to a French emergency database CEMO collected in a previous project. In my thesis, new real recordings in an emergency call center will be collected. The main research topics of my thesis are: Emotional representation and annotation; Speech emotion recognition and ethical implications; Evaluation and real-life trials.
侦测说话者的情绪状态是建设性社会关系的关键组成部分,也是捕捉紧急程度的效率。本文概述了我的博士项目,该项目侧重于紧急呼叫中心的双峰情绪识别,采用最先进的方法,如变压器和零学习,采用深度端到端学习技术。在这项工作中,我们将首先提出一个用于双峰情感识别(副语言和语言)的监督分类系统。然后,我们将研究一个无监督系统,作为前一个系统的补充,以处理“看不见的”情绪和现实生活中情绪的混合。我们之前的研究主要是探索语音情感识别(SER)的声学模态,我们在知名数据库IEMOCAP的临时部分上取得了接近最先进的结果,我们将我们的方法应用于以前项目中收集的法国应急数据库CEMO。在我的论文中,我将收集一个紧急呼叫中心的真实录音。本文的主要研究课题是:情感表征与标注;语音情感识别及其伦理意义评估和现实生活中的试验。
{"title":"Emotion Recognition In Emergency Call Centers: The challenge of real-life emotions","authors":"Théo Deschamps-Berger","doi":"10.1109/aciiw52867.2021.9666308","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666308","url":null,"abstract":"Detected emotional states of speakers are a key component of constructive social relationships but also of efficiency for capturing the degree of emergency. This paper provides an overview of my doctoral project that focuses on bimodal emotion recognition in an emergency call center with deep end-to-end learning techniques using the most advanced approaches such as transformer and zero-shot learning. In this work, we will first propose a supervised classification system for bimodal emotion recognition (paralinguistic and linguistic). Then, we will investigate an unsupervised system as a complement to the previous one in order to deal with “unseen” emotions and mixtures of real-life emotions. Our previous studies mainly explored the acoustic modality of speech emotion recognition (SER), we achieved close to the state-of-the-art results on the improvised part of the well-known database IEMOCAP and we applied our approach to a French emergency database CEMO collected in a previous project. In my thesis, new real recordings in an emergency call center will be collected. The main research topics of my thesis are: Emotional representation and annotation; Speech emotion recognition and ethical implications; Evaluation and real-life trials.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127616898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
In-Corpo-Real Robot-Dreams: Empathy, Skin, and Boundaries 真人机器人梦:移情、皮肤和边界
Dominika Lisy
This document is part of the submission for the doctoral consortium on affective computing and outlines motivation, theoretical background, and my research plan for the PhD project on empathy and social robots. My project idea can be divided into two parts where the first is focusing on theoretical analyses of empathy through binary conceptualisations and re-configuring empathic processes for human-robot-interaction (HRI). I will be drawing from feminist philosophy and empirical work studying signal processing from measurements on the skin and in machines in order to build a model for empathy as a process of crossing boundaries. In the second part I plan to consider implementations of these theoretical ideas in the design of empathic robots. The first part is aiming to understand and dissolve conceptual boundaries whereas the second part is re-establishing material and conceptual boundaries in order to contribute to ethical affective robot design.
这份文件是提交给情感计算博士联盟的一部分,概述了动机,理论背景,以及我对同理心和社交机器人博士项目的研究计划。我的项目想法可以分为两部分,第一部分是通过二元概念化和重新配置人机交互(HRI)的共情过程来关注共情的理论分析。我将借鉴女权主义哲学和实证工作,从皮肤和机器上的测量中研究信号处理,以建立一个同理心的模型,作为一个跨越边界的过程。在第二部分中,我计划考虑这些理论思想在移情机器人设计中的实现。第一部分旨在理解和消解概念边界,而第二部分则是重新建立材料和概念边界,以促进伦理情感机器人设计。
{"title":"In-Corpo-Real Robot-Dreams: Empathy, Skin, and Boundaries","authors":"Dominika Lisy","doi":"10.1109/aciiw52867.2021.9666373","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666373","url":null,"abstract":"This document is part of the submission for the doctoral consortium on affective computing and outlines motivation, theoretical background, and my research plan for the PhD project on empathy and social robots. My project idea can be divided into two parts where the first is focusing on theoretical analyses of empathy through binary conceptualisations and re-configuring empathic processes for human-robot-interaction (HRI). I will be drawing from feminist philosophy and empirical work studying signal processing from measurements on the skin and in machines in order to build a model for empathy as a process of crossing boundaries. In the second part I plan to consider implementations of these theoretical ideas in the design of empathic robots. The first part is aiming to understand and dissolve conceptual boundaries whereas the second part is re-establishing material and conceptual boundaries in order to contribute to ethical affective robot design.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115411347","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Convolutional Neural Network Model for Protective Behavior Detection based on Body Movement Data 基于身体运动数据的多模态卷积神经网络保护行为检测模型
Kim Ngan Phan, Soohyung Kim, Hyung-Jeong Yang, Gueesang Lee
Chronic pain treatment is a significant challenge in the healthcare industry. Physiotherapists tailor physical activity to a patient's activity based on their expression in protective behavior through pain recognition and find the special equipment to help them perform the necessary tasks. The technology can detect and assess pain behavior that could support the delivery of personalized therapies in the long-term and self-directed management of the condition to improve engagement in valued everyday activities. In this paper, we present an approach for task 1 of the Affective Movement Recognition (AffectMove) Challenge in 2021. Our proposed approach using deep learning helps detect persistent protective behavior present or absent during exercise in a person with chronic pain, based on the full-body joint position and back muscle activity of EmoPain challenge 2021 dataset. We employ convolutional neural networks by stacking residual blocks for the multimodal model. Moreover, we suggest new feature groups as additional inputs that help to increase performance for protective behavior. The proposed approach achieves an F1 score of 78.56% on validation set and 59.11% on test set. The proposed approach also outperforms previous baselines in detecting protective behavior from the EmoPain dataset.
慢性疼痛治疗是医疗保健行业的一个重大挑战。物理治疗师根据患者通过疼痛识别而表现出的保护行为,为患者量身定制身体活动,并找到帮助他们完成必要任务的特殊设备。该技术可以检测和评估疼痛行为,从而在长期和自我指导的病情管理中支持个性化治疗的提供,以提高对有价值的日常活动的参与。在本文中,我们提出了2021年情感运动识别(AffectMove)挑战任务1的方法。基于EmoPain挑战2021数据集的全身关节位置和背部肌肉活动,我们提出的使用深度学习的方法有助于检测慢性疼痛患者在运动期间存在或不存在的持续保护行为。对于多模态模型,我们通过堆叠残差块来使用卷积神经网络。此外,我们建议新的功能组作为额外的输入,以帮助提高保护行为的性能。该方法在验证集上的F1得分为78.56%,在测试集上的F1得分为59.11%。该方法在检测EmoPain数据集的保护行为方面也优于以前的基线。
{"title":"Multimodal Convolutional Neural Network Model for Protective Behavior Detection based on Body Movement Data","authors":"Kim Ngan Phan, Soohyung Kim, Hyung-Jeong Yang, Gueesang Lee","doi":"10.1109/aciiw52867.2021.9666290","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666290","url":null,"abstract":"Chronic pain treatment is a significant challenge in the healthcare industry. Physiotherapists tailor physical activity to a patient's activity based on their expression in protective behavior through pain recognition and find the special equipment to help them perform the necessary tasks. The technology can detect and assess pain behavior that could support the delivery of personalized therapies in the long-term and self-directed management of the condition to improve engagement in valued everyday activities. In this paper, we present an approach for task 1 of the Affective Movement Recognition (AffectMove) Challenge in 2021. Our proposed approach using deep learning helps detect persistent protective behavior present or absent during exercise in a person with chronic pain, based on the full-body joint position and back muscle activity of EmoPain challenge 2021 dataset. We employ convolutional neural networks by stacking residual blocks for the multimodal model. Moreover, we suggest new feature groups as additional inputs that help to increase performance for protective behavior. The proposed approach achieves an F1 score of 78.56% on validation set and 59.11% on test set. The proposed approach also outperforms previous baselines in detecting protective behavior from the EmoPain dataset.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"174 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114379675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
The AffectMove 2021 Challenge - Affect Recognition from Naturalistic Movement Data AffectMove 2021挑战-从自然运动数据中进行情感识别
Temitayo A. Olugbade, R. Sagoleo, Simone Ghisio, Nicolas E. Gold, A. Williams, B. Gelder, A. Camurri, G. Volpe, N. Bianchi-Berthouze
We ran the first Affective Movement Recognition (AffectMove) challenge that brings together datasets of affective bodily behaviour across different real-life applications to foster work in this area. Research on automatic detection of naturalistic affective body expressions is still lagging behind detection based on other modalities whereas movement behaviour modelling is a very interesting and very relevant research problem for the affective computing community. The AffectMove challenge aimed to take advantage of existing body movement datasets to address key research problems of automatic recognition of naturalistic and complex affective behaviour from this type of data. Participating teams competed to solve at least one of three tasks based on datasets of different sensors types and real-life problems: multimodal EmoPain dataset for chronic pain physical rehabilitation context, weDraw-l Movement dataset for maths problem solving settings, and multimodal Unige-Maastricht Dance dataset. To foster work across datasets, we also challenged participants to take advantage of the data across datasets to improve performances and also test the generalization of their approach across different applications.
我们进行了第一次情感动作识别(AffectMove)挑战,将不同现实生活应用程序中的情感身体行为数据集汇集在一起,以促进这一领域的工作。自然情感身体表情的自动检测研究仍然落后于基于其他模式的检测,而运动行为建模是情感计算界一个非常有趣且非常相关的研究问题。AffectMove挑战旨在利用现有的身体运动数据集来解决从这类数据中自动识别自然和复杂情感行为的关键研究问题。参赛团队根据不同传感器类型和现实问题的数据集,竞争解决至少三个任务中的一个:用于慢性疼痛物理康复背景的多模态EmoPain数据集,用于数学问题解决设置的weDraw-l运动数据集,以及多模态Unige-Maastricht舞蹈数据集。为了促进跨数据集的工作,我们还要求参与者利用跨数据集的数据来提高性能,并测试他们的方法在不同应用程序中的泛化性。
{"title":"The AffectMove 2021 Challenge - Affect Recognition from Naturalistic Movement Data","authors":"Temitayo A. Olugbade, R. Sagoleo, Simone Ghisio, Nicolas E. Gold, A. Williams, B. Gelder, A. Camurri, G. Volpe, N. Bianchi-Berthouze","doi":"10.1109/aciiw52867.2021.9666322","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666322","url":null,"abstract":"We ran the first Affective Movement Recognition (AffectMove) challenge that brings together datasets of affective bodily behaviour across different real-life applications to foster work in this area. Research on automatic detection of naturalistic affective body expressions is still lagging behind detection based on other modalities whereas movement behaviour modelling is a very interesting and very relevant research problem for the affective computing community. The AffectMove challenge aimed to take advantage of existing body movement datasets to address key research problems of automatic recognition of naturalistic and complex affective behaviour from this type of data. Participating teams competed to solve at least one of three tasks based on datasets of different sensors types and real-life problems: multimodal EmoPain dataset for chronic pain physical rehabilitation context, weDraw-l Movement dataset for maths problem solving settings, and multimodal Unige-Maastricht Dance dataset. To foster work across datasets, we also challenged participants to take advantage of the data across datasets to improve performances and also test the generalization of their approach across different applications.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124509645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Temporal based Emotion Recognition inspired by Activity Recognition models 受活动识别模型启发的基于时间的情绪识别
Balaganesh Mohan, Mirela C. Popa
Affective computing is a subset of the larger field of human-computer interaction, having important connections with cognitive processes, influencing the learning process, decision-making and perception. Out of the multiple means of communication, facial expressions are one of the most widely accepted channels for emotion modulation, receiving an increased attention during the last few years. An important aspect, contributing to their recognition success, concerns modeling the temporal dimension. Therefore, this paper aims to investigate the applicability of current state-of-the-art action recognition techniques to the human emotion recognition task. In particular, two different architectures were investigated, a CNN-based model, named Temporal Shift Module (TSM) that can learn spatiotemporal features in 3D data with the computational complexity of a 2D CNN and a video based vision transformer, employing spatio-temporal self attention. The models were trained and tested on the CREMA-D dataset, demonstrating state-of-the-art performance, with a mean class accuracy of 82% and 77% respectively, while outperforming best previous approaches by at least 3.5%.
情感计算是更大的人机交互领域的一个子集,与认知过程有着重要的联系,影响着学习过程、决策和感知。在多种交流方式中,面部表情是最被广泛接受的情绪调节渠道之一,在过去几年中受到越来越多的关注。一个重要的方面,有助于他们的识别成功,涉及建模的时间维度。因此,本文旨在研究当前最先进的动作识别技术在人类情感识别任务中的适用性。特别地,研究了两种不同的架构,一种是基于CNN的模型,称为时间移位模块(TSM),它可以以2D CNN的计算复杂度学习3D数据中的时空特征,另一种是基于视频的视觉转换器,利用时空自注意。这些模型在CREMA-D数据集上进行了训练和测试,展示了最先进的性能,平均分类准确率分别为82%和77%,而比以前的最佳方法至少高出3.5%。
{"title":"Temporal based Emotion Recognition inspired by Activity Recognition models","authors":"Balaganesh Mohan, Mirela C. Popa","doi":"10.1109/aciiw52867.2021.9666356","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666356","url":null,"abstract":"Affective computing is a subset of the larger field of human-computer interaction, having important connections with cognitive processes, influencing the learning process, decision-making and perception. Out of the multiple means of communication, facial expressions are one of the most widely accepted channels for emotion modulation, receiving an increased attention during the last few years. An important aspect, contributing to their recognition success, concerns modeling the temporal dimension. Therefore, this paper aims to investigate the applicability of current state-of-the-art action recognition techniques to the human emotion recognition task. In particular, two different architectures were investigated, a CNN-based model, named Temporal Shift Module (TSM) that can learn spatiotemporal features in 3D data with the computational complexity of a 2D CNN and a video based vision transformer, employing spatio-temporal self attention. The models were trained and tested on the CREMA-D dataset, demonstrating state-of-the-art performance, with a mean class accuracy of 82% and 77% respectively, while outperforming best previous approaches by at least 3.5%.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127367991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Task-based Classification of Reflective Thinking Using Mixture of Classifiers 混合分类器在反思性思维任务分类中的应用
Saandeep Aathreya, Liza Jivnani, Shivam Srivastava, Saurabh Hinduja, Shaun J. Canavan
This paper studies the problem of Reflective Thinking in children during mathematics related problem solving activities. We present our approach in solving task 2 of the AffectMove challenge, which is Reflective Thinking Detection (RTD) while solving a mathematical activity. We utilize temporal data consisting of 3D joint positions, to construct a series of classifiers that can predict whether the subject appeared to possess reflective thinking ability during the given instance. We tackle the challenge of highly imbalanced data by incorporating and analyzing several meaningful data augmentation techniques and handcrafted features. We then feed different features through a number of machine learning classifiers and select the best performing model. We evaluate our predictions on multiple metrics including accuracy, F1 score, and MCC to work towards a generalized solution for the real-world dataset.
本文研究了儿童在数学相关问题解决活动中的反思性思维问题。我们提出了解决AffectMove挑战任务2的方法,即在解决数学活动时进行反思性思维检测(RTD)。我们利用三维关节位置组成的时间数据来构建一系列分类器,这些分类器可以预测受试者在给定实例中是否具有反思性思维能力。我们通过整合和分析几种有意义的数据增强技术和手工制作的特征来解决高度不平衡数据的挑战。然后,我们通过许多机器学习分类器输入不同的特征,并选择表现最好的模型。我们在多个指标上评估我们的预测,包括准确性、F1分数和MCC,以努力为现实世界的数据集提供一个通用的解决方案。
{"title":"Task-based Classification of Reflective Thinking Using Mixture of Classifiers","authors":"Saandeep Aathreya, Liza Jivnani, Shivam Srivastava, Saurabh Hinduja, Shaun J. Canavan","doi":"10.1109/aciiw52867.2021.9666442","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666442","url":null,"abstract":"This paper studies the problem of Reflective Thinking in children during mathematics related problem solving activities. We present our approach in solving task 2 of the AffectMove challenge, which is Reflective Thinking Detection (RTD) while solving a mathematical activity. We utilize temporal data consisting of 3D joint positions, to construct a series of classifiers that can predict whether the subject appeared to possess reflective thinking ability during the given instance. We tackle the challenge of highly imbalanced data by incorporating and analyzing several meaningful data augmentation techniques and handcrafted features. We then feed different features through a number of machine learning classifiers and select the best performing model. We evaluate our predictions on multiple metrics including accuracy, F1 score, and MCC to work towards a generalized solution for the real-world dataset.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129986446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modeling Emotions as Latent Representations of Appraisals 将情绪建模为评价的潜在表征
Marios A. Fanourakis, Rayan Elalamy, G. Chanel
Emotion recognition is usually achieved by collecting features (physiological signals, events, facial expressions, etc.) to predict an emotional ground truth. This ground truth, however, is subjective and not always an accurate representation of the emotional state of the subject. In this paper, we show that emotion can be learned in the latent space of machine learning methods without relying on an emotional ground truth. Our data consists of physiological measurements during video gameplay, game events, and subjective rankings of game events for the validation of our hypothesis. By calculating the Kendall ${tau}$ rank correlation between the subjective game event rankings and both the rankings derived from Canonical Correlation Analysis (CCA) and a simple neural network, we show that the latent space of these models is correlated with the subjective rankings even though they were not part of the training data.
情感识别通常是通过收集特征(生理信号、事件、面部表情等)来预测情感基础真相来实现的。然而,这个基本真理是主观的,并不总是对主体情感状态的准确反映。在本文中,我们证明了情感可以在机器学习方法的潜在空间中学习,而不依赖于情感基础真理。我们的数据包括视频游戏过程中的生理测量、游戏事件以及验证我们假设的游戏事件的主观排名。通过计算主观比赛事件排名与典型相关分析(CCA)和简单神经网络得出的排名之间的Kendall ${tau}$排名相关性,我们发现这些模型的潜在空间与主观排名相关,即使它们不是训练数据的一部分。
{"title":"Modeling Emotions as Latent Representations of Appraisals","authors":"Marios A. Fanourakis, Rayan Elalamy, G. Chanel","doi":"10.1109/aciiw52867.2021.9666198","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666198","url":null,"abstract":"Emotion recognition is usually achieved by collecting features (physiological signals, events, facial expressions, etc.) to predict an emotional ground truth. This ground truth, however, is subjective and not always an accurate representation of the emotional state of the subject. In this paper, we show that emotion can be learned in the latent space of machine learning methods without relying on an emotional ground truth. Our data consists of physiological measurements during video gameplay, game events, and subjective rankings of game events for the validation of our hypothesis. By calculating the Kendall ${tau}$ rank correlation between the subjective game event rankings and both the rankings derived from Canonical Correlation Analysis (CCA) and a simple neural network, we show that the latent space of these models is correlated with the subjective rankings even though they were not part of the training data.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125296864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simulating Fear as Anticipation of Temporal Differences: An experimental investigation
L. Dai, J. Broekens
Humans use emotional expressions to communicate appraisals. Humans also use emotions in evaluating how they are doing compared to their current goals and desires. The Temporal Difference Reinforcement Learning (TDRL) Theory of Emotion proposes a structure for agents to simulate appropriate emotions during the learning process. In previous work, simulations have shown to reproduce plausible emotion dynamics. In this paper we examine the plausibility and intepretability of TDRL-simulated fear, when expressed by the agent. We presented different TDRL-based fear simulation methods to participants ${left(n=237right)}$ in an online study. Each method used a different action selection protocol for the agent's model-based anticipation process. Results suggest that an ${in}$-greedy fear policy ${left(in=0.1right)}$ combined with a long anticipation horizon provides a plausible fear estimation. This is, to our knowledge, the first experimental evidence detailing some of the predictions of the TDRL Theory of Emotion. Our results are of interest to the design of agent learning methods that are transparent to the user.
人类用情感表达来交流评价。与当前的目标和愿望相比,人类也会用情绪来评估自己的表现。情绪的时间差异强化学习理论(TDRL)提出了一种智能体在学习过程中模拟适当情绪的结构。在之前的工作中,模拟已经显示出再现可信的情绪动态。在这篇论文中,我们研究了由agent表达的tdrl模拟恐惧的合理性和可解释性。在一项在线研究中,我们向参与者${left(n=237right)}$提供了不同的基于tdrl的恐惧模拟方法。每种方法对智能体基于模型的预测过程使用不同的动作选择协议。结果表明,${in}$贪婪的恐惧策略${左(in=0.1右)}$结合较长的预期视界提供了合理的恐惧估计。据我们所知,这是第一个实验证据,详细说明了TDRL情绪理论的一些预测。我们的结果对设计对用户透明的智能体学习方法很有意义。
{"title":"Simulating Fear as Anticipation of Temporal Differences: An experimental investigation","authors":"L. Dai, J. Broekens","doi":"10.1109/aciiw52867.2021.9666259","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666259","url":null,"abstract":"Humans use emotional expressions to communicate appraisals. Humans also use emotions in evaluating how they are doing compared to their current goals and desires. The Temporal Difference Reinforcement Learning (TDRL) Theory of Emotion proposes a structure for agents to simulate appropriate emotions during the learning process. In previous work, simulations have shown to reproduce plausible emotion dynamics. In this paper we examine the plausibility and intepretability of TDRL-simulated fear, when expressed by the agent. We presented different TDRL-based fear simulation methods to participants ${left(n=237right)}$ in an online study. Each method used a different action selection protocol for the agent's model-based anticipation process. Results suggest that an ${in}$-greedy fear policy ${left(in=0.1right)}$ combined with a long anticipation horizon provides a plausible fear estimation. This is, to our knowledge, the first experimental evidence detailing some of the predictions of the TDRL Theory of Emotion. Our results are of interest to the design of agent learning methods that are transparent to the user.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123074635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1