首页 > 最新文献

2015 International Conference on Affective Computing and Intelligent Interaction (ACII)最新文献

英文 中文
Cognitive state measurement from eye gaze analysis in an intelligent virtual reality driving system for autism intervention 自闭症干预智能虚拟现实驾驶系统中眼注视分析的认知状态测量
Lian Zhang, Joshua W. Wade, A. Swanson, A. Weitlauf, Z. Warren, N. Sarkar
Autism Spectrum Disorder (ASD) is a group of neurodevelopmental disabilities with a high prevalence rate. While much research has focused on improving social communication deficits in ASD populations, less emphasis has been devoted to improving skills relevant for adult independent living, such as driving. In this paper, a novel virtual reality (VR)-based driving system with different difficulty levels of tasks is presented to train and improve driving skills of teenagers with ASD. The goal of this paper is to measure the cognitive load experienced by an individual with ASD while he is driving in the VR-based driving system. Several eye gaze features are identified that varied with cognitive load in an experiment participated by 12 teenagers with ASD. Several machine learning methods were compared and the ability of these methods to accurately measure cognitive load was validated with respect to the subjective rating of a therapist. Results will be used to build models in an intelligent VR-based driving system that can sense a participant's real-time cognitive load and offer driving tasks at an appropriate difficulty level in order to maximize the participant's long-term performance.
自闭症谱系障碍(Autism Spectrum Disorder, ASD)是一组患病率较高的神经发育障碍。虽然很多研究都集中在改善ASD人群的社会沟通缺陷上,但很少有人重视提高与成人独立生活相关的技能,比如驾驶。本文提出了一种新的基于虚拟现实(VR)的驾驶系统,通过不同难度的任务来训练和提高青少年ASD的驾驶技能。本文的目的是测量ASD患者在基于vr的驾驶系统中驾驶时所经历的认知负荷。在一项由12名自闭症青少年参与的实验中,发现了几种眼睛注视特征随着认知负荷的变化而变化。比较了几种机器学习方法,并根据治疗师的主观评分验证了这些方法准确测量认知负荷的能力。研究结果将用于在基于vr的智能驾驶系统中建立模型,该系统可以感知参与者的实时认知负荷,并提供适当难度的驾驶任务,以最大限度地提高参与者的长期表现。
{"title":"Cognitive state measurement from eye gaze analysis in an intelligent virtual reality driving system for autism intervention","authors":"Lian Zhang, Joshua W. Wade, A. Swanson, A. Weitlauf, Z. Warren, N. Sarkar","doi":"10.1109/ACII.2015.7344621","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344621","url":null,"abstract":"Autism Spectrum Disorder (ASD) is a group of neurodevelopmental disabilities with a high prevalence rate. While much research has focused on improving social communication deficits in ASD populations, less emphasis has been devoted to improving skills relevant for adult independent living, such as driving. In this paper, a novel virtual reality (VR)-based driving system with different difficulty levels of tasks is presented to train and improve driving skills of teenagers with ASD. The goal of this paper is to measure the cognitive load experienced by an individual with ASD while he is driving in the VR-based driving system. Several eye gaze features are identified that varied with cognitive load in an experiment participated by 12 teenagers with ASD. Several machine learning methods were compared and the ability of these methods to accurately measure cognitive load was validated with respect to the subjective rating of a therapist. Results will be used to build models in an intelligent VR-based driving system that can sense a participant's real-time cognitive load and offer driving tasks at an appropriate difficulty level in order to maximize the participant's long-term performance.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"51 1","pages":"532-538"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79395135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
From simulated speech to natural speech, what are the robust features for emotion recognition? 从模拟语音到自然语音,情感识别的鲁棒性特征是什么?
Ya Li, Linlin Chao, Yazhu Liu, Wei Bao, J. Tao
The earliest research on emotion recognition starts with simulated/acted stereotypical emotional corpus, and then extends to elicited corpus. Recently, the demanding for real application forces the research shift to natural and spontaneous corpus. Previous research shows that accuracies of emotion recognition are gradual decline from simulated speech, to elicited and totally natural speech. This paper aims to investigate the effects of the common utilized spectral, prosody and voice quality features in emotion recognition with the three types of corpus, and finds out the robust feature for emotion recognition with natural speech. Emotion recognition by several common machine learning methods are carried out and thoroughly compared. Three feature selection methods are performed to find the robust features. The results on six common used corpora confirm that recognition accuracies decrease when the corpus changing from simulated to natural corpus. In addition, prosody and voice quality features are robust for emotion recognition on simulated corpus, while spectral feature is robust in elicited and natural corpus.
情绪识别的研究最早是从模拟/行为的刻板印象情绪语料库开始的,然后扩展到引出语料库。近年来,实际应用的需求迫使研究转向自然自发语料库。以往的研究表明,情绪识别的准确性从模拟语音到引出的完全自然的语音是逐渐下降的。本文旨在利用这三种语料库研究常用的频谱、韵律和音质特征在情感识别中的作用,找出用于自然语音情感识别的鲁棒性特征。对几种常用的机器学习方法进行了情感识别,并进行了比较。采用三种特征选择方法寻找鲁棒特征。在6种常用语料库上的实验结果表明,当语料库由模拟语料库转换为自然语料库时,识别准确率下降。此外,韵律和语音质量特征对模拟语料库的情感识别具有鲁棒性,而谱特征对自然语料库和人工语料库的情感识别具有鲁棒性。
{"title":"From simulated speech to natural speech, what are the robust features for emotion recognition?","authors":"Ya Li, Linlin Chao, Yazhu Liu, Wei Bao, J. Tao","doi":"10.1109/ACII.2015.7344597","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344597","url":null,"abstract":"The earliest research on emotion recognition starts with simulated/acted stereotypical emotional corpus, and then extends to elicited corpus. Recently, the demanding for real application forces the research shift to natural and spontaneous corpus. Previous research shows that accuracies of emotion recognition are gradual decline from simulated speech, to elicited and totally natural speech. This paper aims to investigate the effects of the common utilized spectral, prosody and voice quality features in emotion recognition with the three types of corpus, and finds out the robust feature for emotion recognition with natural speech. Emotion recognition by several common machine learning methods are carried out and thoroughly compared. Three feature selection methods are performed to find the robust features. The results on six common used corpora confirm that recognition accuracies decrease when the corpus changing from simulated to natural corpus. In addition, prosody and voice quality features are robust for emotion recognition on simulated corpus, while spectral feature is robust in elicited and natural corpus.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"64 1","pages":"368-373"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79580455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Chinese microblog sentiment classification based on convolution neural network with content extension method 基于内容扩展卷积神经网络的中文微博情感分类
Xiao Sun, Fei Gao, Chengcheng Li, F. Ren
Related research for sentiment analysis on Chinese microblog is aiming at analyzing the emotion of posters. This paper presents a content extension method that combines post with its' comments into a microblog conversation for sentiment analysis. A new convolutional auto encoder which can extract contextual sentiment information from microblog conversation of the post is proposed. Furthermore, a DBN model, which is composed by several layers of RBM(Restricted Boltzmann Machine) stacked together, is implemented to extract some higher level feature for short text of a post. These RBM layers can encoder observed short text to learn hidden structures or semantics information for better feature representation. A ClassRBM (Classification RBM) layer, which is stacked on top of RBM layers, is adapted to achieve the final sentiment classification. The experiment results demonstrate that, with proper structure and parameter, the performance of the proposed deep learning method on sentiment classification is better than state-of-the-art surface learning models such as SVM or NB, which also proves that DBN is suitable for short-length document classification with the proposed feature dimensionality extension method.
中国微博情感分析的相关研究旨在分析海报的情感。本文提出了一种内容扩展方法,将微博帖子及其评论结合到微博会话中进行情感分析。提出了一种新的卷积自编码器,可以从微博帖子的会话中提取上下文情感信息。在此基础上,实现了一种由多层受限玻尔兹曼机(Restricted Boltzmann Machine, RBM)叠加而成的DBN模型,用于文章短文本的高级特征提取。这些RBM层可以对观察到的短文本进行编码器,以学习隐藏的结构或语义信息,以便更好地表示特征。一个ClassRBM (Classification RBM)层,堆叠在RBM层之上,用于实现最终的情感分类。实验结果表明,在适当的结构和参数下,所提出的深度学习方法在情感分类上的性能优于当前最先进的表面学习模型(如SVM或NB),这也证明了DBN适用于使用所提出的特征维数扩展方法进行短长度文档分类。
{"title":"Chinese microblog sentiment classification based on convolution neural network with content extension method","authors":"Xiao Sun, Fei Gao, Chengcheng Li, F. Ren","doi":"10.1109/ACII.2015.7344603","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344603","url":null,"abstract":"Related research for sentiment analysis on Chinese microblog is aiming at analyzing the emotion of posters. This paper presents a content extension method that combines post with its' comments into a microblog conversation for sentiment analysis. A new convolutional auto encoder which can extract contextual sentiment information from microblog conversation of the post is proposed. Furthermore, a DBN model, which is composed by several layers of RBM(Restricted Boltzmann Machine) stacked together, is implemented to extract some higher level feature for short text of a post. These RBM layers can encoder observed short text to learn hidden structures or semantics information for better feature representation. A ClassRBM (Classification RBM) layer, which is stacked on top of RBM layers, is adapted to achieve the final sentiment classification. The experiment results demonstrate that, with proper structure and parameter, the performance of the proposed deep learning method on sentiment classification is better than state-of-the-art surface learning models such as SVM or NB, which also proves that DBN is suitable for short-length document classification with the proposed feature dimensionality extension method.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"38 1","pages":"408-414"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78422426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Gestural and Postural Reactions to Stressful Event: Design of a Haptic Stressful Stimulus 应激事件的手势和姿势反应:触觉应激刺激的设计
Yoren Gaffary, David Antonio Gómez Jáuregui, Jean-Claude Martin, M. Ammi
Previous studies about kinesthetic expressions of emotions are mainly based on acted expressions of affective states, which might be quite different from spontaneous expressions. In a previous study, we proposed a task to collect haptic expressions of a spontaneous stress. In this paper, we explore the effectiveness of this task to induce a spontaneous stress in two ways: a subjective feedback, and a more objective approach-avoidance behavior.
以往关于动觉情绪表达的研究主要是基于情感状态的行为表达,这可能与自发表达有很大的不同。在之前的研究中,我们提出了一个收集自发应激的触觉表达的任务。在本文中,我们探讨了这一任务的有效性,以诱导自发应激两种方式:主观反馈和更客观的方法-回避行为。
{"title":"Gestural and Postural Reactions to Stressful Event: Design of a Haptic Stressful Stimulus","authors":"Yoren Gaffary, David Antonio Gómez Jáuregui, Jean-Claude Martin, M. Ammi","doi":"10.1109/ACII.2015.7344696","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344696","url":null,"abstract":"Previous studies about kinesthetic expressions of emotions are mainly based on acted expressions of affective states, which might be quite different from spontaneous expressions. In a previous study, we proposed a task to collect haptic expressions of a spontaneous stress. In this paper, we explore the effectiveness of this task to induce a spontaneous stress in two ways: a subjective feedback, and a more objective approach-avoidance behavior.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"116 1","pages":"988-992"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76872659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Framework for combination aware AU intensity recognition 组合感知AU强度识别框架
Isabel Gonzalez, W. Verhelst, Meshia Cédric Oveneke, H. Sahli, D. Jiang
We present a framework for combination aware AU intensity recognition. It includes a feature extraction approach that can handle small head movements which does not require face alignment. A three layered structure is used for the AU classification. The first layer is dedicated to independent AU recognition, and the second layer incorporates AU combination knowledge. At a third layer, AU dynamics are handled based on variable duration semi-Markov model. The first two layers are modeled using extreme learning machines (ELMs). ELMs have equal performance to support vector machines but are computationally more efficient, and can handle multi-class classification directly. Moreover, they include feature selection via manifold regularization. We show that the proposed layered classification scheme can improve results by considering AU combinations as well as intensity recognition.
提出了一种组合感知AU强度识别框架。它包括一种特征提取方法,可以处理不需要面部对齐的小头部运动。AU分类采用三层结构。第一层用于独立AU识别,第二层包含AU组合知识。在第三层,基于可变持续时间半马尔可夫模型处理AU动态。前两层使用极限学习机(elm)建模。elm具有与支持向量机相当的性能,但计算效率更高,并且可以直接处理多类分类。此外,它们还包括通过流形正则化进行特征选择。我们表明,通过考虑AU组合和强度识别,提出的分层分类方案可以改善结果。
{"title":"Framework for combination aware AU intensity recognition","authors":"Isabel Gonzalez, W. Verhelst, Meshia Cédric Oveneke, H. Sahli, D. Jiang","doi":"10.1109/ACII.2015.7344631","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344631","url":null,"abstract":"We present a framework for combination aware AU intensity recognition. It includes a feature extraction approach that can handle small head movements which does not require face alignment. A three layered structure is used for the AU classification. The first layer is dedicated to independent AU recognition, and the second layer incorporates AU combination knowledge. At a third layer, AU dynamics are handled based on variable duration semi-Markov model. The first two layers are modeled using extreme learning machines (ELMs). ELMs have equal performance to support vector machines but are computationally more efficient, and can handle multi-class classification directly. Moreover, they include feature selection via manifold regularization. We show that the proposed layered classification scheme can improve results by considering AU combinations as well as intensity recognition.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"21 1","pages":"602-608"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77729160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Adapting sentiment analysis to face-to-face human-agent interactions: From the detection to the evaluation issues 将情感分析应用于面对面的人机交互:从检测到评估问题
Caroline Langlet, C. Clavel
This paper introduces a sentiment analysis method suitable to the human-agent and face-to-face interactions. We present the positioning of our system and its evaluation protocol according to the existing sentiment analysis literature and detail how the proposed system integrates the human-agent interaction issues. Finally, we provide an in-depth analysis of the results obtained by the evaluation, opening the discussion on the different difficulties and the remaining challenges of sentiment analysis in human-agent interactions.
本文介绍了一种适用于人机交互和面对面交互的情感分析方法。根据现有的情感分析文献,我们提出了系统的定位及其评估协议,并详细介绍了所提出的系统如何集成人机交互问题。最后,我们对评估获得的结果进行了深入分析,开始讨论人类智能体交互中情感分析的不同困难和仍然存在的挑战。
{"title":"Adapting sentiment analysis to face-to-face human-agent interactions: From the detection to the evaluation issues","authors":"Caroline Langlet, C. Clavel","doi":"10.1109/ACII.2015.7344545","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344545","url":null,"abstract":"This paper introduces a sentiment analysis method suitable to the human-agent and face-to-face interactions. We present the positioning of our system and its evaluation protocol according to the existing sentiment analysis literature and detail how the proposed system integrates the human-agent interaction issues. Finally, we provide an in-depth analysis of the results obtained by the evaluation, opening the discussion on the different difficulties and the remaining challenges of sentiment analysis in human-agent interactions.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"12 1","pages":"14-20"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81535062","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
A linear regression model to detect user emotion for touch input interactive systems 基于线性回归模型的触摸输入交互系统用户情感检测
S. Bhattacharya
Human emotion plays significant role is affecting our reasoning, learning, cognition and decision making, which in turn may affect usability of interactive systems. Detection of emotion of interactive system users is therefore important, as it can help design for improved user experience. In this work, we propose a model to detect the emotional state of the users of touch screen devices. Although a number of methods were developed to detect human emotion, those are computationally intensive and require setup cost. The model we propose aims to avoid these limitations and make the detection process viable for mobile platforms. We assume three emotional states of a user: positive, negative and neutral. The touch interaction is characterized by a set of seven features, derived from the finger strokes and taps. Our proposed model is a linear combination of these features. The model is developed and validated with empirical data involving 57 participants performing 7 touch input tasks. The validation study demonstrates a high prediction accuracy of 90.47%.
人类的情感影响着我们的推理、学习、认知和决策,进而影响交互系统的可用性。因此,检测交互系统用户的情感是很重要的,因为它可以帮助设计改进用户体验。在这项工作中,我们提出了一个模型来检测触摸屏设备用户的情绪状态。虽然已经开发了许多方法来检测人类的情绪,但这些方法都是计算密集型的,并且需要设置成本。我们提出的模型旨在避免这些限制,并使检测过程适用于移动平台。我们假设用户有三种情绪状态:积极、消极和中性。触摸交互的特点是一组七个特征,源自手指的抚摸和轻拍。我们提出的模型是这些特征的线性组合。该模型通过57名参与者执行7项触摸输入任务的经验数据进行了开发和验证。验证研究表明,该方法的预测准确率为90.47%。
{"title":"A linear regression model to detect user emotion for touch input interactive systems","authors":"S. Bhattacharya","doi":"10.1109/ACII.2015.7344693","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344693","url":null,"abstract":"Human emotion plays significant role is affecting our reasoning, learning, cognition and decision making, which in turn may affect usability of interactive systems. Detection of emotion of interactive system users is therefore important, as it can help design for improved user experience. In this work, we propose a model to detect the emotional state of the users of touch screen devices. Although a number of methods were developed to detect human emotion, those are computationally intensive and require setup cost. The model we propose aims to avoid these limitations and make the detection process viable for mobile platforms. We assume three emotional states of a user: positive, negative and neutral. The touch interaction is characterized by a set of seven features, derived from the finger strokes and taps. Our proposed model is a linear combination of these features. The model is developed and validated with empirical data involving 57 participants performing 7 touch input tasks. The validation study demonstrates a high prediction accuracy of 90.47%.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"51 1","pages":"970-975"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83189095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Detection of negative emotions in speech signals using bags-of-audio-words 利用音频-词袋检测语音信号中的消极情绪
Florian B. Pokorny, F. Graf, F. Pernkopf, Björn Schuller
Boosted by a wide potential application spectrum, emotional speech recognition, i.e., the automatic computer-aided identification of human emotional states based on speech signals, currently describes a popular field of research. However, a variety of studies especially concentrating on the recognition of negative emotions often neglected the specific requirements of real-world scenarios, for example, robustness, real-time capability, and realistic speech corpora. Motivated by these facts, a robust, low-complex classification system for the detection of negative emotions in speech signals was implemented on the basis of a spontaneous, strongly emotionally colored speech corpus. Therefore, an innovative approach in the field of emotion recognition was applied as the core of the system - the bag-of-words approach that is originally known from text and image document retrieval applications. Thorough performance evaluations were carried out and a promising recognition accuracy of 65.6 % for the 2-class paradigm negative versus non-negative emotional states attests to the potential of bags-of-words in speech emotion recognition in the wild.
由于具有广泛的潜在应用范围,情感语音识别,即基于语音信号的人类情绪状态的计算机自动辅助识别,目前是一个热门的研究领域。然而,各种专注于负面情绪识别的研究往往忽略了现实场景的具体要求,如鲁棒性、实时性和真实的语音语料库。基于这些事实,基于一个自发的、具有强烈情感色彩的语音语料库,实现了一个鲁棒的、低复杂度的语音信号负面情绪检测分类系统。因此,我们采用了情感识别领域的一种创新方法作为系统的核心——最初在文本和图像文档检索应用中已知的词袋方法。对两类范式的消极情绪状态和非消极情绪状态进行了全面的性能评估,其识别准确率为65.6%,证明了词袋在野外语音情绪识别中的潜力。
{"title":"Detection of negative emotions in speech signals using bags-of-audio-words","authors":"Florian B. Pokorny, F. Graf, F. Pernkopf, Björn Schuller","doi":"10.1109/ACII.2015.7344678","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344678","url":null,"abstract":"Boosted by a wide potential application spectrum, emotional speech recognition, i.e., the automatic computer-aided identification of human emotional states based on speech signals, currently describes a popular field of research. However, a variety of studies especially concentrating on the recognition of negative emotions often neglected the specific requirements of real-world scenarios, for example, robustness, real-time capability, and realistic speech corpora. Motivated by these facts, a robust, low-complex classification system for the detection of negative emotions in speech signals was implemented on the basis of a spontaneous, strongly emotionally colored speech corpus. Therefore, an innovative approach in the field of emotion recognition was applied as the core of the system - the bag-of-words approach that is originally known from text and image document retrieval applications. Thorough performance evaluations were carried out and a promising recognition accuracy of 65.6 % for the 2-class paradigm negative versus non-negative emotional states attests to the potential of bags-of-words in speech emotion recognition in the wild.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"4 1","pages":"879-884"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88698604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Superframe segmentation based on content-motion correspondence for social video summarization 基于内容-运动对应的超帧分割用于社交视频摘要
Tao Zhuo, Peng Zhang, Kangli Chen, Yanning Zhang
The goal of video summarization is to turn large volume of video data into a compact visual summary that can be easily interpreted by users in a while. Existing summarization strategies employed the point based feature correspondence for the superframe segmentation. Unfortunately, the information carried by those sparse points is far from sufficiency and stability to describe the change of interesting regions of each frame. Therefore, in order to overcome the limitations of point feature, we propose a region correspondence based superframe segmentation to achieve more effective video summarization. Instead of utilizing the motion of feature points, we calculate the similarity of content-motion to obtain the strength of change between the consecutive frames. With the help of circulant structure kernel, the proposed method is able to perform more accurate motion estimation efficiently. Experimental testing on the videos from benchmark database has demonstrate the effectiveness of the proposed method.
视频摘要的目标是将大量的视频数据转化为用户可以在短时间内理解的简洁的视觉摘要。现有的摘要策略采用基于点的特征对应进行超帧分割。遗憾的是,这些稀疏点所携带的信息远远不够充分和稳定,无法描述每帧感兴趣区域的变化。因此,为了克服点特征的局限性,我们提出了一种基于区域对应的超帧分割方法来实现更有效的视频摘要。我们不是利用特征点的运动,而是计算内容运动的相似度来获得连续帧之间的变化强度。在循环结构核的帮助下,该方法能够有效地进行更精确的运动估计。通过对基准数据库视频的实验测试,证明了该方法的有效性。
{"title":"Superframe segmentation based on content-motion correspondence for social video summarization","authors":"Tao Zhuo, Peng Zhang, Kangli Chen, Yanning Zhang","doi":"10.1109/ACII.2015.7344674","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344674","url":null,"abstract":"The goal of video summarization is to turn large volume of video data into a compact visual summary that can be easily interpreted by users in a while. Existing summarization strategies employed the point based feature correspondence for the superframe segmentation. Unfortunately, the information carried by those sparse points is far from sufficiency and stability to describe the change of interesting regions of each frame. Therefore, in order to overcome the limitations of point feature, we propose a region correspondence based superframe segmentation to achieve more effective video summarization. Instead of utilizing the motion of feature points, we calculate the similarity of content-motion to obtain the strength of change between the consecutive frames. With the help of circulant structure kernel, the proposed method is able to perform more accurate motion estimation efficiently. Experimental testing on the videos from benchmark database has demonstrate the effectiveness of the proposed method.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"24 1","pages":"857-862"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87790330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Building autonomous sensitive artificial listeners (Extended abstract) 构建自主敏感的人工听者(扩展摘要)
M. Schröder, Elisabetta Bevacqua, R. Cowie, F. Eyben, H. Gunes, D. Heylen, M. Maat, G. McKeown, Sathish Pammi, M. Pantic, C. Pelachaud, Björn Schuller, E. D. Sevin, M. Valstar, M. Wöllmer
This paper describes a substantial effort to build a real-time interactive multimodal dialogue system with a focus on emotional and non-verbal interaction capabilities. The work is motivated by the aim to provide technology with competences in perceiving and producing the emotional and non-verbal behaviours required to sustain a conversational dialogue. We present the Sensitive Artificial Listener (SAL) scenario as a setting which seems particularly suited for the study of emotional and non-verbal behaviour, since it requires only very limited verbal understanding on the part of the machine. This scenario allows us to concentrate on non-verbal capabilities without having to address at the same time the challenges of spoken language understanding, task modeling etc. We first summarise three prototype versions of the SAL scenario, in which the behaviour of the Sensitive Artificial Listener characters was determined by a human operator. These prototypes served the purpose of verifying the effectiveness of the SAL scenario and allowed us to collect data required for building system components for analysing and synthesising the respective behaviours. We then describe the fully autonomous integrated real-time system we created, which combines incremental analysis of user behaviour, dialogue management, and synthesis of speaker and listener behaviour of a SAL character displayed as a virtual agent. We discuss principles that should underlie the evaluation of SAL-type systems. Since the system is designed for modularity and reuse, and since it is publicly available, the SAL system has potential as a joint research tool in the affective computing research community.
本文描述了建立一个关注情感和非语言交互能力的实时交互多模态对话系统的大量工作。这项工作的动机是为技术提供感知和产生维持会话对话所需的情感和非语言行为的能力。我们提出了敏感人工听者(SAL)场景,它似乎特别适合于研究情感和非语言行为,因为它只需要机器非常有限的语言理解。这种情况使我们能够专注于非语言能力,而不必同时解决口语理解、任务建模等方面的挑战。我们首先总结了SAL场景的三个原型版本,其中敏感人工倾听者角色的行为由人类操作员决定。这些原型用于验证SAL场景的有效性,并允许我们收集构建系统组件所需的数据,以分析和综合各自的行为。然后,我们描述了我们创建的完全自主集成实时系统,该系统结合了用户行为的增量分析、对话管理以及作为虚拟代理显示的SAL角色的说话者和听者行为的综合。我们讨论的原则,应基础的评价saltype系统。由于该系统是为模块化和重用而设计的,并且它是公开可用的,因此SAL系统具有作为情感计算研究社区的联合研究工具的潜力。
{"title":"Building autonomous sensitive artificial listeners (Extended abstract)","authors":"M. Schröder, Elisabetta Bevacqua, R. Cowie, F. Eyben, H. Gunes, D. Heylen, M. Maat, G. McKeown, Sathish Pammi, M. Pantic, C. Pelachaud, Björn Schuller, E. D. Sevin, M. Valstar, M. Wöllmer","doi":"10.1109/ACII.2015.7344610","DOIUrl":"https://doi.org/10.1109/ACII.2015.7344610","url":null,"abstract":"This paper describes a substantial effort to build a real-time interactive multimodal dialogue system with a focus on emotional and non-verbal interaction capabilities. The work is motivated by the aim to provide technology with competences in perceiving and producing the emotional and non-verbal behaviours required to sustain a conversational dialogue. We present the Sensitive Artificial Listener (SAL) scenario as a setting which seems particularly suited for the study of emotional and non-verbal behaviour, since it requires only very limited verbal understanding on the part of the machine. This scenario allows us to concentrate on non-verbal capabilities without having to address at the same time the challenges of spoken language understanding, task modeling etc. We first summarise three prototype versions of the SAL scenario, in which the behaviour of the Sensitive Artificial Listener characters was determined by a human operator. These prototypes served the purpose of verifying the effectiveness of the SAL scenario and allowed us to collect data required for building system components for analysing and synthesising the respective behaviours. We then describe the fully autonomous integrated real-time system we created, which combines incremental analysis of user behaviour, dialogue management, and synthesis of speaker and listener behaviour of a SAL character displayed as a virtual agent. We discuss principles that should underlie the evaluation of SAL-type systems. Since the system is designed for modularity and reuse, and since it is publicly available, the SAL system has potential as a joint research tool in the affective computing research community.","PeriodicalId":6863,"journal":{"name":"2015 International Conference on Affective Computing and Intelligent Interaction (ACII)","volume":"282 1","pages":"456-462"},"PeriodicalIF":0.0,"publicationDate":"2015-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88057551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2015 International Conference on Affective Computing and Intelligent Interaction (ACII)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1