首页 > 最新文献

2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)最新文献

英文 中文
Emotional Paraphrasing Using Pre-trained Language Models 使用预先训练的语言模型进行情感解释
Jacky Casas, Samuel Torche, Karl Daher, E. Mugellini, Omar Abou Khaled
Emotion style transfer is a recent and challenging problem in Natural Language Processing (NLP). Transformer-based language models are becoming extremely powerful, so one wonders if it would be possible to leverage them to perform emotion style transfer. So far, previous work has not used transformer-based models for this task. To address this task, we fine-tune a GPT-2 model with corrupted emotional data. This will train the model to increase the emotional intensity of the input sentence. Coupled with a paraphrasing model, we develop a system capable of transferring an emotion into a paraphrase. We conducted a qualitative study with human judges, as well as a quantitative evaluation. Although the paraphrase metrics show poor performance compared to the state of the art, the transfer of emotion proved to be effective, especially for the emotions fear, sadness, and disgust. The perception of these emotions were improved both in the automatic and human evaluations. Such technology can significantly facilitate the automatic creation of training sentences for natural language understanding (NLU) systems, but it can also be integrated into an emotional or empathic dialogue architecture.
情感风格迁移是自然语言处理(NLP)中一个新兴的、具有挑战性的问题。基于变形器的语言模型正变得非常强大,所以人们想知道是否有可能利用它们来执行情感风格转移。到目前为止,以前的工作还没有使用基于变压器的模型来完成这项任务。为了解决这个问题,我们对带有损坏情绪数据的GPT-2模型进行了微调。这将训练模型增加输入句子的情感强度。结合释义模型,我们开发了一个能够将情感转化为释义的系统。我们进行了定性研究与人类法官,以及定量评价。尽管与目前的技术水平相比,意译指标表现不佳,但情绪转移被证明是有效的,尤其是对恐惧、悲伤和厌恶的情绪。在自动评估和人工评估中,这些情绪的感知都得到了改善。这种技术可以极大地促进自然语言理解(NLU)系统的训练句子的自动创建,但它也可以集成到情感或移情对话架构中。
{"title":"Emotional Paraphrasing Using Pre-trained Language Models","authors":"Jacky Casas, Samuel Torche, Karl Daher, E. Mugellini, Omar Abou Khaled","doi":"10.1109/aciiw52867.2021.9666309","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666309","url":null,"abstract":"Emotion style transfer is a recent and challenging problem in Natural Language Processing (NLP). Transformer-based language models are becoming extremely powerful, so one wonders if it would be possible to leverage them to perform emotion style transfer. So far, previous work has not used transformer-based models for this task. To address this task, we fine-tune a GPT-2 model with corrupted emotional data. This will train the model to increase the emotional intensity of the input sentence. Coupled with a paraphrasing model, we develop a system capable of transferring an emotion into a paraphrase. We conducted a qualitative study with human judges, as well as a quantitative evaluation. Although the paraphrase metrics show poor performance compared to the state of the art, the transfer of emotion proved to be effective, especially for the emotions fear, sadness, and disgust. The perception of these emotions were improved both in the automatic and human evaluations. Such technology can significantly facilitate the automatic creation of training sentences for natural language understanding (NLU) systems, but it can also be integrated into an emotional or empathic dialogue architecture.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"218 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133979716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Real-time Ubiquitous Pain Recognition 实时无处不在的疼痛识别
Iyonna Tynes, Shaun J. Canavan
Emotion recognition is a quickly growing field due to the increased interest in building systems which can classify and respond to emotions. Recent medical crises, such as the opioid overdose epidemic in the United States and the global COVID-19 pandemic has emphasized the importance of emotion recognition applications is areas like Telehealth services. Considering this, we propose an approach to real-time ubiquitous pain recognition from facial images. We have conducted offline experiments using the BP4D dataset, where we investigate the impact of gender and data imbalance. This paper proposes an affordable and easily accessible system which can perform pain recognition inferences. The results from this study found a balanced dataset, in terms of class and gender, results in the highest accuracies for pain recognition. We also detail the difficulties of pain recognition using facial images and propose some future work that can be investigated for this challenging problem.
情绪识别是一个快速发展的领域,因为人们对建立能够对情绪进行分类和反应的系统越来越感兴趣。最近的医疗危机,如美国阿片类药物过量流行和全球COVID-19大流行,都强调了情感识别应用在远程医疗服务等领域的重要性。考虑到这一点,我们提出了一种基于面部图像的实时无处不在的疼痛识别方法。我们使用BP4D数据集进行了离线实验,研究了性别和数据不平衡的影响。本文提出了一种价格合理且易于使用的系统,可以进行疼痛识别推理。这项研究的结果发现了一个平衡的数据集,在阶级和性别方面,疼痛识别的准确性最高。我们还详细介绍了使用面部图像识别疼痛的困难,并提出了一些未来的工作,可以研究这个具有挑战性的问题。
{"title":"Real-time Ubiquitous Pain Recognition","authors":"Iyonna Tynes, Shaun J. Canavan","doi":"10.1109/aciiw52867.2021.9666289","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666289","url":null,"abstract":"Emotion recognition is a quickly growing field due to the increased interest in building systems which can classify and respond to emotions. Recent medical crises, such as the opioid overdose epidemic in the United States and the global COVID-19 pandemic has emphasized the importance of emotion recognition applications is areas like Telehealth services. Considering this, we propose an approach to real-time ubiquitous pain recognition from facial images. We have conducted offline experiments using the BP4D dataset, where we investigate the impact of gender and data imbalance. This paper proposes an affordable and easily accessible system which can perform pain recognition inferences. The results from this study found a balanced dataset, in terms of class and gender, results in the highest accuracies for pain recognition. We also detail the difficulties of pain recognition using facial images and propose some future work that can be investigated for this challenging problem.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130209673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detection of Nudges and Measuring of Alignment in Spoken Interactions 在口语互动中检测暗示和测量对齐度
N. Kalashnikova
Nudges, techniques that indirectly influence human decision making, are little studied in spoken interactions. However, the limits of human-computer spoken interactions are not controlled, allowing machines realize bad nudges. In this context a framework for detecting nudges is needed to enhance the ethics of HCI. The work proposed in this PhD thesis is based on the hypothesis that the detection of nudges lies in measuring linguistic, paralinguistic and emotional alignments between interlocutors. Therefore, this PhD thesis aims to answer two research questions. First, does a high level of linguistic and paralinguistic alignement influence human's potential to be nudged? Second, if a person is resistant to other's emotions is she or he less sensible to be nudged? To get a better understanding of the correlation between alignment and nudges, but also a human's potential to be nudged knowing their level of alignment, we will conduct a series of experiments.
暗示(Nudges)是一种间接影响人类决策的技术,在口语交互中鲜有研究。然而,人机口语交互的局限性并没有得到控制,这使得机器可以实现不良暗示。在这种情况下,我们需要一个检测 "暗示 "的框架,以提高人机交互的道德水平。本博士论文提出的工作假设是,"推搡 "的检测在于测量对话者之间的语言、副语言和情感一致性。因此,本博士论文旨在回答两个研究问题。第一,高水平的语言和副语言一致性是否会影响人类被暗示的潜力?其次,如果一个人对他人的情感有抵触情绪,那么他(她)是否更不容易受到暗示?为了更好地理解一致性与被暗示之间的相关性,同时也为了更好地理解一个人的一致性水平对其被暗示潜力的影响,我们将进行一系列实验。
{"title":"Detection of Nudges and Measuring of Alignment in Spoken Interactions","authors":"N. Kalashnikova","doi":"10.1109/aciiw52867.2021.9666344","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666344","url":null,"abstract":"Nudges, techniques that indirectly influence human decision making, are little studied in spoken interactions. However, the limits of human-computer spoken interactions are not controlled, allowing machines realize bad nudges. In this context a framework for detecting nudges is needed to enhance the ethics of HCI. The work proposed in this PhD thesis is based on the hypothesis that the detection of nudges lies in measuring linguistic, paralinguistic and emotional alignments between interlocutors. Therefore, this PhD thesis aims to answer two research questions. First, does a high level of linguistic and paralinguistic alignement influence human's potential to be nudged? Second, if a person is resistant to other's emotions is she or he less sensible to be nudged? To get a better understanding of the correlation between alignment and nudges, but also a human's potential to be nudged knowing their level of alignment, we will conduct a series of experiments.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127386920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A system for collecting emotionally annotated physiological signals in daily life using wearables 一种利用可穿戴设备收集日常生活中带有情感注释的生理信号的系统
Stanisław Saganowski, Maciej Behnke, Joanna Komoszyńska, Dominika Kunc, Bartosz Perz, Przemyslaw Kazienko
Several obstacles have to be overcome in order to recognize emotions and affect in daily life. One of them is collecting a large amount of emotionally annotated data necessary to create data-greedy machine learning-based predictive models. Hence, we propose the Emognition system supporting the collection of rich emotional samples in everyday-life scenarios. The system utilizes smart-wearables to record physiological signals unobtrusively and smartphones to gather self-assessments. We have performed a two-week pilot study with 15 participants and devices available on the market to validate the system. The outcomes of the study, alongside the discussion and lessons learned, are provided.
为了识别日常生活中的情绪和影响,必须克服几个障碍。其中之一是收集大量的情感注释数据,这些数据是创建基于数据贪婪机器学习的预测模型所必需的。因此,我们提出了支持在日常生活场景中收集丰富情绪样本的emooging系统。该系统利用智能可穿戴设备不显眼地记录生理信号,并利用智能手机收集自我评估。我们对15名参与者和市场上可用的设备进行了为期两周的试点研究,以验证该系统。本文提供了研究结果、讨论和经验教训。
{"title":"A system for collecting emotionally annotated physiological signals in daily life using wearables","authors":"Stanisław Saganowski, Maciej Behnke, Joanna Komoszyńska, Dominika Kunc, Bartosz Perz, Przemyslaw Kazienko","doi":"10.1109/aciiw52867.2021.9666272","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666272","url":null,"abstract":"Several obstacles have to be overcome in order to recognize emotions and affect in daily life. One of them is collecting a large amount of emotionally annotated data necessary to create data-greedy machine learning-based predictive models. Hence, we propose the Emognition system supporting the collection of rich emotional samples in everyday-life scenarios. The system utilizes smart-wearables to record physiological signals unobtrusively and smartphones to gather self-assessments. We have performed a two-week pilot study with 15 participants and devices available on the market to validate the system. The outcomes of the study, alongside the discussion and lessons learned, are provided.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122444296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A simple baseline for evaluating Expression Transfer and Anonymisation in Video Transfer 评价视频传输中表达转移和匿名化的简单基线
Gabriel Haddon-Hill, Keerthy Kusumam, M. Valstar
Video-to-video synthesis methods provide increasingly accessible solutions for training models on privacy-sensitive and limited-size datasets frequently encountered in domains such as affect analysis. However, there are no existing baselines that explicitly measure the extent of reliable expression transfer or privacy preservation in the generated data. In this paper, we evaluate a general-purpose video transfer method, vid2vid, on these two key tasks: expression transfer and anonymisation of identities, as well as its suitability for training affect prediction models. We provide results that form a strong baseline for future comparisons, and further motivate the need for purpose-built methods for conducting expression-preserving video transfer. Our results indicate that a significant limitation of vid2vid's expression transfer arises from conditioning on facial landmarks and optical flow, which do not carry sufficient information to preserve facial expressions. Finally, we demonstrate that vid2vid can adequately anonymise videos in some cases, though not consistently, and that the anonymisation can be improved by applying random perturbations to input landmarks, at the cost of reduced expression transfer.
视频到视频合成方法为在影响分析等领域经常遇到的隐私敏感和有限大小的数据集上训练模型提供了越来越容易获得的解决方案。然而,目前还没有明确衡量生成数据中可靠的表达传递或隐私保护程度的基线。在本文中,我们评估了一种通用的视频传输方法,vid2vid,在这两个关键任务上:表达转移和身份匿名,以及它对训练影响预测模型的适用性。我们提供的结果为未来的比较提供了强有力的基线,并进一步激发了对用于进行保留表达的视频传输的专用方法的需求。我们的研究结果表明,vid2vid的表情传递受到面部标志和光流的限制,它们没有携带足够的信息来保存面部表情。最后,我们证明了vid2vid在某些情况下可以充分地匿名化视频,尽管不是一致的,并且可以通过对输入地标应用随机扰动来改进匿名化,以减少表达转移为代价。
{"title":"A simple baseline for evaluating Expression Transfer and Anonymisation in Video Transfer","authors":"Gabriel Haddon-Hill, Keerthy Kusumam, M. Valstar","doi":"10.1109/aciiw52867.2021.9666292","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666292","url":null,"abstract":"Video-to-video synthesis methods provide increasingly accessible solutions for training models on privacy-sensitive and limited-size datasets frequently encountered in domains such as affect analysis. However, there are no existing baselines that explicitly measure the extent of reliable expression transfer or privacy preservation in the generated data. In this paper, we evaluate a general-purpose video transfer method, vid2vid, on these two key tasks: expression transfer and anonymisation of identities, as well as its suitability for training affect prediction models. We provide results that form a strong baseline for future comparisons, and further motivate the need for purpose-built methods for conducting expression-preserving video transfer. Our results indicate that a significant limitation of vid2vid's expression transfer arises from conditioning on facial landmarks and optical flow, which do not carry sufficient information to preserve facial expressions. Finally, we demonstrate that vid2vid can adequately anonymise videos in some cases, though not consistently, and that the anonymisation can be improved by applying random perturbations to input landmarks, at the cost of reduced expression transfer.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115489477","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SUGO-MIMI: A Waggle Ear-Type Device Linked to Eyebrows SUGO-MIMI:与眉毛相连的摆动耳朵式装置
Shoko Kimura, Ayaka Fujii, Seiichi Harata, Takuto Sakuma, Shohei Kato
Facial expressions convey emotions. However, not all people are good at understanding such expressions in daily communication. To address this issue, we herein use “SUGO-MIMI,” a lightweight device that does not require a power source to expand facial expressions. SUGO-MIMI expands eyebrow movements by connecting the eyebrows to thin plates, imitating cat ears attached to the headband with wires. In our experiment of conveying facial expressions, the emotion denoting happiness was conveyed better when wearing SUGO-MIMI.
面部表情传达情感。然而,并不是所有的人都善于在日常交流中理解这些表达。为了解决这个问题,我们在这里使用了“SUGO-MIMI”,一种不需要电源来扩展面部表情的轻量级设备。SUGO-MIMI将眉毛连接到薄板上,模仿用电线连接在发带上的猫耳朵,扩大眉毛的活动。在我们的面部表情传递实验中,穿着SUGO-MIMI能更好地传达快乐的情绪。
{"title":"SUGO-MIMI: A Waggle Ear-Type Device Linked to Eyebrows","authors":"Shoko Kimura, Ayaka Fujii, Seiichi Harata, Takuto Sakuma, Shohei Kato","doi":"10.1109/aciiw52867.2021.9666368","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666368","url":null,"abstract":"Facial expressions convey emotions. However, not all people are good at understanding such expressions in daily communication. To address this issue, we herein use “SUGO-MIMI,” a lightweight device that does not require a power source to expand facial expressions. SUGO-MIMI expands eyebrow movements by connecting the eyebrows to thin plates, imitating cat ears attached to the headband with wires. In our experiment of conveying facial expressions, the emotion denoting happiness was conveyed better when wearing SUGO-MIMI.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"90 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123271422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Relationship between Mood Improvement and Questioning to Evaluate Automatic Thoughts in Cognitive Restructuring with a Virtual Agent 基于虚拟Agent的认知重构中自动思维的情绪改善与质疑的关系
Kazuhiro Shidara, Hiroki Tanaka, Hiroyoshi Adachi, D. Kanayama, Yukako Sakagami, Takashi Kudo, Satoshi Nakamura
Cognitive restructuring is a therapeutic technique of cognitive behavior therapy that helps convert negative automatic thoughts to balanced thoughts. Automatic thoughts can be converted by the patient/participant's own objective evaluation from a different perspective. Human therapists ask questions intending to guide the evaluation of automatic thoughts more effectively. Virtual agents, as therapists, have a great potential to support cognitive restructuring. We investigated how a virtual agent could affect the participant's mood when asked questions to evaluate automatic thoughts in cognitive restructuring. We implemented a virtual agent that performs scenario-based dialogue with two types of dialogue scenarios: with and without questions to evaluate automatic thoughts. We conducted a dialogue experiment with 20 healthy graduate students and divided them into two groups of ten, finding that the participant's negative mood significantly improved when the virtual agent asked questions to evaluate the automatic thoughts. Furthermore, the number of helpful questions was significantly correlated with the degree of mood change (ρ=0.81). The results suggest that it is important to provide appropriate questions for cognitive restructuring and that the number of helpful questions reflects the dialogue's effectiveness.
认知重组是认知行为疗法的一种治疗技术,有助于将消极的自动思维转化为平衡的思维。自动思维可以通过患者/参与者自己从不同角度的客观评价来转换。人类治疗师提出问题是为了更有效地指导对自动思维的评估。作为治疗师,虚拟代理具有支持认知重构的巨大潜力。我们调查了虚拟代理如何影响参与者的情绪,当被问及评估认知重构中的自动思维时。我们实现了一个虚拟代理,它与两种类型的对话场景进行基于场景的对话:有问题和没有问题,以评估自动思维。我们对20名健康的研究生进行了对话实验,并将他们分为两组,每组10人,发现当虚拟代理提出问题来评估自动思维时,参与者的负面情绪显著改善。此外,有用问题的数量与情绪变化程度显著相关(ρ=0.81)。研究结果表明,为认知重构提供合适的问题是很重要的,而有帮助的问题的数量反映了对话的有效性。
{"title":"Relationship between Mood Improvement and Questioning to Evaluate Automatic Thoughts in Cognitive Restructuring with a Virtual Agent","authors":"Kazuhiro Shidara, Hiroki Tanaka, Hiroyoshi Adachi, D. Kanayama, Yukako Sakagami, Takashi Kudo, Satoshi Nakamura","doi":"10.1109/aciiw52867.2021.9666312","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666312","url":null,"abstract":"Cognitive restructuring is a therapeutic technique of cognitive behavior therapy that helps convert negative automatic thoughts to balanced thoughts. Automatic thoughts can be converted by the patient/participant's own objective evaluation from a different perspective. Human therapists ask questions intending to guide the evaluation of automatic thoughts more effectively. Virtual agents, as therapists, have a great potential to support cognitive restructuring. We investigated how a virtual agent could affect the participant's mood when asked questions to evaluate automatic thoughts in cognitive restructuring. We implemented a virtual agent that performs scenario-based dialogue with two types of dialogue scenarios: with and without questions to evaluate automatic thoughts. We conducted a dialogue experiment with 20 healthy graduate students and divided them into two groups of ten, finding that the participant's negative mood significantly improved when the virtual agent asked questions to evaluate the automatic thoughts. Furthermore, the number of helpful questions was significantly correlated with the degree of mood change (ρ=0.81). The results suggest that it is important to provide appropriate questions for cognitive restructuring and that the number of helpful questions reflects the dialogue's effectiveness.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128962814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Multimodal Engagement-Aware Recommender System for People with Dementia 痴呆患者多模式参与感知推荐系统
Lars Steinert
Dementia places an immeasurable burden on affected individuals and caregivers. In non-pharmacological therapy, physical, social, and cognitive activation of People with Dementia (PwD) is known to be crucial. However, effective activation requires sustained engagement. Technical activation systems thus require a means to automatically recognize if the user is engaged. While research has shown that engagement can be automatically recognized in healthy individuals, this task is especially challenging for PwD who might suffer from aphasia or blunted affect. In this project, I aim to investigate whether PwD provide sufficient verbal and non-verbal signals for the automatic recognition of engagement. Next, I aim to build a multimodal engagement recognition system for PwD using a technical activation system. Lastly, I aim to leverage this knowledge to build and evaluate an engagement-aware recommender system to promote the usage of engaging activation contents.
痴呆症给受影响的个人和照顾者带来了无法估量的负担。在非药物治疗中,痴呆症患者(PwD)的身体、社会和认知激活是至关重要的。然而,有效的激活需要持续的参与。因此,技术激活系统需要一种自动识别用户是否参与的方法。虽然研究表明,健康的人可以自动识别参与,但这项任务对可能患有失语症或情感迟钝的残疾人来说尤其具有挑战性。在这个项目中,我的目的是研究残疾人士是否为自动识别参与提供了足够的语言和非语言信号。接下来,我的目标是使用技术激活系统为PwD构建一个多模式交战识别系统。最后,我的目标是利用这些知识来构建和评估一个参与性感知的推荐系统,以促进参与性激活内容的使用。
{"title":"A Multimodal Engagement-Aware Recommender System for People with Dementia","authors":"Lars Steinert","doi":"10.1109/aciiw52867.2021.9666306","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666306","url":null,"abstract":"Dementia places an immeasurable burden on affected individuals and caregivers. In non-pharmacological therapy, physical, social, and cognitive activation of People with Dementia (PwD) is known to be crucial. However, effective activation requires sustained engagement. Technical activation systems thus require a means to automatically recognize if the user is engaged. While research has shown that engagement can be automatically recognized in healthy individuals, this task is especially challenging for PwD who might suffer from aphasia or blunted affect. In this project, I aim to investigate whether PwD provide sufficient verbal and non-verbal signals for the automatic recognition of engagement. Next, I aim to build a multimodal engagement recognition system for PwD using a technical activation system. Lastly, I aim to leverage this knowledge to build and evaluate an engagement-aware recommender system to promote the usage of engaging activation contents.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"22 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130213632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Quantifying the Intensity of Toxicity for Discussions and Speakers 量化讨论和演讲者的毒性强度
Samiha Samrose, E. Hoque
In this work, from YouTube News-show multimodal dataset with dyadic speakers having heated discussions, we analyze the toxicity through audio-visual signals. Firstly, as different speakers may contribute differently towards the toxicity, we propose a speaker-wise toxicity score revealing individual proportionate contribution. As discussions with disagreements may reflect some signals of toxicity, in order to identify discussions needing more attention we categorize discussions into binary high-low toxicity levels. By analyzing visual features, we show that the levels correlate with facial expressions as Upper Lid Raiser (associated with ‘surprise’), Dimpler (associated with ‘contempť), and Lip Corner Depressor (associated with ‘disgust’) remain statistically significant in separating high-low intensities of disrespect. Secondly, we investigate the impact of audio-based features such as pitch and intensity that can significantly elicit disrespect, and utilize the signals in classifying disrespect and non-disrespect samples by applying logistic regression model achieving 79.86% accuracy. Our findings shed light on the potential of utilizing audio-visual signals in adding important context towards understanding toxic discussions.
在这项工作中,我们从YouTube新闻节目多模态数据集与二元发言者进行了激烈的讨论,我们通过视听信号分析毒性。首先,由于不同的讲话者对毒性的贡献不同,我们提出了一个讲话者毒性评分,揭示了个人的比例贡献。由于有分歧的讨论可能反映出一些毒性信号,为了确定需要更多关注的讨论,我们将讨论分为二元高-低毒性水平。通过分析视觉特征,我们发现该水平与面部表情相关,如上眼睑抬高(与“惊讶”相关),酒窝(与“蔑视”相关)和唇角下降(与“厌恶”相关)在区分高低强度的不尊重方面仍然具有统计学意义。其次,我们研究了音调和强度等基于音频的特征对不尊重行为的影响,并利用逻辑回归模型对不尊重和非不尊重样本进行分类,准确率达到79.86%。我们的研究结果揭示了利用视听信号为理解毒性讨论增加重要背景的潜力。
{"title":"Quantifying the Intensity of Toxicity for Discussions and Speakers","authors":"Samiha Samrose, E. Hoque","doi":"10.1109/aciiw52867.2021.9666258","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666258","url":null,"abstract":"In this work, from YouTube News-show multimodal dataset with dyadic speakers having heated discussions, we analyze the toxicity through audio-visual signals. Firstly, as different speakers may contribute differently towards the toxicity, we propose a speaker-wise toxicity score revealing individual proportionate contribution. As discussions with disagreements may reflect some signals of toxicity, in order to identify discussions needing more attention we categorize discussions into binary high-low toxicity levels. By analyzing visual features, we show that the levels correlate with facial expressions as Upper Lid Raiser (associated with ‘surprise’), Dimpler (associated with ‘contempť), and Lip Corner Depressor (associated with ‘disgust’) remain statistically significant in separating high-low intensities of disrespect. Secondly, we investigate the impact of audio-based features such as pitch and intensity that can significantly elicit disrespect, and utilize the signals in classifying disrespect and non-disrespect samples by applying logistic regression model achieving 79.86% accuracy. Our findings shed light on the potential of utilizing audio-visual signals in adding important context towards understanding toxic discussions.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123559581","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Embracing and Exploiting Annotator Emotional Subjectivity: An Affective Rater Ensemble Model 拥抱和开发注释者情感主体性:一个情感评注者的整体模型
Lukas Stappen, Lea Schumann, A. Batliner, Björn Schuller
Automated recognition of continuous emotions in audio-visual data is a growing area of study that aids in understanding human-machine interaction. Training such systems presupposes human annotation of the data. The annotation process, however, is laborious and expensive given that several human ratings are required for every data sample to compensate for the subjectivity of emotion perception. As a consequence, labelled data for emotion recognition are rare and the existing corpora are limited when compared to other state-of-the-art deep learning datasets. In this study, we explore different ways in which existing emotion annotations can be utilised more effectively to exploit available labelled information to the fullest. To reach this objective, we exploit individual raters’ opinions by employing an ensemble of rater-specific models, one for each annotator, by that reducing the loss of information which is a byproduct of annotation aggregation; we find that individual models can indeed infer subjective opinions. Furthermore, we explore the fusion of such ensemble predictions using different fusion techniques. Our ensemble model with only two annotators outperforms the regular Arousal baseline on the test set of the MuSe-CaR corpus. While no considerable improvements on valence could be obtained, using all annotators increases the prediction performance of arousal by up to. 07 Concordance Correlation Coefficient absolute improvement on test - solely trained on rate-specific models and fused by an attention-enhanced Long-short Term Memory-Recurrent Neural Network.
对视听数据中连续情绪的自动识别是一个不断发展的研究领域,有助于理解人机交互。训练这样的系统的前提是人类对数据进行注释。然而,由于每个数据样本都需要几个人工评级来弥补情感感知的主观性,因此注释过程既费力又昂贵。因此,与其他最先进的深度学习数据集相比,用于情感识别的标记数据很少,现有的语料库也很有限。在这项研究中,我们探索了不同的方法,在这些方法中,现有的情感注释可以更有效地利用现有的标记信息。为了达到这一目标,我们通过使用一个特定于评分者的模型集合来利用个人评分者的意见,每个评分者一个模型,通过减少注释聚合的副产品信息损失;我们发现单个模型确实可以推断主观意见。此外,我们使用不同的融合技术探索了这些集合预测的融合。我们的集成模型只有两个注释器,在MuSe-CaR语料库的测试集上优于常规的唤醒基线。虽然在效价上没有明显的改善,但使用所有注释器可以使唤醒的预测性能提高高达。07一致性相关系数在测试中的绝对改进-仅在特定率模型上进行训练,并由注意增强的长短期记忆循环神经网络融合。
{"title":"Embracing and Exploiting Annotator Emotional Subjectivity: An Affective Rater Ensemble Model","authors":"Lukas Stappen, Lea Schumann, A. Batliner, Björn Schuller","doi":"10.1109/aciiw52867.2021.9666407","DOIUrl":"https://doi.org/10.1109/aciiw52867.2021.9666407","url":null,"abstract":"Automated recognition of continuous emotions in audio-visual data is a growing area of study that aids in understanding human-machine interaction. Training such systems presupposes human annotation of the data. The annotation process, however, is laborious and expensive given that several human ratings are required for every data sample to compensate for the subjectivity of emotion perception. As a consequence, labelled data for emotion recognition are rare and the existing corpora are limited when compared to other state-of-the-art deep learning datasets. In this study, we explore different ways in which existing emotion annotations can be utilised more effectively to exploit available labelled information to the fullest. To reach this objective, we exploit individual raters’ opinions by employing an ensemble of rater-specific models, one for each annotator, by that reducing the loss of information which is a byproduct of annotation aggregation; we find that individual models can indeed infer subjective opinions. Furthermore, we explore the fusion of such ensemble predictions using different fusion techniques. Our ensemble model with only two annotators outperforms the regular Arousal baseline on the test set of the MuSe-CaR corpus. While no considerable improvements on valence could be obtained, using all annotators increases the prediction performance of arousal by up to. 07 Concordance Correlation Coefficient absolute improvement on test - solely trained on rate-specific models and fused by an attention-enhanced Long-short Term Memory-Recurrent Neural Network.","PeriodicalId":105376,"journal":{"name":"2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123786951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 9th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos (ACIIW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1