Global-View and Speaker-Aware Emotion Cause Extraction in Conversations

IF 4.1 2区 计算机科学 Q1 ACOUSTICS IEEE/ACM Transactions on Audio, Speech, and Language Processing Pub Date : 2023-10-09 DOI:10.1109/TASLP.2023.3319990
Jiaming An;Zixiang Ding;Ke Li;Rui Xia
{"title":"Global-View and Speaker-Aware Emotion Cause Extraction in Conversations","authors":"Jiaming An;Zixiang Ding;Ke Li;Rui Xia","doi":"10.1109/TASLP.2023.3319990","DOIUrl":null,"url":null,"abstract":"Emotion cause extraction in conversations, the task of recognizing and extracting the causes behind the emotions in a conversation, is a new and under-explored task. It was previously treated as an utterance-level task, that can only extract cause of one emotion from one utterance at a time and is difficult to model the correlation between different emotions and causes in the conversation. The role of speakers was also not fully utilized in the previous methods. In this article, we introduce a global-view and speaker-aware conversational emotion cause extraction framework. It can fully model the interaction between utterances and emotions in the conversation and simultaneously extract all the causes corresponding to all emotions or one given emotion in a conversation, and can be applied to both real-time and non-real-time task settings. We further propose a Speaker-aware Couple-Decoder Module and a Speaker-Emotion Graph Attention Network, to better model the role of speakers in the conversation. The experimental results prove our approach's advantages in both emotion cause extraction performance and computational efficiency.","PeriodicalId":13332,"journal":{"name":"IEEE/ACM Transactions on Audio, Speech, and Language Processing","volume":"31 ","pages":"3814-3823"},"PeriodicalIF":4.1000,"publicationDate":"2023-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE/ACM Transactions on Audio, Speech, and Language Processing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10274611/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ACOUSTICS","Score":null,"Total":0}
引用次数: 0

Abstract

Emotion cause extraction in conversations, the task of recognizing and extracting the causes behind the emotions in a conversation, is a new and under-explored task. It was previously treated as an utterance-level task, that can only extract cause of one emotion from one utterance at a time and is difficult to model the correlation between different emotions and causes in the conversation. The role of speakers was also not fully utilized in the previous methods. In this article, we introduce a global-view and speaker-aware conversational emotion cause extraction framework. It can fully model the interaction between utterances and emotions in the conversation and simultaneously extract all the causes corresponding to all emotions or one given emotion in a conversation, and can be applied to both real-time and non-real-time task settings. We further propose a Speaker-aware Couple-Decoder Module and a Speaker-Emotion Graph Attention Network, to better model the role of speakers in the conversation. The experimental results prove our approach's advantages in both emotion cause extraction performance and computational efficiency.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
全局观与说话人感知的会话情感原因提取
会话中的情绪原因提取,即识别和提取会话中情绪背后的原因,是一项新的、未被充分探索的任务。它以前被视为一个话语层面的任务,一次只能从一个话语中提取一种情绪的原因,并且很难对对话中不同情绪和原因之间的相关性进行建模。发言者的作用在以前的方法中也没有得到充分利用。在本文中,我们介绍了一个全局观和说话人感知的会话情感原因提取框架。它可以完全建模对话中话语和情绪之间的互动,同时提取对话中所有情绪或一种给定情绪对应的所有原因,可以应用于实时和非实时任务设置。我们进一步提出了一个说话人感知耦合解码器模块和说话人情感图注意力网络,以更好地模拟说话人在对话中的角色。实验结果证明了我们的方法在情感原因提取性能和计算效率方面的优势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE/ACM Transactions on Audio, Speech, and Language Processing
IEEE/ACM Transactions on Audio, Speech, and Language Processing ACOUSTICS-ENGINEERING, ELECTRICAL & ELECTRONIC
CiteScore
11.30
自引率
11.10%
发文量
217
期刊介绍: The IEEE/ACM Transactions on Audio, Speech, and Language Processing covers audio, speech and language processing and the sciences that support them. In audio processing: transducers, room acoustics, active sound control, human audition, analysis/synthesis/coding of music, and consumer audio. In speech processing: areas such as speech analysis, synthesis, coding, speech and speaker recognition, speech production and perception, and speech enhancement. In language processing: speech and text analysis, understanding, generation, dialog management, translation, summarization, question answering and document indexing and retrieval, as well as general language modeling.
期刊最新文献
Enhancing Robustness of Speech Watermarking Using a Transformer-Based Framework Exploiting Acoustic Features FxLMS/F Based Tap Decomposed Adaptive Filter for Decentralized Active Noise Control System MRC-PASCL: A Few-Shot Machine Reading Comprehension Approach via Post-Training and Answer Span-Oriented Contrastive Learning Knowledge-Guided Transformer for Joint Theme and Emotion Classification of Chinese Classical Poetry WEDA: Exploring Copyright Protection for Large Language Model Downstream Alignment
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1