{"title":"An Effective Data Fusion Methodology for Multi-modal Emotion Recognition: A Survey","authors":"","doi":"10.30534/ijeter/2024/011272024","DOIUrl":null,"url":null,"abstract":"Emotion recognition is a pivotal area of research with applications spanning education, healthcare, and intelligent customer service. Multimodal emotion recognition (MER) has emerged as a superior approach by integrating multiple modalities such as speech, text, and facial expressions, offering enhanced accuracy and robustness over unimodal methods. This paper reviews the evolution and current state of MER, highlighting its significance, challenges, and methodologies. We delve into various datasets, including IEMOCAP and MELD, providing a comparative analysis of their strengths and limitations. The literature review covers recent advancements in deep learning techniques, focusing on fusion strategies like early, late, and hybrid fusion. Identified gaps include issues related to data redundancy, feature extraction complexity, and real-time detection. Our proposed methodology leverages deep learning for feature extraction and a hybrid fusion approach to improve emotion detection accuracy. This research aims to guide future studies in addressing current limitations and advancing the field of MER. The main of this paper review recent methodologies in multimodal emotion recognition, analyze different data fusion techniques,identify challenges and research gaps.","PeriodicalId":13964,"journal":{"name":"International Journal of Emerging Trends in Engineering Research","volume":"113 15","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Emerging Trends in Engineering Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.30534/ijeter/2024/011272024","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Engineering","Score":null,"Total":0}
引用次数: 0
Abstract
Emotion recognition is a pivotal area of research with applications spanning education, healthcare, and intelligent customer service. Multimodal emotion recognition (MER) has emerged as a superior approach by integrating multiple modalities such as speech, text, and facial expressions, offering enhanced accuracy and robustness over unimodal methods. This paper reviews the evolution and current state of MER, highlighting its significance, challenges, and methodologies. We delve into various datasets, including IEMOCAP and MELD, providing a comparative analysis of their strengths and limitations. The literature review covers recent advancements in deep learning techniques, focusing on fusion strategies like early, late, and hybrid fusion. Identified gaps include issues related to data redundancy, feature extraction complexity, and real-time detection. Our proposed methodology leverages deep learning for feature extraction and a hybrid fusion approach to improve emotion detection accuracy. This research aims to guide future studies in addressing current limitations and advancing the field of MER. The main of this paper review recent methodologies in multimodal emotion recognition, analyze different data fusion techniques,identify challenges and research gaps.
情感识别是一个关键的研究领域,其应用范围涵盖教育、医疗保健和智能客户服务。多模态情感识别(MER)通过整合语音、文本和面部表情等多种模态而成为一种卓越的方法,与单模态方法相比,它具有更高的准确性和鲁棒性。本文回顾了 MER 的演变和现状,强调了其意义、挑战和方法。我们深入研究了各种数据集,包括 IEMOCAP 和 MELD,对它们的优势和局限性进行了比较分析。文献综述涵盖了深度学习技术的最新进展,重点关注早期、晚期和混合融合等融合策略。发现的差距包括与数据冗余、特征提取复杂性和实时检测相关的问题。我们提出的方法利用深度学习的特征提取和混合融合方法来提高情感检测的准确性。这项研究旨在指导未来的研究,解决目前的局限性,推动 MER 领域的发展。本文主要回顾了多模态情感识别的最新方法,分析了不同的数据融合技术,指出了面临的挑战和研究空白。