EMOTION ANALYSIS USING SIGNAL AND IMAGE PROCESSING APPROACH BY IMPLEMENTING DEEP NEURAL NETWORK

S. Shuma, T. Bobby, S. Malathi
{"title":"EMOTION ANALYSIS USING SIGNAL AND IMAGE PROCESSING APPROACH BY IMPLEMENTING DEEP NEURAL NETWORK","authors":"S. Shuma, T. Bobby, S. Malathi","doi":"10.34107/yhpn9422.04313","DOIUrl":null,"url":null,"abstract":"Emotion recognition is important in human communication and to achieve a complete interaction between humans and machines. In medical applications, emotion recognition is used to assist the children with Autism Spectrum Disorder (ASD to improve their socio-emotional communication, helps doctors with diagnosis of diseases such as depression and dementia and also helps the caretakers of older patients to monitor their well-being. This paper discusses the application of feature level fusion of speech and facial expressions of different emotions such as neutral, happy, sad, angry, surprise, fearful and disgust. Also, to explore how best to build the deep learning networks to classify the emotions independently and jointly from these two modalities. VGG-model is utilized to extract features from facial images, and spectral features are extracted from speech signals. Further, feature level fusion technique is adopted to fuse the features extracted from the two modalities. Principal Component Analysis (PCA is implemented to choose the significant features. The proposed method achieved a maximum score of 90% on training set and 82% on validation set. The recognition rate in case of multimodal data improved greatly when compared to unimodal system. The multimodal system gave an improvement of 9% compared to the performance of the system based on speech. Thus, result shows that the proposed Multimodal Emotion Recognition (MER outperform the unimodal emotion recognition system.","PeriodicalId":75599,"journal":{"name":"Biomedical sciences instrumentation","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical sciences instrumentation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.34107/yhpn9422.04313","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Emotion recognition is important in human communication and to achieve a complete interaction between humans and machines. In medical applications, emotion recognition is used to assist the children with Autism Spectrum Disorder (ASD to improve their socio-emotional communication, helps doctors with diagnosis of diseases such as depression and dementia and also helps the caretakers of older patients to monitor their well-being. This paper discusses the application of feature level fusion of speech and facial expressions of different emotions such as neutral, happy, sad, angry, surprise, fearful and disgust. Also, to explore how best to build the deep learning networks to classify the emotions independently and jointly from these two modalities. VGG-model is utilized to extract features from facial images, and spectral features are extracted from speech signals. Further, feature level fusion technique is adopted to fuse the features extracted from the two modalities. Principal Component Analysis (PCA is implemented to choose the significant features. The proposed method achieved a maximum score of 90% on training set and 82% on validation set. The recognition rate in case of multimodal data improved greatly when compared to unimodal system. The multimodal system gave an improvement of 9% compared to the performance of the system based on speech. Thus, result shows that the proposed Multimodal Emotion Recognition (MER outperform the unimodal emotion recognition system.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
情感分析采用信号和图像处理方法,实现深度神经网络
情感识别在人类交流和实现人与机器之间的完整互动中非常重要。在医学应用中,情绪识别用于帮助患有自闭症谱系障碍的儿童(ASD可以改善他们的社会情感交流,帮助医生诊断抑郁症和痴呆症等疾病,也可以帮助老年患者的护理人员监测他们的健康状况。本文讨论了中性、快乐、悲伤、愤怒、惊讶、恐惧和厌恶等不同情绪的言语和面部表情的特征级融合的应用或者如何最好地构建深度学习网络,从这两种模式中独立和联合地对情绪进行分类。利用VGG模型从人脸图像中提取特征,并从语音信号中提取频谱特征。此外,采用特征级融合技术对从两种模态中提取的特征进行融合。主成分分析(主成分分析用于选择显著特征。该方法在训练集上的最高得分为90%,在验证集上的得分为82%。与单模态系统相比,多模态数据的识别率大大提高。与基于语音的系统相比,该多模态系统的性能提高了9%。因此,结果表明模态情感识别(MER)优于单一模态情感识别系统。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
0.50
自引率
0.00%
发文量
0
期刊最新文献
The Roles of Echocardiography in Teaching of Cardiovascular Physiology at Pre-Clinical Level of Undergraduate Medical Education Positively Charged Water as a Tumor Growth Stimulator Impact of COVID-19 on Liver Function Tests Among Sudanese Patients: A Cross-Sectional Study of Khartoum State Analysis of the Mechanism of Salvia miltiorrhiza in the Treatment of Pancreatic Cancer Based on Network Pharmacology Serological Status of Viral Hepatitis B and Associated Factors Among Sex Workers in Douala (Littoral-Cameroon)
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1