基于小波变换的语音信号时频分析用于语音情感自动识别

IF 2.4 3区 计算机科学 Q2 ACOUSTICS Speech Communication Pub Date : 2023-09-23 DOI:10.1016/j.specom.2023.102986
Siba Prasad Mishra, Pankaj Warule, Suman Deb
{"title":"基于小波变换的语音信号时频分析用于语音情感自动识别","authors":"Siba Prasad Mishra,&nbsp;Pankaj Warule,&nbsp;Suman Deb","doi":"10.1016/j.specom.2023.102986","DOIUrl":null,"url":null,"abstract":"<div><p>Nowadays, the recognition of emotion using the speech signal has gained popularity because of its vast number of applications in different fields like medicine, online marketing, online search engines, the education system, criminal investigations, traffic collisions, etc. Many researchers have adopted different methodologies to improve emotion classification accuracy using speech signals. In our study, time–frequency (TF) analysis-based features were used to analyze the emotion classification performance. We used a novel TF analysis method called the chirplet transform (CT) to find the TF matrix of the speech signal. We then calculated the proposed TF-based permutation entropy (TFPE) feature using the TF matrix of the speech signal. To reduce the feature dimension and select the most informative emotional feature, we employed the genetic algorithm (GA) feature selection method. Then, the selected TFPE features are used as input to machine learning classifiers such as SVM, RF, DT, and KNN to classify the emotions in the speech signal. We obtained classification accuracy of 77.2%, 69.57%, 68.78%, 56.9%, and 99.1% for the EMO-DB, EMOVO, SAVEE, IEMOCAP, and TESS datasets without the GA feature selection method. The emotion classification accuracy increased to 85.6%, 78.33%, 77.76%, 63.15%, and 100% with the GA feature selection method. We compared our results with other methods and found that our method performed better in emotion classification than the state-of-the-art methods.</p></div>","PeriodicalId":49485,"journal":{"name":"Speech Communication","volume":"155 ","pages":"Article 102986"},"PeriodicalIF":2.4000,"publicationDate":"2023-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Chirplet transform based time frequency analysis of speech signal for automated speech emotion recognition\",\"authors\":\"Siba Prasad Mishra,&nbsp;Pankaj Warule,&nbsp;Suman Deb\",\"doi\":\"10.1016/j.specom.2023.102986\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Nowadays, the recognition of emotion using the speech signal has gained popularity because of its vast number of applications in different fields like medicine, online marketing, online search engines, the education system, criminal investigations, traffic collisions, etc. Many researchers have adopted different methodologies to improve emotion classification accuracy using speech signals. In our study, time–frequency (TF) analysis-based features were used to analyze the emotion classification performance. We used a novel TF analysis method called the chirplet transform (CT) to find the TF matrix of the speech signal. We then calculated the proposed TF-based permutation entropy (TFPE) feature using the TF matrix of the speech signal. To reduce the feature dimension and select the most informative emotional feature, we employed the genetic algorithm (GA) feature selection method. Then, the selected TFPE features are used as input to machine learning classifiers such as SVM, RF, DT, and KNN to classify the emotions in the speech signal. We obtained classification accuracy of 77.2%, 69.57%, 68.78%, 56.9%, and 99.1% for the EMO-DB, EMOVO, SAVEE, IEMOCAP, and TESS datasets without the GA feature selection method. The emotion classification accuracy increased to 85.6%, 78.33%, 77.76%, 63.15%, and 100% with the GA feature selection method. We compared our results with other methods and found that our method performed better in emotion classification than the state-of-the-art methods.</p></div>\",\"PeriodicalId\":49485,\"journal\":{\"name\":\"Speech Communication\",\"volume\":\"155 \",\"pages\":\"Article 102986\"},\"PeriodicalIF\":2.4000,\"publicationDate\":\"2023-09-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Speech Communication\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167639323001206\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ACOUSTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Speech Communication","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167639323001206","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ACOUSTICS","Score":null,"Total":0}
引用次数: 1

摘要

如今,利用语音信号识别情绪已经受到欢迎,因为它在医学、在线营销、在线搜索引擎、教育系统、刑事调查、交通碰撞等不同领域有着广泛的应用。许多研究人员采用了不同的方法来提高利用语音信号进行情绪分类的准确性。在我们的研究中,使用基于时间-频率(TF)分析的特征来分析情绪分类性能。我们使用了一种新的TF分析方法,称为啁啾变换(CT)来找到语音信号的TF矩阵。然后,我们使用语音信号的TF矩阵来计算所提出的基于TF的排列熵(TFPE)特征。为了降低特征维数并选择信息量最大的情感特征,我们采用了遗传算法(GA)的特征选择方法。然后,将所选择的TFPE特征用作机器学习分类器(如SVM、RF、DT和KNN)的输入,以对语音信号中的情绪进行分类。在没有GA特征选择方法的情况下,我们对EMO-DB、EMOVO、SAVEE、IEMOCAP和TESS数据集的分类准确率分别为77.2%、69.57%、68.78%、56.9%和99.1%。采用GA特征选择方法,情绪分类准确率分别提高到85.6%、78.33%、77.76%、63.15%和100%。我们将我们的结果与其他方法进行了比较,发现我们的方法在情绪分类方面比最先进的方法表现得更好。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Chirplet transform based time frequency analysis of speech signal for automated speech emotion recognition

Nowadays, the recognition of emotion using the speech signal has gained popularity because of its vast number of applications in different fields like medicine, online marketing, online search engines, the education system, criminal investigations, traffic collisions, etc. Many researchers have adopted different methodologies to improve emotion classification accuracy using speech signals. In our study, time–frequency (TF) analysis-based features were used to analyze the emotion classification performance. We used a novel TF analysis method called the chirplet transform (CT) to find the TF matrix of the speech signal. We then calculated the proposed TF-based permutation entropy (TFPE) feature using the TF matrix of the speech signal. To reduce the feature dimension and select the most informative emotional feature, we employed the genetic algorithm (GA) feature selection method. Then, the selected TFPE features are used as input to machine learning classifiers such as SVM, RF, DT, and KNN to classify the emotions in the speech signal. We obtained classification accuracy of 77.2%, 69.57%, 68.78%, 56.9%, and 99.1% for the EMO-DB, EMOVO, SAVEE, IEMOCAP, and TESS datasets without the GA feature selection method. The emotion classification accuracy increased to 85.6%, 78.33%, 77.76%, 63.15%, and 100% with the GA feature selection method. We compared our results with other methods and found that our method performed better in emotion classification than the state-of-the-art methods.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Speech Communication
Speech Communication 工程技术-计算机:跨学科应用
CiteScore
6.80
自引率
6.20%
发文量
94
审稿时长
19.2 weeks
期刊介绍: Speech Communication is an interdisciplinary journal whose primary objective is to fulfil the need for the rapid dissemination and thorough discussion of basic and applied research results. The journal''s primary objectives are: • to present a forum for the advancement of human and human-machine speech communication science; • to stimulate cross-fertilization between different fields of this domain; • to contribute towards the rapid and wide diffusion of scientifically sound contributions in this domain.
期刊最新文献
Fixed frequency range empirical wavelet transform based acoustic and entropy features for speech emotion recognition AFP-Conformer: Asymptotic feature pyramid conformer for spoofing speech detection A robust temporal map of speech monitoring from planning to articulation The combined effects of bilingualism and musicianship on listeners’ perception of non-native lexical tones Evaluating the effects of continuous pitch and speech tempo modifications on perceptual speaker verification performance by familiar and unfamiliar listeners
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1