{"title":"Chirplet transform based time frequency analysis of speech signal for automated speech emotion recognition","authors":"Siba Prasad Mishra, Pankaj Warule, Suman Deb","doi":"10.1016/j.specom.2023.102986","DOIUrl":null,"url":null,"abstract":"<div><p>Nowadays, the recognition of emotion using the speech signal has gained popularity because of its vast number of applications in different fields like medicine, online marketing, online search engines, the education system, criminal investigations, traffic collisions, etc. Many researchers have adopted different methodologies to improve emotion classification accuracy using speech signals. In our study, time–frequency (TF) analysis-based features were used to analyze the emotion classification performance. We used a novel TF analysis method called the chirplet transform (CT) to find the TF matrix of the speech signal. We then calculated the proposed TF-based permutation entropy (TFPE) feature using the TF matrix of the speech signal. To reduce the feature dimension and select the most informative emotional feature, we employed the genetic algorithm (GA) feature selection method. Then, the selected TFPE features are used as input to machine learning classifiers such as SVM, RF, DT, and KNN to classify the emotions in the speech signal. We obtained classification accuracy of 77.2%, 69.57%, 68.78%, 56.9%, and 99.1% for the EMO-DB, EMOVO, SAVEE, IEMOCAP, and TESS datasets without the GA feature selection method. The emotion classification accuracy increased to 85.6%, 78.33%, 77.76%, 63.15%, and 100% with the GA feature selection method. We compared our results with other methods and found that our method performed better in emotion classification than the state-of-the-art methods.</p></div>","PeriodicalId":49485,"journal":{"name":"Speech Communication","volume":"155 ","pages":"Article 102986"},"PeriodicalIF":2.4000,"publicationDate":"2023-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Speech Communication","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167639323001206","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ACOUSTICS","Score":null,"Total":0}
引用次数: 1
Abstract
Nowadays, the recognition of emotion using the speech signal has gained popularity because of its vast number of applications in different fields like medicine, online marketing, online search engines, the education system, criminal investigations, traffic collisions, etc. Many researchers have adopted different methodologies to improve emotion classification accuracy using speech signals. In our study, time–frequency (TF) analysis-based features were used to analyze the emotion classification performance. We used a novel TF analysis method called the chirplet transform (CT) to find the TF matrix of the speech signal. We then calculated the proposed TF-based permutation entropy (TFPE) feature using the TF matrix of the speech signal. To reduce the feature dimension and select the most informative emotional feature, we employed the genetic algorithm (GA) feature selection method. Then, the selected TFPE features are used as input to machine learning classifiers such as SVM, RF, DT, and KNN to classify the emotions in the speech signal. We obtained classification accuracy of 77.2%, 69.57%, 68.78%, 56.9%, and 99.1% for the EMO-DB, EMOVO, SAVEE, IEMOCAP, and TESS datasets without the GA feature selection method. The emotion classification accuracy increased to 85.6%, 78.33%, 77.76%, 63.15%, and 100% with the GA feature selection method. We compared our results with other methods and found that our method performed better in emotion classification than the state-of-the-art methods.
期刊介绍:
Speech Communication is an interdisciplinary journal whose primary objective is to fulfil the need for the rapid dissemination and thorough discussion of basic and applied research results.
The journal''s primary objectives are:
• to present a forum for the advancement of human and human-machine speech communication science;
• to stimulate cross-fertilization between different fields of this domain;
• to contribute towards the rapid and wide diffusion of scientifically sound contributions in this domain.