{"title":"利用基于频谱和熵特征的多分辨率希尔伯特变换进行语音情感识别","authors":"Siba Prasad Mishra, Pankaj Warule, Suman Deb","doi":"10.1016/j.apacoust.2024.110403","DOIUrl":null,"url":null,"abstract":"<div><div>Speech emotion recognition (SER) is essential for addressing many personal and professional challenges in our everyday lives. The application of SER has shown potential in a number of domains, such as medical intervention, fortification of security systems, online marketing and educational platforms, personal communication, strengthening of devices and human interaction, and numerous other domains. Due to its extensive variety of applications, this subject has attracted the attention of several researchers for more than three decades. The performance of SER can be improved by adopting a suitable methodology for extracting the feature and using it to classify speech emotion. In our study, we used a novel technique known as the multi-resolution Hilbert transform (MRHT) method to extract the speech feature. We used the multi-resolution signal decomposition (MRSD) method to break down the speech signal frame (SSF) into a number of sub-frequency band signals, which are called modes or intrinsic mode functions (IMFs). Then, Hilbert transform (HT) is applied to each IMF signal to find the MRHT-based instantaneous amplitude (MRHIA) and MRHT-based instantaneous frequency (MRHIF) signal vectors. Features such as MRHT-based approximate entropy (MRHAE), MRHT-based permutation entropy (MRHPE), MRHT-based increment entropy (MRHIE), MRHT-based spectral entropy (MRHSE), and MRHT-based sample entropy (MRHSME) were calculated using each MRHIA and MRHIF signal vectors and the mel frequency cepstral coefficient (MFCC) feature extracted using the speech signals. The combinations of the proposed MRHT-based features (MRHAE + MRHPE + MRHIE + MRHSE + MRHSME) are known as the MRHT-based entropy feature (MRHEF). Subsequently, the MRHEF and MFCC features are used both alone and in conjunction to categorize speech emotion using a deep neural network (DNN) classifier. This results in emotion classification accuracies of 89.67%, 85.42%, and 83.48% for the EMO-DB, EMOVO, and SAVEE datasets, respectively. Comparing our experimental results with the other approaches, we found that the proposed feature combinations (MFCC + MRHEF) using a DNN classifier outperformed the state-of-the-art methods in SER.</div></div>","PeriodicalId":55506,"journal":{"name":"Applied Acoustics","volume":"229 ","pages":"Article 110403"},"PeriodicalIF":3.4000,"publicationDate":"2024-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Speech emotion recognition using multi resolution Hilbert transform based spectral and entropy features\",\"authors\":\"Siba Prasad Mishra, Pankaj Warule, Suman Deb\",\"doi\":\"10.1016/j.apacoust.2024.110403\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Speech emotion recognition (SER) is essential for addressing many personal and professional challenges in our everyday lives. The application of SER has shown potential in a number of domains, such as medical intervention, fortification of security systems, online marketing and educational platforms, personal communication, strengthening of devices and human interaction, and numerous other domains. Due to its extensive variety of applications, this subject has attracted the attention of several researchers for more than three decades. The performance of SER can be improved by adopting a suitable methodology for extracting the feature and using it to classify speech emotion. In our study, we used a novel technique known as the multi-resolution Hilbert transform (MRHT) method to extract the speech feature. We used the multi-resolution signal decomposition (MRSD) method to break down the speech signal frame (SSF) into a number of sub-frequency band signals, which are called modes or intrinsic mode functions (IMFs). Then, Hilbert transform (HT) is applied to each IMF signal to find the MRHT-based instantaneous amplitude (MRHIA) and MRHT-based instantaneous frequency (MRHIF) signal vectors. Features such as MRHT-based approximate entropy (MRHAE), MRHT-based permutation entropy (MRHPE), MRHT-based increment entropy (MRHIE), MRHT-based spectral entropy (MRHSE), and MRHT-based sample entropy (MRHSME) were calculated using each MRHIA and MRHIF signal vectors and the mel frequency cepstral coefficient (MFCC) feature extracted using the speech signals. The combinations of the proposed MRHT-based features (MRHAE + MRHPE + MRHIE + MRHSE + MRHSME) are known as the MRHT-based entropy feature (MRHEF). Subsequently, the MRHEF and MFCC features are used both alone and in conjunction to categorize speech emotion using a deep neural network (DNN) classifier. This results in emotion classification accuracies of 89.67%, 85.42%, and 83.48% for the EMO-DB, EMOVO, and SAVEE datasets, respectively. Comparing our experimental results with the other approaches, we found that the proposed feature combinations (MFCC + MRHEF) using a DNN classifier outperformed the state-of-the-art methods in SER.</div></div>\",\"PeriodicalId\":55506,\"journal\":{\"name\":\"Applied Acoustics\",\"volume\":\"229 \",\"pages\":\"Article 110403\"},\"PeriodicalIF\":3.4000,\"publicationDate\":\"2024-11-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Acoustics\",\"FirstCategoryId\":\"101\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0003682X24005541\",\"RegionNum\":2,\"RegionCategory\":\"物理与天体物理\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ACOUSTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Acoustics","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0003682X24005541","RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ACOUSTICS","Score":null,"Total":0}
Speech emotion recognition using multi resolution Hilbert transform based spectral and entropy features
Speech emotion recognition (SER) is essential for addressing many personal and professional challenges in our everyday lives. The application of SER has shown potential in a number of domains, such as medical intervention, fortification of security systems, online marketing and educational platforms, personal communication, strengthening of devices and human interaction, and numerous other domains. Due to its extensive variety of applications, this subject has attracted the attention of several researchers for more than three decades. The performance of SER can be improved by adopting a suitable methodology for extracting the feature and using it to classify speech emotion. In our study, we used a novel technique known as the multi-resolution Hilbert transform (MRHT) method to extract the speech feature. We used the multi-resolution signal decomposition (MRSD) method to break down the speech signal frame (SSF) into a number of sub-frequency band signals, which are called modes or intrinsic mode functions (IMFs). Then, Hilbert transform (HT) is applied to each IMF signal to find the MRHT-based instantaneous amplitude (MRHIA) and MRHT-based instantaneous frequency (MRHIF) signal vectors. Features such as MRHT-based approximate entropy (MRHAE), MRHT-based permutation entropy (MRHPE), MRHT-based increment entropy (MRHIE), MRHT-based spectral entropy (MRHSE), and MRHT-based sample entropy (MRHSME) were calculated using each MRHIA and MRHIF signal vectors and the mel frequency cepstral coefficient (MFCC) feature extracted using the speech signals. The combinations of the proposed MRHT-based features (MRHAE + MRHPE + MRHIE + MRHSE + MRHSME) are known as the MRHT-based entropy feature (MRHEF). Subsequently, the MRHEF and MFCC features are used both alone and in conjunction to categorize speech emotion using a deep neural network (DNN) classifier. This results in emotion classification accuracies of 89.67%, 85.42%, and 83.48% for the EMO-DB, EMOVO, and SAVEE datasets, respectively. Comparing our experimental results with the other approaches, we found that the proposed feature combinations (MFCC + MRHEF) using a DNN classifier outperformed the state-of-the-art methods in SER.
期刊介绍:
Since its launch in 1968, Applied Acoustics has been publishing high quality research papers providing state-of-the-art coverage of research findings for engineers and scientists involved in applications of acoustics in the widest sense.
Applied Acoustics looks not only at recent developments in the understanding of acoustics but also at ways of exploiting that understanding. The Journal aims to encourage the exchange of practical experience through publication and in so doing creates a fund of technological information that can be used for solving related problems. The presentation of information in graphical or tabular form is especially encouraged. If a report of a mathematical development is a necessary part of a paper it is important to ensure that it is there only as an integral part of a practical solution to a problem and is supported by data. Applied Acoustics encourages the exchange of practical experience in the following ways: • Complete Papers • Short Technical Notes • Review Articles; and thereby provides a wealth of technological information that can be used to solve related problems.
Manuscripts that address all fields of applications of acoustics ranging from medicine and NDT to the environment and buildings are welcome.