{"title":"Emotion recognition in Arabic speech","authors":"Imene Hadjadji, L. Falek, Lyes Demri, H. Teffahi","doi":"10.1109/ICAEE47123.2019.9014809","DOIUrl":null,"url":null,"abstract":"The general objective of this paper is to build a system in order to automatically recognize emotion in speech. The linguistic material used is a corpus of Arabic expressive sentences phonetically balanced. The dependence of the system on speaker is an encountered problem in this field; in this work we will study the influence of this phenomenon on our result. The targeted emotions are joy, sadness, anger and neutral. After an analytical study of a large number of speech acoustic parameters, we chose the cepstral parameters, their first and second derivatives, the Shimmer, the Jitter and the duration of the sentence. A classifier based on a multilayer perceptron neural network to recognize emotion on the basis of the chosen feature vector that has been developed. The recognition rate could reach more than 98% in the case of an intra-speaker classification and 54.75% in inter-speaker classification. We can see the system’s dependence on speaker clearly.","PeriodicalId":197612,"journal":{"name":"2019 International Conference on Advanced Electrical Engineering (ICAEE)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on Advanced Electrical Engineering (ICAEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAEE47123.2019.9014809","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
The general objective of this paper is to build a system in order to automatically recognize emotion in speech. The linguistic material used is a corpus of Arabic expressive sentences phonetically balanced. The dependence of the system on speaker is an encountered problem in this field; in this work we will study the influence of this phenomenon on our result. The targeted emotions are joy, sadness, anger and neutral. After an analytical study of a large number of speech acoustic parameters, we chose the cepstral parameters, their first and second derivatives, the Shimmer, the Jitter and the duration of the sentence. A classifier based on a multilayer perceptron neural network to recognize emotion on the basis of the chosen feature vector that has been developed. The recognition rate could reach more than 98% in the case of an intra-speaker classification and 54.75% in inter-speaker classification. We can see the system’s dependence on speaker clearly.