{"title":"阿拉伯语语音中的情感识别","authors":"Imene Hadjadji, L. Falek, Lyes Demri, H. Teffahi","doi":"10.1109/ICAEE47123.2019.9014809","DOIUrl":null,"url":null,"abstract":"The general objective of this paper is to build a system in order to automatically recognize emotion in speech. The linguistic material used is a corpus of Arabic expressive sentences phonetically balanced. The dependence of the system on speaker is an encountered problem in this field; in this work we will study the influence of this phenomenon on our result. The targeted emotions are joy, sadness, anger and neutral. After an analytical study of a large number of speech acoustic parameters, we chose the cepstral parameters, their first and second derivatives, the Shimmer, the Jitter and the duration of the sentence. A classifier based on a multilayer perceptron neural network to recognize emotion on the basis of the chosen feature vector that has been developed. The recognition rate could reach more than 98% in the case of an intra-speaker classification and 54.75% in inter-speaker classification. We can see the system’s dependence on speaker clearly.","PeriodicalId":197612,"journal":{"name":"2019 International Conference on Advanced Electrical Engineering (ICAEE)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Emotion recognition in Arabic speech\",\"authors\":\"Imene Hadjadji, L. Falek, Lyes Demri, H. Teffahi\",\"doi\":\"10.1109/ICAEE47123.2019.9014809\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The general objective of this paper is to build a system in order to automatically recognize emotion in speech. The linguistic material used is a corpus of Arabic expressive sentences phonetically balanced. The dependence of the system on speaker is an encountered problem in this field; in this work we will study the influence of this phenomenon on our result. The targeted emotions are joy, sadness, anger and neutral. After an analytical study of a large number of speech acoustic parameters, we chose the cepstral parameters, their first and second derivatives, the Shimmer, the Jitter and the duration of the sentence. A classifier based on a multilayer perceptron neural network to recognize emotion on the basis of the chosen feature vector that has been developed. The recognition rate could reach more than 98% in the case of an intra-speaker classification and 54.75% in inter-speaker classification. We can see the system’s dependence on speaker clearly.\",\"PeriodicalId\":197612,\"journal\":{\"name\":\"2019 International Conference on Advanced Electrical Engineering (ICAEE)\",\"volume\":\"25 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 International Conference on Advanced Electrical Engineering (ICAEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAEE47123.2019.9014809\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on Advanced Electrical Engineering (ICAEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAEE47123.2019.9014809","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The general objective of this paper is to build a system in order to automatically recognize emotion in speech. The linguistic material used is a corpus of Arabic expressive sentences phonetically balanced. The dependence of the system on speaker is an encountered problem in this field; in this work we will study the influence of this phenomenon on our result. The targeted emotions are joy, sadness, anger and neutral. After an analytical study of a large number of speech acoustic parameters, we chose the cepstral parameters, their first and second derivatives, the Shimmer, the Jitter and the duration of the sentence. A classifier based on a multilayer perceptron neural network to recognize emotion on the basis of the chosen feature vector that has been developed. The recognition rate could reach more than 98% in the case of an intra-speaker classification and 54.75% in inter-speaker classification. We can see the system’s dependence on speaker clearly.