{"title":"SVM-MLP-PNN Classifiers on Speech Emotion Recognition Field - A Comparative Study","authors":"Theodoros Iliou, C. Anagnostopoulos","doi":"10.1109/ICDT.2010.8","DOIUrl":null,"url":null,"abstract":"In this paper, we present a comparative analysisof three classifiers for speech signal emotion recognition.Recognition was performed on emotional Berlin Database.This work focuses on speaker and utterance (phrase)dependent and independent framework. One hundred thirtythree (133) sound/speech features were extracted from Pitch,Mel Frequency Cepstral Coefficients, Energy and Formantsand were evaluated in order to create a feature set sufficient todiscriminate between seven emotions in acted speech. A set of26 features was selected by statistical method and MultilayerPercepton, Probabilistic Neural Networks and Support VectorMachine were used for the Emotion Classification at sevenclasses: anger, happiness, anxiety/fear, sadness, boredom,disgust and neutral. In speaker dependent framework,Probabilistic Neural Network classifier reached very highaccuracy of 94%, whereas in speaker independent framework,Support Vector Machine classification reached the bestaccuracy of 80%. The results of numerical experiments aregiven and discussed in the paper.","PeriodicalId":322589,"journal":{"name":"2010 Fifth International Conference on Digital Telecommunications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"24","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 Fifth International Conference on Digital Telecommunications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDT.2010.8","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 24
Abstract
In this paper, we present a comparative analysisof three classifiers for speech signal emotion recognition.Recognition was performed on emotional Berlin Database.This work focuses on speaker and utterance (phrase)dependent and independent framework. One hundred thirtythree (133) sound/speech features were extracted from Pitch,Mel Frequency Cepstral Coefficients, Energy and Formantsand were evaluated in order to create a feature set sufficient todiscriminate between seven emotions in acted speech. A set of26 features was selected by statistical method and MultilayerPercepton, Probabilistic Neural Networks and Support VectorMachine were used for the Emotion Classification at sevenclasses: anger, happiness, anxiety/fear, sadness, boredom,disgust and neutral. In speaker dependent framework,Probabilistic Neural Network classifier reached very highaccuracy of 94%, whereas in speaker independent framework,Support Vector Machine classification reached the bestaccuracy of 80%. The results of numerical experiments aregiven and discussed in the paper.