{"title":"基于CCBC和神经网络的语音情感识别研究","authors":"Zhiyan Han, Shuxian Lun, Jian Wang","doi":"10.1109/ICCSEE.2012.128","DOIUrl":null,"url":null,"abstract":"This paper described a novel speech emotion recognition approach aiming at improving speech emotion recognition rate. Seven discrete emotional states (anger, disgust, fear, joy, neutral, sadness, surprise) are classified throughout the work. Firstly, series preprocessing of speech signals are done. Secondly, extracting features are done, and then we consider incorporating Canonical Correlation Based on Compensation (CCBC) to cope with the mismatch between training and test set. The mismatch between training and test conditions can be simply clustered into three classes: differences of speakers, changes of recording channel and effects of noisy environment. Finally, we evaluated the system using Back-propagation Neural Networks (BPNN). Results are given using the Chinese Corpus of emotional speech synthesis database, recognition experiments show that the method is effective and high speech for emotion recognition.","PeriodicalId":132465,"journal":{"name":"2012 International Conference on Computer Science and Electronics Engineering","volume":"70 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":"{\"title\":\"A Study on Speech Emotion Recognition Based on CCBC and Neural Network\",\"authors\":\"Zhiyan Han, Shuxian Lun, Jian Wang\",\"doi\":\"10.1109/ICCSEE.2012.128\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper described a novel speech emotion recognition approach aiming at improving speech emotion recognition rate. Seven discrete emotional states (anger, disgust, fear, joy, neutral, sadness, surprise) are classified throughout the work. Firstly, series preprocessing of speech signals are done. Secondly, extracting features are done, and then we consider incorporating Canonical Correlation Based on Compensation (CCBC) to cope with the mismatch between training and test set. The mismatch between training and test conditions can be simply clustered into three classes: differences of speakers, changes of recording channel and effects of noisy environment. Finally, we evaluated the system using Back-propagation Neural Networks (BPNN). Results are given using the Chinese Corpus of emotional speech synthesis database, recognition experiments show that the method is effective and high speech for emotion recognition.\",\"PeriodicalId\":132465,\"journal\":{\"name\":\"2012 International Conference on Computer Science and Electronics Engineering\",\"volume\":\"70 1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-03-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"11\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 International Conference on Computer Science and Electronics Engineering\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCSEE.2012.128\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 International Conference on Computer Science and Electronics Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCSEE.2012.128","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Study on Speech Emotion Recognition Based on CCBC and Neural Network
This paper described a novel speech emotion recognition approach aiming at improving speech emotion recognition rate. Seven discrete emotional states (anger, disgust, fear, joy, neutral, sadness, surprise) are classified throughout the work. Firstly, series preprocessing of speech signals are done. Secondly, extracting features are done, and then we consider incorporating Canonical Correlation Based on Compensation (CCBC) to cope with the mismatch between training and test set. The mismatch between training and test conditions can be simply clustered into three classes: differences of speakers, changes of recording channel and effects of noisy environment. Finally, we evaluated the system using Back-propagation Neural Networks (BPNN). Results are given using the Chinese Corpus of emotional speech synthesis database, recognition experiments show that the method is effective and high speech for emotion recognition.