Y. Liao, Jia-Jang Tu, Sen-Chia Chang, Chin-Hui Lee
{"title":"An enhanced minimum classification error learning framework for balancing insertion, deletion and substitution errors","authors":"Y. Liao, Jia-Jang Tu, Sen-Chia Chang, Chin-Hui Lee","doi":"10.1109/ASRU.2007.4430178","DOIUrl":null,"url":null,"abstract":"In continuous speech recognition substitution, insertion and deletion errors usually not only vary in numbers but also have different degrees of impact on optimizing a set of acoustic models. To balance their contributions to the overall error, an enhanced minimum classification error (E-MCE) learning framework is developed. The basic idea is to partition acoustic model optimization into three subtasks, i.e., minimum substitution errors (MSE), insertion errors (MIE) and deletion errors (MDE), and select/generate three corresponding sets of competing hypotheses, one for each individual sub-problem. MSE, MIE and MDE are then sequentially executed to gradually reduce the overall word error rates. Experimental results on continuous Mandarin digit recognition of five different data sets collected over various acoustic conditions have consistently shown the effectiveness of the proposed E-MCE learning framework.","PeriodicalId":371729,"journal":{"name":"2007 IEEE Workshop on Automatic Speech Recognition & Understanding (ASRU)","volume":"6 11","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2007 IEEE Workshop on Automatic Speech Recognition & Understanding (ASRU)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ASRU.2007.4430178","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
In continuous speech recognition substitution, insertion and deletion errors usually not only vary in numbers but also have different degrees of impact on optimizing a set of acoustic models. To balance their contributions to the overall error, an enhanced minimum classification error (E-MCE) learning framework is developed. The basic idea is to partition acoustic model optimization into three subtasks, i.e., minimum substitution errors (MSE), insertion errors (MIE) and deletion errors (MDE), and select/generate three corresponding sets of competing hypotheses, one for each individual sub-problem. MSE, MIE and MDE are then sequentially executed to gradually reduce the overall word error rates. Experimental results on continuous Mandarin digit recognition of five different data sets collected over various acoustic conditions have consistently shown the effectiveness of the proposed E-MCE learning framework.