{"title":"Comparison of Distance Computation Methods in Trained HMM Clustering for Huge-Scale Online Handwritten Chinese Character Recognition","authors":"KwangSeob Kim, J. Ha","doi":"10.1109/FGCNS.2008.29","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a clustering method to solve computing resource problem that may arise in very large scale on-line Chinese character recognition using HMM and structure code. The basic concept of our method is to collect HMMs that have same number of parameters, then to cluster those HMMs. In our system, the number of classes is 98,639, which makes it almost impossible to load all the models in main memory. We load only cluster centers in main memory while the individual HMM is loaded only when it is needed. We got 0.92 sec/char recognition speed and 96.03% 30-candidate recognition accuracy for Kanji, using less than 250 MB RAM for the recognition system.","PeriodicalId":370780,"journal":{"name":"2008 Second International Conference on Future Generation Communication and Networking Symposia","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 Second International Conference on Future Generation Communication and Networking Symposia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FGCNS.2008.29","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we propose a clustering method to solve computing resource problem that may arise in very large scale on-line Chinese character recognition using HMM and structure code. The basic concept of our method is to collect HMMs that have same number of parameters, then to cluster those HMMs. In our system, the number of classes is 98,639, which makes it almost impossible to load all the models in main memory. We load only cluster centers in main memory while the individual HMM is loaded only when it is needed. We got 0.92 sec/char recognition speed and 96.03% 30-candidate recognition accuracy for Kanji, using less than 250 MB RAM for the recognition system.