{"title":"大规模在线手写体汉字识别训练HMM聚类距离计算方法比较","authors":"KwangSeob Kim, J. Ha","doi":"10.1109/FGCNS.2008.29","DOIUrl":null,"url":null,"abstract":"In this paper, we propose a clustering method to solve computing resource problem that may arise in very large scale on-line Chinese character recognition using HMM and structure code. The basic concept of our method is to collect HMMs that have same number of parameters, then to cluster those HMMs. In our system, the number of classes is 98,639, which makes it almost impossible to load all the models in main memory. We load only cluster centers in main memory while the individual HMM is loaded only when it is needed. We got 0.92 sec/char recognition speed and 96.03% 30-candidate recognition accuracy for Kanji, using less than 250 MB RAM for the recognition system.","PeriodicalId":370780,"journal":{"name":"2008 Second International Conference on Future Generation Communication and Networking Symposia","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-12-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Comparison of Distance Computation Methods in Trained HMM Clustering for Huge-Scale Online Handwritten Chinese Character Recognition\",\"authors\":\"KwangSeob Kim, J. Ha\",\"doi\":\"10.1109/FGCNS.2008.29\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we propose a clustering method to solve computing resource problem that may arise in very large scale on-line Chinese character recognition using HMM and structure code. The basic concept of our method is to collect HMMs that have same number of parameters, then to cluster those HMMs. In our system, the number of classes is 98,639, which makes it almost impossible to load all the models in main memory. We load only cluster centers in main memory while the individual HMM is loaded only when it is needed. We got 0.92 sec/char recognition speed and 96.03% 30-candidate recognition accuracy for Kanji, using less than 250 MB RAM for the recognition system.\",\"PeriodicalId\":370780,\"journal\":{\"name\":\"2008 Second International Conference on Future Generation Communication and Networking Symposia\",\"volume\":\"27 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-12-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 Second International Conference on Future Generation Communication and Networking Symposia\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/FGCNS.2008.29\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 Second International Conference on Future Generation Communication and Networking Symposia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FGCNS.2008.29","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Comparison of Distance Computation Methods in Trained HMM Clustering for Huge-Scale Online Handwritten Chinese Character Recognition
In this paper, we propose a clustering method to solve computing resource problem that may arise in very large scale on-line Chinese character recognition using HMM and structure code. The basic concept of our method is to collect HMMs that have same number of parameters, then to cluster those HMMs. In our system, the number of classes is 98,639, which makes it almost impossible to load all the models in main memory. We load only cluster centers in main memory while the individual HMM is loaded only when it is needed. We got 0.92 sec/char recognition speed and 96.03% 30-candidate recognition accuracy for Kanji, using less than 250 MB RAM for the recognition system.