首页 > 最新文献

Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)最新文献

英文 中文
Iterated Document Content Classification 迭代文档内容分类
Chang An, H. Baird, Pingping Xiu
We report an improved methodology for training classifiers for document image content extraction, that is, the location and segmentation of regions containing handwriting, machine-printed text, photographs, blank space, etc. Our previous methods classified each individual pixel separately (rather than regions): this avoids the arbitrariness and restrictiveness that result from constraining region shapes (to, e.g., rectangles). However, this policy also allows content classes to vary frequently within small regions, often yielding areas where several content classes are mixed together. This does not reflect the way that real content is organized: typically almost all small local regions are of uniform class. This observation suggested a post-classification methodology which enforces local uniformity without imposing a restricted class of region shapes. We choose features extracted from small local regions (e.g. 4-5 pixels radius) with which we train classifiers that operate on the output of previous classifiers, guided by ground truth. This provides a sequence of post-classifiers, each trained separately on the results of the previous classifier. Experiments on a highly diverse test set of 83 document images show that this method reduces per-pixel classification errors by 23%, and it dramatically increases the occurrence of large contiguous regions of uniform class, thus providing highly usable near-solid 'masks' with which to segment the images into distinct classes. It continues to allow a wide range of complex, non-rectilinear region shapes.
我们报告了一种用于文档图像内容提取的训练分类器的改进方法,即包含手写,机器打印文本,照片,空白等的区域的定位和分割。我们之前的方法分别对每个单独的像素(而不是区域)进行分类:这避免了由于约束区域形状(例如矩形)而导致的随意性和限制性。但是,此策略还允许内容类在小区域内频繁变化,通常会产生几个内容类混合在一起的区域。这并没有反映真实内容的组织方式:通常几乎所有小的局部区域都是统一的类。这一观察提出了一种后分类方法,该方法在不强加区域形状的限制类别的情况下强制局部一致性。我们选择从小的局部区域(例如4-5个像素半径)提取的特征,我们用这些特征来训练分类器,这些分类器在ground truth的指导下对先前分类器的输出进行操作。这提供了一系列后分类器,每个后分类器分别在前一个分类器的结果上进行训练。在83张高度多样化的文档图像测试集上进行的实验表明,该方法将每像素的分类误差降低了23%,并且显著增加了统一类别的大型连续区域的出现,从而提供了高度可用的近固体“掩模”,用于将图像分割为不同的类别。它继续允许广泛的复杂,非直线区域形状。
{"title":"Iterated Document Content Classification","authors":"Chang An, H. Baird, Pingping Xiu","doi":"10.1109/ICDAR.2007.148","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.148","url":null,"abstract":"We report an improved methodology for training classifiers for document image content extraction, that is, the location and segmentation of regions containing handwriting, machine-printed text, photographs, blank space, etc. Our previous methods classified each individual pixel separately (rather than regions): this avoids the arbitrariness and restrictiveness that result from constraining region shapes (to, e.g., rectangles). However, this policy also allows content classes to vary frequently within small regions, often yielding areas where several content classes are mixed together. This does not reflect the way that real content is organized: typically almost all small local regions are of uniform class. This observation suggested a post-classification methodology which enforces local uniformity without imposing a restricted class of region shapes. We choose features extracted from small local regions (e.g. 4-5 pixels radius) with which we train classifiers that operate on the output of previous classifiers, guided by ground truth. This provides a sequence of post-classifiers, each trained separately on the results of the previous classifier. Experiments on a highly diverse test set of 83 document images show that this method reduces per-pixel classification errors by 23%, and it dramatically increases the occurrence of large contiguous regions of uniform class, thus providing highly usable near-solid 'masks' with which to segment the images into distinct classes. It continues to allow a wide range of complex, non-rectilinear region shapes.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130404835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Skew Detection for Chinese Handwriting by Horizontal Stroke Histogram 基于水平笔画直方图的汉字书写歪斜检测
Tong-Hua Su, Tian-Wen Zhang, Hu-Jie Huang, Yu Zhou
This paper proposes a skew detection method for real Chinese handwritten documents. After analyzing the characteristics of Chinese characters, it utilizes the horizontal stroke histogram. Its accuracy, ability to increase the recall rate of text line separation, and CPU time consuming are investigated using 853 real Chinese handwritten documents. The results show that: 1) the method can identify 98.83% of the skew angles within one degree, with an improvement of 8.44% than Wigner-Ville distribution (WVD) method; 2) when incorporated into text line separation, the recall rate has an improvement of 2.54% than WVD method; 3) the method only consumes one-twentieth of WVD method on the same test environment.
提出了一种针对真实中文手写文档的倾斜检测方法。在分析了汉字的特点后,采用了横画直方图。以853份真实中文手写文档为例,考察了该方法的准确率、提高文本行分离查全率的能力和CPU耗时。结果表明:1)该方法能在1度内识别出98.83%的倾斜角,比WVD方法提高了8.44%;2)结合文本行分离,召回率比WVD方法提高了2.54%;3)在相同的测试环境下,该方法的能耗仅为WVD法的二十分之一。
{"title":"Skew Detection for Chinese Handwriting by Horizontal Stroke Histogram","authors":"Tong-Hua Su, Tian-Wen Zhang, Hu-Jie Huang, Yu Zhou","doi":"10.1109/ICDAR.2007.233","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.233","url":null,"abstract":"This paper proposes a skew detection method for real Chinese handwritten documents. After analyzing the characteristics of Chinese characters, it utilizes the horizontal stroke histogram. Its accuracy, ability to increase the recall rate of text line separation, and CPU time consuming are investigated using 853 real Chinese handwritten documents. The results show that: 1) the method can identify 98.83% of the skew angles within one degree, with an improvement of 8.44% than Wigner-Ville distribution (WVD) method; 2) when incorporated into text line separation, the recall rate has an improvement of 2.54% than WVD method; 3) the method only consumes one-twentieth of WVD method on the same test environment.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117177738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Deriving Symbol Dependent Edit Weights for Text Correction_The Use of Error Dictionaries 为文本纠错派生与符号相关的编辑权重——错误字典的使用
Christoph Ringlstetter, Ulrich Reffle, Annette Gotscharek, K. Schulz
Most systems for correcting errors in texts make use of specific word distance measures such as the Levenshtein distance. In many experiments it has been shown that correction accuracy is improved when using edit weights that depend on the particular symbols of the edit operation. However, most proposed approaches so far rely on high amounts of training data where errors and their corrections are collected. In practice, the preparation of suitable ground truth data is often too costly, which means that uniform edit costs are used. In this paper we evaluate approaches for deriving symbol dependent edit weights that do not need any ground truth training data, comparing them with methods based on ground truth training. We suggest a new approach where special error dictionaries are used to estimate weights. The method is simple and very efficient, needing one pass of the document to be corrected. Our experiments with different OCR systems and textual data show that the method consistently improves correction accuracy in a significant way, often leading to results comparable to those achieved with ground truth training.
大多数文本纠错系统都使用特定的单词距离度量,如Levenshtein距离。在许多实验中已经证明,当使用依赖于编辑操作的特定符号的编辑权重时,校正精度得到了提高。然而,到目前为止,大多数提出的方法都依赖于大量的训练数据,其中收集了错误及其纠正。在实践中,准备合适的地面真值数据往往成本过高,这意味着使用统一的编辑成本。在本文中,我们评估了不需要任何基础真值训练数据的符号相关编辑权的推导方法,并将它们与基于基础真值训练的方法进行了比较。我们提出了一种新的方法,使用特殊的错误字典来估计权重。该方法简单有效,只需通过一次文件即可进行校正。我们对不同OCR系统和文本数据的实验表明,该方法持续地显著提高了校正精度,通常导致与地面真值训练相媲美的结果。
{"title":"Deriving Symbol Dependent Edit Weights for Text Correction_The Use of Error Dictionaries","authors":"Christoph Ringlstetter, Ulrich Reffle, Annette Gotscharek, K. Schulz","doi":"10.1109/ICDAR.2007.99","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.99","url":null,"abstract":"Most systems for correcting errors in texts make use of specific word distance measures such as the Levenshtein distance. In many experiments it has been shown that correction accuracy is improved when using edit weights that depend on the particular symbols of the edit operation. However, most proposed approaches so far rely on high amounts of training data where errors and their corrections are collected. In practice, the preparation of suitable ground truth data is often too costly, which means that uniform edit costs are used. In this paper we evaluate approaches for deriving symbol dependent edit weights that do not need any ground truth training data, comparing them with methods based on ground truth training. We suggest a new approach where special error dictionaries are used to estimate weights. The method is simple and very efficient, needing one pass of the document to be corrected. Our experiments with different OCR systems and textual data show that the method consistently improves correction accuracy in a significant way, often leading to results comparable to those achieved with ground truth training.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130971510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Handwritten Chinese Character Recognition Using Modified LDA and Kernel FDA 基于改进LDA和核FDA的手写汉字识别
Duanduan Yang, Lianwen Jin
The effectiveness of kernel fisher discrimination analysis (KFDA) has been demonstrated by many pattern recognition applications. However, due to the large size of Gram matrix to be trained, how to use KFDA to solve large vocabulary pattern recognition task such as Chinese Characters recognition is still a challenging problem. In this paper, a two-stage KFDA approach is presented for handwritten Chinese character recognition. In the first stage, a new modified linear discriminant analysis method is developed to get the recognition candidates. In the second stage, KFDA is used to determine the final recognition result. Experiments on 1034 categories of Chinese character from 120 sets of handwriting samples shows that a 3.37% improvement of recognition rate is obtained, which suggests the effectiveness of the proposed method.
核费雪判别分析(KFDA)的有效性已被许多模式识别应用所证明。然而,由于待训练的Gram矩阵规模较大,如何利用KFDA解决像汉字识别这样的大词汇量模式识别任务仍然是一个具有挑战性的问题。本文提出了一种两阶段KFDA的手写体汉字识别方法。在第一阶段,提出了一种新的改进的线性判别分析方法来获得识别候选者。在第二阶段,由KFDA确定最终的识别结果。对120组手写样本中的1034类汉字进行了实验,结果表明,该方法的识别率提高了3.37%,表明了该方法的有效性。
{"title":"Handwritten Chinese Character Recognition Using Modified LDA and Kernel FDA","authors":"Duanduan Yang, Lianwen Jin","doi":"10.1109/ICDAR.2007.128","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.128","url":null,"abstract":"The effectiveness of kernel fisher discrimination analysis (KFDA) has been demonstrated by many pattern recognition applications. However, due to the large size of Gram matrix to be trained, how to use KFDA to solve large vocabulary pattern recognition task such as Chinese Characters recognition is still a challenging problem. In this paper, a two-stage KFDA approach is presented for handwritten Chinese character recognition. In the first stage, a new modified linear discriminant analysis method is developed to get the recognition candidates. In the second stage, KFDA is used to determine the final recognition result. Experiments on 1034 categories of Chinese character from 120 sets of handwriting samples shows that a 3.37% improvement of recognition rate is obtained, which suggests the effectiveness of the proposed method.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132559338","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Combination of OCR Engines for Page Segmentation Based on Performance Evaluation 基于性能评估的组合OCR引擎页面分割
Miquel A. Ferrer, Ernest Valveny
In this paper we present a method to improve the performance of individual page segmentation engines based on the combination of the output of several engines. The rules of combination are designed after analyzing the results of each individual method. This analysis is performed using a performance evaluation framework that aims at characterizing each method according to its strengths and weaknesses rather than computing a single performance measure telling which is the "best" segmentation method.
在本文中,我们提出了一种基于多个引擎输出的组合来提高单个页面分割引擎性能的方法。在分析了每种方法的结果后,设计了组合规则。此分析使用性能评估框架执行,该框架旨在根据每种方法的优点和缺点来描述其特征,而不是计算单个性能度量来告诉哪个是“最佳”分割方法。
{"title":"Combination of OCR Engines for Page Segmentation Based on Performance Evaluation","authors":"Miquel A. Ferrer, Ernest Valveny","doi":"10.1109/ICDAR.2007.83","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.83","url":null,"abstract":"In this paper we present a method to improve the performance of individual page segmentation engines based on the combination of the output of several engines. The rules of combination are designed after analyzing the results of each individual method. This analysis is performed using a performance evaluation framework that aims at characterizing each method according to its strengths and weaknesses rather than computing a single performance measure telling which is the \"best\" segmentation method.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132733743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Two Stage Recognition Scheme for Handwritten Tamil Characters 手写泰米尔字符的两阶段识别方案
U. Bhattacharya, S. Ghosh, S. K. Parui
India is a multilingual multiscript country with more than 18 languages and 10 different major scripts. Not enough research work towards recognition of handwritten characters of these Indian scripts has been done. Tamil, an official as well as popular script of the southern part of India, Singapore, Malaysia, and Sri Lanka has a large character set which includes many compound characters. Only a few works towards handwriting recognition of this large character set has been reported in the literature. Recently, HP Labs India developed a database of handwritten Tamil characters. In the present paper, we describe an off-line recognition approach based on this database. The proposed method consists of two stages. In the first stage, we apply an unsupervised clustering method to create a smaller number of groups of handwritten Tamil character classes. In the second stage, we consider a supervised classification technique in each of these smaller groups for final recognition. The features considered in the two stages are different. The proposed two-stage recognition scheme provided acceptable classification accuracies on both the training and test sets of the present database.
印度是一个多语言多文字的国家,有超过18种语言和10种不同的主要文字。对这些印度文字的手写体的识别研究工作还不够。泰米尔语是印度南部、新加坡、马来西亚和斯里兰卡的一种官方和流行的文字,它有一个很大的字符集,其中包括许多复合字。只有少数的工作对这种大字符集的手写识别已在文献中报道。最近,惠普印度实验室开发了一个手写泰米尔文字数据库。在本文中,我们描述了一种基于该数据库的离线识别方法。该方法分为两个阶段。在第一阶段,我们应用无监督聚类方法来创建较少数量的手写泰米尔字符类组。在第二阶段,我们考虑在每个较小的组中使用监督分类技术进行最终识别。这两个阶段所考虑的特性是不同的。提出的两阶段识别方案在现有数据库的训练集和测试集上都提供了可接受的分类精度。
{"title":"A Two Stage Recognition Scheme for Handwritten Tamil Characters","authors":"U. Bhattacharya, S. Ghosh, S. K. Parui","doi":"10.1109/ICDAR.2007.37","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.37","url":null,"abstract":"India is a multilingual multiscript country with more than 18 languages and 10 different major scripts. Not enough research work towards recognition of handwritten characters of these Indian scripts has been done. Tamil, an official as well as popular script of the southern part of India, Singapore, Malaysia, and Sri Lanka has a large character set which includes many compound characters. Only a few works towards handwriting recognition of this large character set has been reported in the literature. Recently, HP Labs India developed a database of handwritten Tamil characters. In the present paper, we describe an off-line recognition approach based on this database. The proposed method consists of two stages. In the first stage, we apply an unsupervised clustering method to create a smaller number of groups of handwritten Tamil character classes. In the second stage, we consider a supervised classification technique in each of these smaller groups for final recognition. The features considered in the two stages are different. The proposed two-stage recognition scheme provided acceptable classification accuracies on both the training and test sets of the present database.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131686054","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 58
A hybrid approach for off-line Arabic handwriting recognition based on a Planar Hidden Markov modeling 一种基于平面隐马尔可夫建模的离线阿拉伯手写识别混合方法
Sameh Masmoudi Touj, N. Amara, H. Amiri
A novel approach for the Arabic handwriting recognition is presented. The use of a planar hidden Markov model (PHMM) has permitted to split the Arabic script into five homogeneous horizontal regions. Each region was described by a 1D-HMM. This modeling is based on different levels of segmentation: horizontal, natural and vertical. Both holistic and analytical approaches have been tested for the description of the median band of the Arabic writing. We show finally that a hybrid approach conducted to the improvement of the whole system performances.
提出了一种新的阿拉伯语手写识别方法。平面隐马尔可夫模型(PHMM)的使用允许将阿拉伯文字划分为五个均匀的水平区域。每个区域用1D-HMM来描述。这种建模基于不同层次的分割:水平、自然和垂直。对于阿拉伯文字的中间带的描述,整体性和分析性两种方法都进行了测试。最后,我们证明了一种混合方法可以改善整个系统的性能。
{"title":"A hybrid approach for off-line Arabic handwriting recognition based on a Planar Hidden Markov modeling","authors":"Sameh Masmoudi Touj, N. Amara, H. Amiri","doi":"10.1109/ICDAR.2007.14","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.14","url":null,"abstract":"A novel approach for the Arabic handwriting recognition is presented. The use of a planar hidden Markov model (PHMM) has permitted to split the Arabic script into five homogeneous horizontal regions. Each region was described by a 1D-HMM. This modeling is based on different levels of segmentation: horizontal, natural and vertical. Both holistic and analytical approaches have been tested for the description of the median band of the Arabic writing. We show finally that a hybrid approach conducted to the improvement of the whole system performances.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134096675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Hidden Markov Models for Online Handwritten Tamil Word Recognition 隐马尔可夫模型用于在线手写泰米尔语单词识别
A. Bharath, S. Madhvanath
Hidden Markov models (HMM) have long been a popular choice for Western cursive handwriting recognition following their success in speech recognition. Even for the recognition of Oriental scripts such as Chinese, Japanese and Korean, hidden Markov models are increasingly being used to model substrokes of characters. However, when it comes to Indie script recognition, the published work employing HMMs is limited, and generally focussed on isolated character recognition. In this effort, a data-driven HMM-based online handwritten word recognition system for Tamil, an Indie script, is proposed. The accuracies obtained ranged from 98% to 92.2% with different lexicon sizes (IK to 20 K words). These initial results are promising and warrant further research in this direction. The results are also encouraging to explore possibilities for adopting the approach to other Indie scripts as well.
隐马尔可夫模型(HMM)在语音识别领域取得成功后,一直是西方手写体识别领域的热门选择。即使是汉字、日文、韩文等东方文字的识别,也越来越多地使用隐马尔可夫模型来模拟汉字的笔划。然而,当涉及到独立脚本识别时,使用hmm的出版作品是有限的,并且通常集中在孤立的字符识别上。在此基础上,提出了一种基于数据驱动的独立语言泰米尔语的在线手写单词识别系统。在不同的词汇量(IK到20k单词)下,准确率在98%到92.2%之间。这些初步结果是有希望的,值得在这个方向上进一步研究。结果也鼓励我们探索将这种方法应用于其他独立脚本的可能性。
{"title":"Hidden Markov Models for Online Handwritten Tamil Word Recognition","authors":"A. Bharath, S. Madhvanath","doi":"10.1109/ICDAR.2007.131","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.131","url":null,"abstract":"Hidden Markov models (HMM) have long been a popular choice for Western cursive handwriting recognition following their success in speech recognition. Even for the recognition of Oriental scripts such as Chinese, Japanese and Korean, hidden Markov models are increasingly being used to model substrokes of characters. However, when it comes to Indie script recognition, the published work employing HMMs is limited, and generally focussed on isolated character recognition. In this effort, a data-driven HMM-based online handwritten word recognition system for Tamil, an Indie script, is proposed. The accuracies obtained ranged from 98% to 92.2% with different lexicon sizes (IK to 20 K words). These initial results are promising and warrant further research in this direction. The results are also encouraging to explore possibilities for adopting the approach to other Indie scripts as well.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133060010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 68
Pàtrà: A Novel Document Architecture for Integrating Handwriting with Audio-Visual Information Pàtrà:一种集成手写和视听信息的新型文档体系结构
Gaurav Harit, V. Mankar, S. Chaudhury
In this paper we present Patra - an integrated document architecture which incorporates handwritten illustrations captured and rendered in a temporal fashion synchronized with audio, video, text, and image data. The architecture of Patra permits non-linear growth in the form of multiple hierarchically organized play streams. Semantic metadata is also an integral part of Patra which serves a useful purpose of organizing such documents in a collection. We have developed an email application in which the users are provided with an authoring and rendering environment to compose, view, and reply to messages in the form of Patra.
在本文中,我们介绍了Patra——一个集成的文档架构,它包含了以与音频、视频、文本和图像数据同步的时间方式捕获和渲染的手写插图。Patra的架构允许以多种层次组织的游戏流的形式非线性增长。语义元数据也是Patra的一个组成部分,它有助于在集合中组织这样的文档。我们开发了一个电子邮件应用程序,为用户提供了一个创作和呈现环境,以便以Patra的形式编写、查看和回复消息。
{"title":"Pàtrà: A Novel Document Architecture for Integrating Handwriting with Audio-Visual Information","authors":"Gaurav Harit, V. Mankar, S. Chaudhury","doi":"10.1109/ICDAR.2007.204","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.204","url":null,"abstract":"In this paper we present Patra - an integrated document architecture which incorporates handwritten illustrations captured and rendered in a temporal fashion synchronized with audio, video, text, and image data. The architecture of Patra permits non-linear growth in the form of multiple hierarchically organized play streams. Semantic metadata is also an integral part of Patra which serves a useful purpose of organizing such documents in a collection. We have developed an email application in which the users are provided with an authoring and rendering environment to compose, view, and reply to messages in the form of Patra.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132187379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
PRAAD: Preprocessing and Analysis Tool for Arabic Ancient Documents PRAAD:阿拉伯语古代文献预处理与分析工具
Wafa Boussellaa, Abderrazak Zahour, B. Taconet, A. Alimi, A. BenAbdelhafid
This paper presents the new system PRAAD for preprocessing and analysis of Arabic historical documents. It is composed of two important parts: pre-processing and analysis of ancient documents. After digitization, the color or greyscale ancient documents images are distorted by the presence of strong background artefacts such as scan optical blur and noise, show-through and bleed-through effects and spots. In order to preserve and exploit this cultural heritage documents, we intend to create efficient tool that achieves restoration, binarisation, and analyses the document layout. The developed tool is done by adapting our expertise in document image processing of Arabic ancient documents, printed or manuscripts. The different functions of PRAAD system are tested on a set of Arabic ancient documents from the national library and the National Archives of Tunisia.
本文介绍了一种新的阿拉伯语历史文献预处理分析系统PRAAD。它由两个重要部分组成:古代文献的预处理和分析。数字化后的彩色或灰度古代文献图像由于扫描光学模糊和噪声、透显和透透效果和斑点等强背景伪影的存在而发生畸变。为了保护和利用这些文化遗产文件,我们打算创建一个有效的工具来实现文件布局的修复、二值化和分析。开发的工具是通过调整我们在阿拉伯古代文件,印刷或手稿的文件图像处理方面的专业知识来完成的。以突尼斯国家图书馆和国家档案馆的一组阿拉伯文古文献为例,对PRAAD系统的不同功能进行了测试。
{"title":"PRAAD: Preprocessing and Analysis Tool for Arabic Ancient Documents","authors":"Wafa Boussellaa, Abderrazak Zahour, B. Taconet, A. Alimi, A. BenAbdelhafid","doi":"10.1109/ICDAR.2007.209","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.209","url":null,"abstract":"This paper presents the new system PRAAD for preprocessing and analysis of Arabic historical documents. It is composed of two important parts: pre-processing and analysis of ancient documents. After digitization, the color or greyscale ancient documents images are distorted by the presence of strong background artefacts such as scan optical blur and noise, show-through and bleed-through effects and spots. In order to preserve and exploit this cultural heritage documents, we intend to create efficient tool that achieves restoration, binarisation, and analyses the document layout. The developed tool is done by adapting our expertise in document image processing of Arabic ancient documents, printed or manuscripts. The different functions of PRAAD system are tested on a set of Arabic ancient documents from the national library and the National Archives of Tunisia.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"10 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114038748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1