首页 > 最新文献

Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)最新文献

英文 中文
Iterated Document Content Classification 迭代文档内容分类
Chang An, H. Baird, Pingping Xiu
We report an improved methodology for training classifiers for document image content extraction, that is, the location and segmentation of regions containing handwriting, machine-printed text, photographs, blank space, etc. Our previous methods classified each individual pixel separately (rather than regions): this avoids the arbitrariness and restrictiveness that result from constraining region shapes (to, e.g., rectangles). However, this policy also allows content classes to vary frequently within small regions, often yielding areas where several content classes are mixed together. This does not reflect the way that real content is organized: typically almost all small local regions are of uniform class. This observation suggested a post-classification methodology which enforces local uniformity without imposing a restricted class of region shapes. We choose features extracted from small local regions (e.g. 4-5 pixels radius) with which we train classifiers that operate on the output of previous classifiers, guided by ground truth. This provides a sequence of post-classifiers, each trained separately on the results of the previous classifier. Experiments on a highly diverse test set of 83 document images show that this method reduces per-pixel classification errors by 23%, and it dramatically increases the occurrence of large contiguous regions of uniform class, thus providing highly usable near-solid 'masks' with which to segment the images into distinct classes. It continues to allow a wide range of complex, non-rectilinear region shapes.
我们报告了一种用于文档图像内容提取的训练分类器的改进方法,即包含手写,机器打印文本,照片,空白等的区域的定位和分割。我们之前的方法分别对每个单独的像素(而不是区域)进行分类:这避免了由于约束区域形状(例如矩形)而导致的随意性和限制性。但是,此策略还允许内容类在小区域内频繁变化,通常会产生几个内容类混合在一起的区域。这并没有反映真实内容的组织方式:通常几乎所有小的局部区域都是统一的类。这一观察提出了一种后分类方法,该方法在不强加区域形状的限制类别的情况下强制局部一致性。我们选择从小的局部区域(例如4-5个像素半径)提取的特征,我们用这些特征来训练分类器,这些分类器在ground truth的指导下对先前分类器的输出进行操作。这提供了一系列后分类器,每个后分类器分别在前一个分类器的结果上进行训练。在83张高度多样化的文档图像测试集上进行的实验表明,该方法将每像素的分类误差降低了23%,并且显著增加了统一类别的大型连续区域的出现,从而提供了高度可用的近固体“掩模”,用于将图像分割为不同的类别。它继续允许广泛的复杂,非直线区域形状。
{"title":"Iterated Document Content Classification","authors":"Chang An, H. Baird, Pingping Xiu","doi":"10.1109/ICDAR.2007.148","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.148","url":null,"abstract":"We report an improved methodology for training classifiers for document image content extraction, that is, the location and segmentation of regions containing handwriting, machine-printed text, photographs, blank space, etc. Our previous methods classified each individual pixel separately (rather than regions): this avoids the arbitrariness and restrictiveness that result from constraining region shapes (to, e.g., rectangles). However, this policy also allows content classes to vary frequently within small regions, often yielding areas where several content classes are mixed together. This does not reflect the way that real content is organized: typically almost all small local regions are of uniform class. This observation suggested a post-classification methodology which enforces local uniformity without imposing a restricted class of region shapes. We choose features extracted from small local regions (e.g. 4-5 pixels radius) with which we train classifiers that operate on the output of previous classifiers, guided by ground truth. This provides a sequence of post-classifiers, each trained separately on the results of the previous classifier. Experiments on a highly diverse test set of 83 document images show that this method reduces per-pixel classification errors by 23%, and it dramatically increases the occurrence of large contiguous regions of uniform class, thus providing highly usable near-solid 'masks' with which to segment the images into distinct classes. It continues to allow a wide range of complex, non-rectilinear region shapes.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130404835","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Elastic Matching of Online Handwritten Tamil and Telugu Scripts Using Local Features 使用本地特征的在线手写泰米尔语和泰卢固语脚本的弹性匹配
Prasanth, Jagadeesh Babu, Raghunath Sharma, Prabhakara Rao, Dinesh Mandalapu, L. Prasanth, V. Jagadeesh Babu, R. Raghunath, Sharma Dinesh, M. G. V. P. Rao
This paper describes character based elastic matching using local features for recognizing online handwritten data. Dynamic time warping (DTW) has been used with four different feature sets: x-y features, shape context (SC) and tangent angle (TA) features, generalized shape context feature (GSC) and the fourth set containing x-y, normalized first and second derivatives and curvature features. Nearest neighborhood classifier with DTW distance was used as the classifier. In comparison, the SC and TA feature set was found to be the slowest and the fourth set was best among all in the recognition rate. The results have been compiled for the online handwritten Tamil and Telugu data. On Telugu data we obtained an accuracy of 90.6% with a speed of 0.166 symbols/sec. To increase the speed we have proposed a 2-stage recognition scheme using which we obtained accuracy of 89.77% but with a speed of 3.977 symbols/sec.
本文描述了一种基于字符的弹性匹配方法,利用局部特征识别在线手写数据。动态时间翘曲(DTW)被用于四种不同的特征集:x-y特征、形状上下文(SC)和切角(TA)特征、广义形状上下文特征(GSC)和包含x-y、标准化一阶导数和二阶导数以及曲率特征的第四集。采用DTW距离的最近邻分类器作为分类器。对比发现,SC和TA特征集的识别率是最慢的,而第四特征集的识别率是最好的。结果已汇编为在线手写泰米尔语和泰卢固语数据。在泰卢固语数据上,我们以0.166个符号/秒的速度获得了90.6%的准确率。为了提高速度,我们提出了一种两阶段识别方案,使用该方案我们获得了89.77%的准确率,但速度为3.977个符号/秒。
{"title":"Elastic Matching of Online Handwritten Tamil and Telugu Scripts Using Local Features","authors":"Prasanth, Jagadeesh Babu, Raghunath Sharma, Prabhakara Rao, Dinesh Mandalapu, L. Prasanth, V. Jagadeesh Babu, R. Raghunath, Sharma Dinesh, M. G. V. P. Rao","doi":"10.1109/ICDAR.2007.106","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.106","url":null,"abstract":"This paper describes character based elastic matching using local features for recognizing online handwritten data. Dynamic time warping (DTW) has been used with four different feature sets: x-y features, shape context (SC) and tangent angle (TA) features, generalized shape context feature (GSC) and the fourth set containing x-y, normalized first and second derivatives and curvature features. Nearest neighborhood classifier with DTW distance was used as the classifier. In comparison, the SC and TA feature set was found to be the slowest and the fourth set was best among all in the recognition rate. The results have been compiled for the online handwritten Tamil and Telugu data. On Telugu data we obtained an accuracy of 90.6% with a speed of 0.166 symbols/sec. To increase the speed we have proposed a 2-stage recognition scheme using which we obtained accuracy of 89.77% but with a speed of 3.977 symbols/sec.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129983384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 58
Deriving Symbol Dependent Edit Weights for Text Correction_The Use of Error Dictionaries 为文本纠错派生与符号相关的编辑权重——错误字典的使用
Christoph Ringlstetter, Ulrich Reffle, Annette Gotscharek, K. Schulz
Most systems for correcting errors in texts make use of specific word distance measures such as the Levenshtein distance. In many experiments it has been shown that correction accuracy is improved when using edit weights that depend on the particular symbols of the edit operation. However, most proposed approaches so far rely on high amounts of training data where errors and their corrections are collected. In practice, the preparation of suitable ground truth data is often too costly, which means that uniform edit costs are used. In this paper we evaluate approaches for deriving symbol dependent edit weights that do not need any ground truth training data, comparing them with methods based on ground truth training. We suggest a new approach where special error dictionaries are used to estimate weights. The method is simple and very efficient, needing one pass of the document to be corrected. Our experiments with different OCR systems and textual data show that the method consistently improves correction accuracy in a significant way, often leading to results comparable to those achieved with ground truth training.
大多数文本纠错系统都使用特定的单词距离度量,如Levenshtein距离。在许多实验中已经证明,当使用依赖于编辑操作的特定符号的编辑权重时,校正精度得到了提高。然而,到目前为止,大多数提出的方法都依赖于大量的训练数据,其中收集了错误及其纠正。在实践中,准备合适的地面真值数据往往成本过高,这意味着使用统一的编辑成本。在本文中,我们评估了不需要任何基础真值训练数据的符号相关编辑权的推导方法,并将它们与基于基础真值训练的方法进行了比较。我们提出了一种新的方法,使用特殊的错误字典来估计权重。该方法简单有效,只需通过一次文件即可进行校正。我们对不同OCR系统和文本数据的实验表明,该方法持续地显著提高了校正精度,通常导致与地面真值训练相媲美的结果。
{"title":"Deriving Symbol Dependent Edit Weights for Text Correction_The Use of Error Dictionaries","authors":"Christoph Ringlstetter, Ulrich Reffle, Annette Gotscharek, K. Schulz","doi":"10.1109/ICDAR.2007.99","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.99","url":null,"abstract":"Most systems for correcting errors in texts make use of specific word distance measures such as the Levenshtein distance. In many experiments it has been shown that correction accuracy is improved when using edit weights that depend on the particular symbols of the edit operation. However, most proposed approaches so far rely on high amounts of training data where errors and their corrections are collected. In practice, the preparation of suitable ground truth data is often too costly, which means that uniform edit costs are used. In this paper we evaluate approaches for deriving symbol dependent edit weights that do not need any ground truth training data, comparing them with methods based on ground truth training. We suggest a new approach where special error dictionaries are used to estimate weights. The method is simple and very efficient, needing one pass of the document to be corrected. Our experiments with different OCR systems and textual data show that the method consistently improves correction accuracy in a significant way, often leading to results comparable to those achieved with ground truth training.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130971510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Off-line Signature Verification System Performance against Image Acquisition Resolution 针对图像采集分辨率的离线签名验证系统性能
Pub Date : 2007-09-23 DOI: 10.1109/ICDAR.2007.4377032
J. Vargas-Bonilla, M. A. Ferrer-Ballester, C. Travieso-González, J. B. Alonso
The effect of changing the image resolution over an off-line signature verification system performance is analyzed. The geometrical features used for the system analyzed in this paper are based on two vectors which represent the envelope and the interior stroke distribution in polar and Cartesian coordinates. Image resolution is progressively diminished from an initial 600 ppp resolution till 45 ppp. The robustness of the analyzed system for random and simple forgeries is tested out with a hidden Markov model. The results show that 150 ppp offers a good trade-off between performance and image resolution for static features.
分析了改变图像分辨率对离线签名验证系统性能的影响。本文所分析的系统的几何特征是基于在极坐标和直角坐标下表示包络和内部行程分布的两个向量。图像分辨率从最初的600ppp分辨率逐渐降低到45ppp。利用隐马尔可夫模型验证了系统对随机伪造和简单伪造的鲁棒性。结果表明,150ppp在静态特征的性能和图像分辨率之间提供了一个很好的权衡。
{"title":"Off-line Signature Verification System Performance against Image Acquisition Resolution","authors":"J. Vargas-Bonilla, M. A. Ferrer-Ballester, C. Travieso-González, J. B. Alonso","doi":"10.1109/ICDAR.2007.4377032","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.4377032","url":null,"abstract":"The effect of changing the image resolution over an off-line signature verification system performance is analyzed. The geometrical features used for the system analyzed in this paper are based on two vectors which represent the envelope and the interior stroke distribution in polar and Cartesian coordinates. Image resolution is progressively diminished from an initial 600 ppp resolution till 45 ppp. The robustness of the analyzed system for random and simple forgeries is tested out with a hidden Markov model. The results show that 150 ppp offers a good trade-off between performance and image resolution for static features.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131096448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Machine Dating of Handwritten Manuscripts 手抄本的机器年代测定
Pub Date : 2007-09-23 DOI: 10.1109/ICDAR.2007.4377017
Utpal Garain, S. K. Parui, T. Paquet, L. Heutte
This paper presents a pioneering study on automatic dating of handwritten manuscripts. Analysis of handwriting style forms the core of the dating method. Initially, it is hypothesized that a manuscript can be dated, to a certain level of accuracy, by looking at the way it is written. The hypothesis is then verified with real samples of known dates. A general framework is proposed for machine dating of handwritten manuscripts. Experiments on a database containing manuscripts of Gustave Flaubert (1821- 1880), the famous French novelist reports about 62% accuracy when manuscripts are dated within a range of five calendar years with respect to their exact year of writing.
本文介绍了一项手写体手稿自动测年的开创性研究。笔迹风格的分析是年代测定方法的核心。最初,人们假设一份手稿可以通过观察它的写作方式来确定它的年代,达到一定程度的准确性。然后用已知日期的真实样本验证该假设。提出了手写体手稿机器定年的一般框架。法国著名小说家古斯塔夫·福楼拜(Gustave Flaubert, 1821- 1880)在一个包含他手稿的数据库中进行了实验,结果显示,当手稿的日期在5个日历年的范围内时,其准确写作年份的准确率为62%。
{"title":"Machine Dating of Handwritten Manuscripts","authors":"Utpal Garain, S. K. Parui, T. Paquet, L. Heutte","doi":"10.1109/ICDAR.2007.4377017","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.4377017","url":null,"abstract":"This paper presents a pioneering study on automatic dating of handwritten manuscripts. Analysis of handwriting style forms the core of the dating method. Initially, it is hypothesized that a manuscript can be dated, to a certain level of accuracy, by looking at the way it is written. The hypothesis is then verified with real samples of known dates. A general framework is proposed for machine dating of handwritten manuscripts. Experiments on a database containing manuscripts of Gustave Flaubert (1821- 1880), the famous French novelist reports about 62% accuracy when manuscripts are dated within a range of five calendar years with respect to their exact year of writing.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122430564","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Learning to Group Text Lines and Regions in Freeform Handwritten Notes 学习在自由形式的手写笔记中分组文本行和区域
Ming Ye, Paul A. Viola, Sashi Raghupathy, H. Sutanto, Chengyang Li
This paper proposes a machine learning approach to grouping problems in ink parsing. Starting from an initial segmentation, hypotheses are generated by perturbing local configurations and processed in a high-confidence-first fashion, where the confidence of each hypothesis is produced by a data-driven AdaBoost decision-tree classifier with a set of intuitive features. This framework has successfully applied to grouping text lines and regions in complex freeform digital ink notes from real TabletPC users. It holds great potential in solving many other grouping problems in the ink parsing and document image analysis domains.
本文提出了一种机器学习方法来解决油墨解析中的分组问题。从初始分割开始,通过扰动局部配置生成假设,并以高置信度优先的方式进行处理,其中每个假设的置信度由数据驱动的AdaBoost决策树分类器产生,该分类器具有一组直观的特征。该框架已成功地应用于对来自真实TabletPC用户的复杂自由格式数字墨水笔记中的文本行和区域进行分组。它在解决油墨解析和文档图像分析领域的许多其他分组问题方面具有很大的潜力。
{"title":"Learning to Group Text Lines and Regions in Freeform Handwritten Notes","authors":"Ming Ye, Paul A. Viola, Sashi Raghupathy, H. Sutanto, Chengyang Li","doi":"10.1109/ICDAR.2007.159","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.159","url":null,"abstract":"This paper proposes a machine learning approach to grouping problems in ink parsing. Starting from an initial segmentation, hypotheses are generated by perturbing local configurations and processed in a high-confidence-first fashion, where the confidence of each hypothesis is produced by a data-driven AdaBoost decision-tree classifier with a set of intuitive features. This framework has successfully applied to grouping text lines and regions in complex freeform digital ink notes from real TabletPC users. It holds great potential in solving many other grouping problems in the ink parsing and document image analysis domains.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123107893","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A New Method for Writer Identification and Verification Based on Farsi/Arabic Handwritten Texts 基于波斯语/阿拉伯语手写体文本的作者识别与验证新方法
F. Nejad, M. Rahmati
Most studies about writer identification are based on English documents and to our knowledge no research has been reported on Farsi or Arabic documents. In this paper, we have proposed a method for off-line writer identification and verification based on Farsi handwriting, which is text-dependent. Based on the idea that has been presented in the previous studies, here we assume handwriting as texture image and after normalization step, the Gabor filters are applied to image and then new features are extracted. Substantially, the property of proposed method is using of the bank of Gabor filters which is appropriate for the structure of Farsi handwritten texts and vision system. Also, a new method for feature extraction from output of Gabor filters is proposed which is based on moments and nonlinear transform. In this paper, with definition a confidence criterion, a new method for writer verification is proposed. Evaluation of other methods and proposed method demonstrates that proposed method achieves better performance on Farsi handwritten from 40 peoples.
大多数关于作者识别的研究都是基于英语文献,据我们所知,没有关于波斯语或阿拉伯语文献的研究报道。在本文中,我们提出了一种基于波斯语手写的离线写作者识别和验证方法,该方法依赖于文本。基于之前的研究思路,我们假设笔迹为纹理图像,经过归一化步骤后,对图像进行Gabor滤波,提取新的特征。从本质上讲,该方法的特点是使用了适合波斯语手写文本结构和视觉系统的Gabor滤波器库。同时,提出了一种基于矩量和非线性变换的Gabor滤波器输出特征提取方法。本文通过定义置信准则,提出了一种新的作者验证方法。对其他方法和本文提出的方法进行了比较,结果表明本文提出的方法对40个民族的波斯语手写体具有较好的识别效果。
{"title":"A New Method for Writer Identification and Verification Based on Farsi/Arabic Handwritten Texts","authors":"F. Nejad, M. Rahmati","doi":"10.1109/ICDAR.2007.24","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.24","url":null,"abstract":"Most studies about writer identification are based on English documents and to our knowledge no research has been reported on Farsi or Arabic documents. In this paper, we have proposed a method for off-line writer identification and verification based on Farsi handwriting, which is text-dependent. Based on the idea that has been presented in the previous studies, here we assume handwriting as texture image and after normalization step, the Gabor filters are applied to image and then new features are extracted. Substantially, the property of proposed method is using of the bank of Gabor filters which is appropriate for the structure of Farsi handwritten texts and vision system. Also, a new method for feature extraction from output of Gabor filters is proposed which is based on moments and nonlinear transform. In this paper, with definition a confidence criterion, a new method for writer verification is proposed. Evaluation of other methods and proposed method demonstrates that proposed method achieves better performance on Farsi handwritten from 40 peoples.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126614275","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 56
Simultaneous Layout Style and Logical Entity Recognition in a Heterogeneous Collection of Documents 异构文档集合中的同步布局样式和逻辑实体识别
Siyuan Chen, Song Mao, G. Thoma
Logical entity recognition in heterogeneous collections of document page images remains a challenging problem since the performance of traditional supervised methods degrades dramatically in case of many distinct layout styles. In this paper we present an unsupervised method where layout style information is explicitly used in both training and recognition phases. We represent the layout style, local features, and logical labels of physical regions of a document compactly by an ordered labeled X-Y tree. Style dissimilarity of two document pages is represented by the distance between their respective trees. During the training phase, document pages with true logical labels in training set are classified into distinct layout styles by unsupervised clustering. During the recognition phase, the layout style and logical entities of an input document are recognized simultaneously by matching the input tree to the trees in closest- matched layout style cluster of training set. Experimental results show that our algorithm is robust with both balanced and unbalanced style cluster sizes, zone over-segmentation, zone length variation, and variation in tree representations of the same layout style.
在异构文档页面图像集合中的逻辑实体识别仍然是一个具有挑战性的问题,因为传统的监督方法在许多不同布局样式的情况下性能显著下降。在本文中,我们提出了一种无监督的方法,其中布局样式信息在训练和识别阶段都被明确地使用。我们通过有序的标记X-Y树紧凑地表示文档的物理区域的布局样式、局部特征和逻辑标签。两个文档页面的样式差异由它们各自树之间的距离表示。在训练阶段,通过无监督聚类将训练集中具有真实逻辑标签的文档页面划分为不同的布局样式。在识别阶段,通过将输入树与训练集最匹配的布局样式簇中的树进行匹配,同时识别输入文档的布局样式和逻辑实体。实验结果表明,该算法对平衡和不平衡样式簇大小、区域过度分割、区域长度变化和相同布局样式的树表示变化都具有鲁棒性。
{"title":"Simultaneous Layout Style and Logical Entity Recognition in a Heterogeneous Collection of Documents","authors":"Siyuan Chen, Song Mao, G. Thoma","doi":"10.1109/ICDAR.2007.231","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.231","url":null,"abstract":"Logical entity recognition in heterogeneous collections of document page images remains a challenging problem since the performance of traditional supervised methods degrades dramatically in case of many distinct layout styles. In this paper we present an unsupervised method where layout style information is explicitly used in both training and recognition phases. We represent the layout style, local features, and logical labels of physical regions of a document compactly by an ordered labeled X-Y tree. Style dissimilarity of two document pages is represented by the distance between their respective trees. During the training phase, document pages with true logical labels in training set are classified into distinct layout styles by unsupervised clustering. During the recognition phase, the layout style and logical entities of an input document are recognized simultaneously by matching the input tree to the trees in closest- matched layout style cluster of training set. Experimental results show that our algorithm is robust with both balanced and unbalanced style cluster sizes, zone over-segmentation, zone length variation, and variation in tree representations of the same layout style.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"130 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127838342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Text Input System Using Online Overlapped Handwriting Recognition for Mobile Devices 移动设备上使用在线重叠手写识别的文本输入系统
Yojiro Tonouchi, A. Kawamura
This paper proposes a novel online overlapped handwriting recognition system for mobile devices such as cellular phones. Users can input characters continuously without pauses on the single writing area. It has three features: small writing area, quick response and direct operations with handwritten gestures. Therefore, it is suitable for mobile devices such as cellular phones. The system realizes a new handwriting interface similar to touch-typing. We evaluated the system by two experiments: character recognition performance and text entry speed of Japanese sentences. Through these experiments we showed the effectiveness of the proposed system.
本文提出了一种新的用于移动设备(如手机)的在线重叠手写识别系统。用户可以在单个书写区域连续输入字符而不停顿。它有三个特点:书写面积小,反应快,手写手势操作直接。因此,它适用于手机等移动设备。该系统实现了一种类似于触摸打字的新型手写界面。我们通过两个实验来评估系统:字符识别性能和日语句子的文本输入速度。通过这些实验,我们证明了所提出系统的有效性。
{"title":"Text Input System Using Online Overlapped Handwriting Recognition for Mobile Devices","authors":"Yojiro Tonouchi, A. Kawamura","doi":"10.1109/ICDAR.2007.243","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.243","url":null,"abstract":"This paper proposes a novel online overlapped handwriting recognition system for mobile devices such as cellular phones. Users can input characters continuously without pauses on the single writing area. It has three features: small writing area, quick response and direct operations with handwritten gestures. Therefore, it is suitable for mobile devices such as cellular phones. The system realizes a new handwriting interface similar to touch-typing. We evaluated the system by two experiments: character recognition performance and text entry speed of Japanese sentences. Through these experiments we showed the effectiveness of the proposed system.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127756317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
PRAAD: Preprocessing and Analysis Tool for Arabic Ancient Documents PRAAD:阿拉伯语古代文献预处理与分析工具
Wafa Boussellaa, Abderrazak Zahour, B. Taconet, A. Alimi, A. BenAbdelhafid
This paper presents the new system PRAAD for preprocessing and analysis of Arabic historical documents. It is composed of two important parts: pre-processing and analysis of ancient documents. After digitization, the color or greyscale ancient documents images are distorted by the presence of strong background artefacts such as scan optical blur and noise, show-through and bleed-through effects and spots. In order to preserve and exploit this cultural heritage documents, we intend to create efficient tool that achieves restoration, binarisation, and analyses the document layout. The developed tool is done by adapting our expertise in document image processing of Arabic ancient documents, printed or manuscripts. The different functions of PRAAD system are tested on a set of Arabic ancient documents from the national library and the National Archives of Tunisia.
本文介绍了一种新的阿拉伯语历史文献预处理分析系统PRAAD。它由两个重要部分组成:古代文献的预处理和分析。数字化后的彩色或灰度古代文献图像由于扫描光学模糊和噪声、透显和透透效果和斑点等强背景伪影的存在而发生畸变。为了保护和利用这些文化遗产文件,我们打算创建一个有效的工具来实现文件布局的修复、二值化和分析。开发的工具是通过调整我们在阿拉伯古代文件,印刷或手稿的文件图像处理方面的专业知识来完成的。以突尼斯国家图书馆和国家档案馆的一组阿拉伯文古文献为例,对PRAAD系统的不同功能进行了测试。
{"title":"PRAAD: Preprocessing and Analysis Tool for Arabic Ancient Documents","authors":"Wafa Boussellaa, Abderrazak Zahour, B. Taconet, A. Alimi, A. BenAbdelhafid","doi":"10.1109/ICDAR.2007.209","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.209","url":null,"abstract":"This paper presents the new system PRAAD for preprocessing and analysis of Arabic historical documents. It is composed of two important parts: pre-processing and analysis of ancient documents. After digitization, the color or greyscale ancient documents images are distorted by the presence of strong background artefacts such as scan optical blur and noise, show-through and bleed-through effects and spots. In order to preserve and exploit this cultural heritage documents, we intend to create efficient tool that achieves restoration, binarisation, and analyses the document layout. The developed tool is done by adapting our expertise in document image processing of Arabic ancient documents, printed or manuscripts. The different functions of PRAAD system are tested on a set of Arabic ancient documents from the national library and the National Archives of Tunisia.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"10 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114038748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1