首页 > 最新文献

Proceedings of the Fourth International Conference on Document Analysis and Recognition最新文献

英文 中文
Image and text coupling for creating electronic books from manuscripts 从手稿中创建电子图书的图像和文本耦合
Laurent Robert, Laurence Likforman-Sulem, É. Lecolinet
Presents the first achievements of HERS (Hypermedia Edit and Read Station), which is devoted to the browsing and editing of hypermedia documents from literary material including document images. Concerning the editing, our purpose as twofold. First, capabilities are offered to transcribe manuscripts. Transcribing the text consists of coupling lines typed on the keyboard with their corresponding text lines in the manuscript images. A collaborative system, based on computer-human interaction and document analysis, is proposed for performing this task. Second, interactive tools are offered to organize the electronic document and establish hypermedia links between its different components (image areas, transcribed words or lines, or other kinds of heterogeneous data). Concerning the browsing, we developed an approach based on information visualization in order to provide users with an idea of the overall organization of the hyperdocument and so help them to navigate through it.
介绍了HERS(超媒体编辑和阅读站)的第一批成果,该工作站致力于浏览和编辑包括文档图像在内的文学材料中的超媒体文档。关于编辑,我们的目的是双重的。首先,提供抄写手稿的能力。抄写文本由键盘上键入的行与手稿图像中相应的文本行耦合组成。提出了一种基于人机交互和文档分析的协作系统来完成这一任务。其次,提供交互式工具来组织电子文档,并在其不同组件(图像区域、转录的单词或行或其他类型的异构数据)之间建立超媒体链接。关于浏览,我们开发了一种基于信息可视化的方法,以便为用户提供超文档的总体组织的概念,从而帮助他们浏览它。
{"title":"Image and text coupling for creating electronic books from manuscripts","authors":"Laurent Robert, Laurence Likforman-Sulem, É. Lecolinet","doi":"10.1109/ICDAR.1997.620626","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620626","url":null,"abstract":"Presents the first achievements of HERS (Hypermedia Edit and Read Station), which is devoted to the browsing and editing of hypermedia documents from literary material including document images. Concerning the editing, our purpose as twofold. First, capabilities are offered to transcribe manuscripts. Transcribing the text consists of coupling lines typed on the keyboard with their corresponding text lines in the manuscript images. A collaborative system, based on computer-human interaction and document analysis, is proposed for performing this task. Second, interactive tools are offered to organize the electronic document and establish hypermedia links between its different components (image areas, transcribed words or lines, or other kinds of heterogeneous data). Concerning the browsing, we developed an approach based on information visualization in order to provide users with an idea of the overall organization of the hyperdocument and so help them to navigate through it.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114080858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Table image segmentation 表格图像分割
Konstantin Zuyev
Algorithm for table image segmentation, a part of complete document recognition system is presented. The proposed approach introduces a concept of table grid which can serve for advanced methods of table structure analysis. It provides a layer of terminal symbols for the table, which is used by syntactical methods. Detailed discussion of grid detection is presented which is performed through the analysis of connected components projection profile. Simple rules for analysis of table structure cover majority of real life tables. The system is implemented, rested, and is now extensively used in FineReader OCR product.
给出了完整文档识别系统的一部分表图像分割算法。该方法引入了表网格的概念,为表结构分析提供了先进的方法。它为表提供了一层终端符号,供语法方法使用。详细讨论了通过分析连通构件投影轮廓来实现网格检测的方法。表结构分析的简单规则涵盖了大多数实际生活表。该系统已经实现,并且现在广泛应用于FineReader OCR产品中。
{"title":"Table image segmentation","authors":"Konstantin Zuyev","doi":"10.1109/ICDAR.1997.620599","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620599","url":null,"abstract":"Algorithm for table image segmentation, a part of complete document recognition system is presented. The proposed approach introduces a concept of table grid which can serve for advanced methods of table structure analysis. It provides a layer of terminal symbols for the table, which is used by syntactical methods. Detailed discussion of grid detection is presented which is performed through the analysis of connected components projection profile. Simple rules for analysis of table structure cover majority of real life tables. The system is implemented, rested, and is now extensively used in FineReader OCR product.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"183 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121034191","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 48
An Image Consulting Framework for document analysis of Internet graphics 一个用于网络图形文档分析的图像咨询框架
M. Köppen, L. Lohmann, B. Nickolay
A new system approach for image understanding, called the Image Consulting Framework, is proposed. It allows for the validation of image properties. The kinds of image properties considered are textual, textural, hierarchical, color and symbolic. Its main application field is information filtering from images used in World Wide Web documents. The Image Consulting Framework consists of four stages: the color separation stage, the information granulation-verification modules (GVMs), the task stage and the recognition stage. At the base of the framework are the GVMs, which are designed to solve very special tasks. They consists of three parts: a method maintainer, a parameter chooser and a tester (verifier). The parameter chooser uses a given set of parameter settings for different runs of the maintained method on the input images of the GVM. The resulting images are tested for the occurrence of the property for which the GVM is designed. All successful images are put into a queue. The task stage calls new GVMs due to the filling of the queue, and it also assigns input images to the GVMs. All fully-treated images are passed to the recognition stage, where the information extraction is performed.
提出了一种新的图像理解系统方法,称为图像咨询框架。它允许对图像属性进行验证。所考虑的图像属性类型包括文本、纹理、层次、颜色和符号。它的主要应用领域是万维网文档中图像的信息过滤。图像咨询框架包括四个阶段:分色阶段、信息颗粒验证模块(GVMs)、任务阶段和识别阶段。该框架的基础是gvm,它们被设计用来解决非常特殊的任务。它们由三部分组成:方法维护者、参数选择者和测试者(验证者)。参数选择器为维护方法在GVM的输入映像上的不同运行使用一组给定的参数设置。对生成的映像进行测试,以确定是否存在为GVM设计的属性。所有成功的映像都放入队列中。由于队列填充,任务阶段调用新的gvm,并将输入映像分配给gvm。所有经过充分处理的图像都被传递到识别阶段,在那里执行信息提取。
{"title":"An Image Consulting Framework for document analysis of Internet graphics","authors":"M. Köppen, L. Lohmann, B. Nickolay","doi":"10.1109/ICDAR.1997.620625","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620625","url":null,"abstract":"A new system approach for image understanding, called the Image Consulting Framework, is proposed. It allows for the validation of image properties. The kinds of image properties considered are textual, textural, hierarchical, color and symbolic. Its main application field is information filtering from images used in World Wide Web documents. The Image Consulting Framework consists of four stages: the color separation stage, the information granulation-verification modules (GVMs), the task stage and the recognition stage. At the base of the framework are the GVMs, which are designed to solve very special tasks. They consists of three parts: a method maintainer, a parameter chooser and a tester (verifier). The parameter chooser uses a given set of parameter settings for different runs of the maintained method on the input images of the GVM. The resulting images are tested for the occurrence of the property for which the GVM is designed. All successful images are put into a queue. The task stage calls new GVMs due to the filling of the queue, and it also assigns input images to the GVMs. All fully-treated images are passed to the recognition stage, where the information extraction is performed.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125233012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Retrieval methods for English-text with missrecognized OCR characters 英文文本OCR字符识别错误的检索方法
Manabu Ohta, A. Takasu, J. Adachi
This paper presents three probabilistic text retrieval methods designed to carry out a full-text search of English documents containing OCR errors. By searching for any query term on the premise that there are errors in the recognized text, the methods presented can tolerate such errors, and therefore costly manual post-editing is not required after OCR recognition. In the applied approach, confusion matrices are used to store characters which are likely to be interchanged when a particular character is missrecognized, and the respective probability of each occurrence. Moreover, a 2-gram matrix is used to store probabilities of character connection, i.e., which letter is likely to come after another. Multiple search terms are generated for an input query term by making reference to confusion matrices, after which a full-text search is run for each search term. The validity of retrieved terms is determined based on error-occurrence and character connection probabilities. The performance of these methods is experimentally evaluated by determining retrieval effectiveness, i.e., by calculating recall and precision rates. Results indicate marked improvement in comparison with exact matching.
本文提出了三种概率文本检索方法,用于对包含OCR错误的英文文档进行全文检索。通过在识别文本中存在错误的前提下搜索任何查询项,所提出的方法可以容忍这种错误,因此在OCR识别后不需要进行昂贵的人工后期编辑。在应用方法中,混淆矩阵用于存储当某个特定字符被错误识别时可能互换的字符,以及每个字符各自出现的概率。此外,一个2克矩阵用于存储字符连接的概率,即哪个字母可能出现在另一个字母之后。通过引用混淆矩阵为输入查询词生成多个搜索词,然后为每个搜索词运行全文搜索。根据错误发生概率和字符连接概率确定检索词的有效性。这些方法的性能是通过实验来评估检索效率,即通过计算召回率和准确率。结果表明,与精确匹配相比,有明显的改善。
{"title":"Retrieval methods for English-text with missrecognized OCR characters","authors":"Manabu Ohta, A. Takasu, J. Adachi","doi":"10.1109/ICDAR.1997.620651","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620651","url":null,"abstract":"This paper presents three probabilistic text retrieval methods designed to carry out a full-text search of English documents containing OCR errors. By searching for any query term on the premise that there are errors in the recognized text, the methods presented can tolerate such errors, and therefore costly manual post-editing is not required after OCR recognition. In the applied approach, confusion matrices are used to store characters which are likely to be interchanged when a particular character is missrecognized, and the respective probability of each occurrence. Moreover, a 2-gram matrix is used to store probabilities of character connection, i.e., which letter is likely to come after another. Multiple search terms are generated for an input query term by making reference to confusion matrices, after which a full-text search is run for each search term. The validity of retrieved terms is determined based on error-occurrence and character connection probabilities. The performance of these methods is experimentally evaluated by determining retrieval effectiveness, i.e., by calculating recall and precision rates. Results indicate marked improvement in comparison with exact matching.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121878919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
Generalized contextual recognition of hand-printed documents using semantic trees with lazy evaluation 基于延迟求值的语义树手印文档的广义上下文识别
L. Du, A. Downton, S. Lucas, Badr Al-Badr
Describes a new general-purpose contextual architecture which provides a unified framework for efficiently combining all types and levels of context in hand-print recognition applications. The architecture has been designed and built as a C++ class library and utilised within an initial demonstrator which implements full contextual constraints for a combination of postcode and corresponding postal address. Preliminary evaluation of the demonstrator suggests the system has the potential to achieve genuinely remarkable performance compared with previous context systems: its memory requirements are an order of magnitude less than an equivalent trie-based dictionary; its search speed is at least an order of magnitude faster than the trie, and actually gets faster as the dictionary size increases(!); and its error rate is virtually zero if suitable contextual constraints can be applied. Using this architecture, it appears to be possible to build real-time solutions to large-scale heterogeneous contextual problems.
描述一种新的通用上下文体系结构,该体系结构为手印识别应用程序中有效地组合所有类型和级别的上下文提供了统一的框架。该体系结构是作为c++类库设计和构建的,并在一个初始演示中使用,该演示实现了邮政编码和相应邮政地址组合的完整上下文约束。对演示器的初步评估表明,与以前的上下文系统相比,该系统有潜力实现真正卓越的性能:它的内存需求比同等的基于尝试的字典少一个数量级;它的搜索速度至少比trie快一个数量级,并且实际上随着字典大小的增加而变得更快(!);如果可以应用合适的上下文约束,它的错误率几乎为零。使用这种体系结构,似乎可以为大规模异构上下文问题构建实时解决方案。
{"title":"Generalized contextual recognition of hand-printed documents using semantic trees with lazy evaluation","authors":"L. Du, A. Downton, S. Lucas, Badr Al-Badr","doi":"10.1109/ICDAR.1997.619848","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619848","url":null,"abstract":"Describes a new general-purpose contextual architecture which provides a unified framework for efficiently combining all types and levels of context in hand-print recognition applications. The architecture has been designed and built as a C++ class library and utilised within an initial demonstrator which implements full contextual constraints for a combination of postcode and corresponding postal address. Preliminary evaluation of the demonstrator suggests the system has the potential to achieve genuinely remarkable performance compared with previous context systems: its memory requirements are an order of magnitude less than an equivalent trie-based dictionary; its search speed is at least an order of magnitude faster than the trie, and actually gets faster as the dictionary size increases(!); and its error rate is virtually zero if suitable contextual constraints can be applied. Using this architecture, it appears to be possible to build real-time solutions to large-scale heterogeneous contextual problems.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127664035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Information capture and semantic indexing of digital libraries through machine learning techniques 基于机器学习技术的数字图书馆信息捕获和语义索引
F. Esposito, D. Malerba, G. Semeraro, Cesare Daniele Antifora, G. D. Gennaro
This paper presents a prototypical digital library service. It integrates machine learning tools and techniques in order to make effective, efficient and economically feasible the process of capturing the information that should be stored and indexed by content in the digital library. Infact, information capture is one of the main bottleneck when building a digital library, since it involves complex pattern recognition problems, such as document analysis, classification and understanding. Experimental results show that learning systems can solve effectively and efficiently these problems.
本文提出了一种典型的数字图书馆服务。它集成了机器学习工具和技术,以便有效、高效和经济可行地捕获应该在数字图书馆中存储和索引内容的信息。事实上,信息捕获是构建数字图书馆的主要瓶颈之一,因为它涉及复杂的模式识别问题,如文档分析、分类和理解。实验结果表明,学习系统可以有效地解决这些问题。
{"title":"Information capture and semantic indexing of digital libraries through machine learning techniques","authors":"F. Esposito, D. Malerba, G. Semeraro, Cesare Daniele Antifora, G. D. Gennaro","doi":"10.1109/ICDAR.1997.620603","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620603","url":null,"abstract":"This paper presents a prototypical digital library service. It integrates machine learning tools and techniques in order to make effective, efficient and economically feasible the process of capturing the information that should be stored and indexed by content in the digital library. Infact, information capture is one of the main bottleneck when building a digital library, since it involves complex pattern recognition problems, such as document analysis, classification and understanding. Experimental results show that learning systems can solve effectively and efficiently these problems.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133635929","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Variations on the analysis of architectural drawings 建筑图纸分析的变化
Christian Ah-Soon, K. Tombre
Lately, our team has begun investigating the analysis of architectural drawings. The paper presents our first results. After a brief introduction to the specificities of architectural drawings and a short review of existing work, we describe the low level processing steps we have implemented: segmentation, vectorization and detection of arcs, loops detection. In the present state of our work, we have investigated two complementary techniques for higher level analysis; one is based on geometric analysis and symbol recognition, the other relies on the idea that architecture is a combination of spaces, and is therefore based on spatial analysis. We present our current results with these two techniques, and we suggest a number of perspectives for the continuation of this work.
最近,我们的团队开始研究建筑图纸的分析。这篇论文介绍了我们的初步结果。在简要介绍了建筑图纸的特殊性和对现有工作的简短回顾之后,我们描述了我们已经实现的低级处理步骤:分割,矢量化和弧线检测,环路检测。在我们目前的工作状态下,我们研究了两种互补的技术用于更高层次的分析;一种基于几何分析和符号识别,另一种依赖于建筑是空间组合的想法,因此基于空间分析。我们介绍了目前这两种技术的结果,并为这项工作的继续提出了一些观点。
{"title":"Variations on the analysis of architectural drawings","authors":"Christian Ah-Soon, K. Tombre","doi":"10.1109/ICDAR.1997.619869","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619869","url":null,"abstract":"Lately, our team has begun investigating the analysis of architectural drawings. The paper presents our first results. After a brief introduction to the specificities of architectural drawings and a short review of existing work, we describe the low level processing steps we have implemented: segmentation, vectorization and detection of arcs, loops detection. In the present state of our work, we have investigated two complementary techniques for higher level analysis; one is based on geometric analysis and symbol recognition, the other relies on the idea that architecture is a combination of spaces, and is therefore based on spatial analysis. We present our current results with these two techniques, and we suggest a number of perspectives for the continuation of this work.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134397290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 47
New features for Chinese character recognition 新增汉字识别功能
T. Caesar
The wide range of shape variations for Chinese characters requires an adequate representation of the discriminating features for classification. For the recognition of Latin characters or numerals pixel values of a normalized raster image are proper features to reach very good recognition rates. But Chinese characters require a much higher resolution of the normalized raster image to enable a discrimination of complex shaped characters which leads to a feature space dimensionality of prohibitive computational effort for classification. Therefore feature extraction algorithms are needed which capture the discriminative characteristics of character shapes in a compact form. Several algorithms were proposed in the past and many of them are based on the contour data. This paper also introduces a contour based approach which is very time efficient and overcomes the problem of vanishing lines during anisotropic size normalization.
汉字的形状变化范围很广,因此需要对汉字的识别特征进行充分的表征。对于拉丁字符或数字的识别,归一化光栅图像的像素值是达到非常好的识别率的适当特征。但是汉字需要更高分辨率的归一化栅格图像来识别复杂形状的字符,这导致分类的特征空间维度难以计算。因此,需要一种能够以紧凑的形式捕获字符形状的判别特征的特征提取算法。过去提出了几种算法,其中许多算法都是基于等高线数据的。本文还介绍了一种基于轮廓线的方法,该方法非常省时,并且克服了各向异性尺寸归一化过程中线消失的问题。
{"title":"New features for Chinese character recognition","authors":"T. Caesar","doi":"10.1109/ICDAR.1997.620571","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620571","url":null,"abstract":"The wide range of shape variations for Chinese characters requires an adequate representation of the discriminating features for classification. For the recognition of Latin characters or numerals pixel values of a normalized raster image are proper features to reach very good recognition rates. But Chinese characters require a much higher resolution of the normalized raster image to enable a discrimination of complex shaped characters which leads to a feature space dimensionality of prohibitive computational effort for classification. Therefore feature extraction algorithms are needed which capture the discriminative characteristics of character shapes in a compact form. Several algorithms were proposed in the past and many of them are based on the contour data. This paper also introduces a contour based approach which is very time efficient and overcomes the problem of vanishing lines during anisotropic size normalization.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"198 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133853297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Confidence computation improvement in an optical field reading system 光学场读数系统置信度计算改进
A. Benedetti, Z. Kovács-Vajna
An expression in closed form is derived for the recognition error vs. rejection rate of optical character or word recognition systems. This expression allows to define a lower bound for the error rate of any recognition system employing a rejection process based on the definition of a confidence threshold. This relation has also proved to be useful to make a quantitative comparison between two confidence computation methods implemented in a system for reading USA Census '90 hand-written forms. The newly proposed method is based upon a confidence model integrating single-character confidence levels, digram statistics and other information from the dictionary matching phase. At a 50% rejection rate, the field error rate calculated using the new confidence computation algorithm decreased from 47.7% to 44.6%, which represents a considerable improvement, given a theoretical lower bound of 40.8% on the error rate.
导出了光学字符或单词识别系统的识别误差与拒绝率的封闭表达式。该表达式允许定义基于置信度阈值定义的任何采用拒绝过程的识别系统的错误率的下界。这种关系也被证明是有用的,使两种置信度计算方法之间的定量比较在一个系统中实施的阅读美国人口普查'90手写表格。新提出的方法基于一个置信度模型,该模型集成了单字符置信度、图统计和字典匹配阶段的其他信息。在拒绝率为50%时,使用新的置信度计算算法计算的现场错误率从47.7%下降到44.6%,这是一个相当大的改进,因为错误率的理论下限为40.8%。
{"title":"Confidence computation improvement in an optical field reading system","authors":"A. Benedetti, Z. Kovács-Vajna","doi":"10.1109/ICDAR.1997.620629","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620629","url":null,"abstract":"An expression in closed form is derived for the recognition error vs. rejection rate of optical character or word recognition systems. This expression allows to define a lower bound for the error rate of any recognition system employing a rejection process based on the definition of a confidence threshold. This relation has also proved to be useful to make a quantitative comparison between two confidence computation methods implemented in a system for reading USA Census '90 hand-written forms. The newly proposed method is based upon a confidence model integrating single-character confidence levels, digram statistics and other information from the dictionary matching phase. At a 50% rejection rate, the field error rate calculated using the new confidence computation algorithm decreased from 47.7% to 44.6%, which represents a considerable improvement, given a theoretical lower bound of 40.8% on the error rate.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122955681","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Handwritten ZIP code recognition 手写邮政编码识别
Gregory I. Dzuba, Alexander Filatov, A. Volgunin
The encoding of delivery point code (DPC) for a handwritten address is one of the most complex problems of the US mail delivery automation. This paper describes a real-time system intended to recognize the 5-digit ZIP code part of DPC. To increase the system performance the results of ZIP code recognition are cross-validated with those of city and state name recognition. The main principles of the handwritten word recognizer which provide the core of the system are explained. The system throughput is 40,000 address blocks per hour. Experimental results on live mail pieces are presented. The ZIP code recognition rate is 73% with 1% error rate.
手写地址的投递点编码(DPC)是美国邮件投递自动化中最复杂的问题之一。本文介绍了一种实时识别邮政编码中5位邮编的系统。为了提高系统的性能,将邮政编码识别结果与城市和州名识别结果进行交叉验证。介绍了手写体文字识别器的主要工作原理,它构成了系统的核心。系统吞吐量为每小时4万个地址块。给出了在实弹邮件上的实验结果。邮政编码识别率为73%,错误率为1%。
{"title":"Handwritten ZIP code recognition","authors":"Gregory I. Dzuba, Alexander Filatov, A. Volgunin","doi":"10.1109/ICDAR.1997.620613","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620613","url":null,"abstract":"The encoding of delivery point code (DPC) for a handwritten address is one of the most complex problems of the US mail delivery automation. This paper describes a real-time system intended to recognize the 5-digit ZIP code part of DPC. To increase the system performance the results of ZIP code recognition are cross-validated with those of city and state name recognition. The main principles of the handwritten word recognizer which provide the core of the system are explained. The system throughput is 40,000 address blocks per hour. Experimental results on live mail pieces are presented. The ZIP code recognition rate is 73% with 1% error rate.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121784133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
期刊
Proceedings of the Fourth International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1