首页 > 最新文献

Proceedings of 3rd International Conference on Document Analysis and Recognition最新文献

英文 中文
Evaluation of an interactive tool for handwritten form description 评估手写表单描述的交互式工具
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602135
G. Leedham, D. Monger
A highly time-consuming activity in many areas of commerce and business is the manual entry into computer of data handwritten on forms. All forms in widespread use contain discrete fields where specific information can be entered. Automatic recognition of these forms could be achieved using existing state-of-the-art OCR algorithms for numerals, alphabetic characters, cursive words, signatures and mark sensing if they could be rapidly configured along with any inter-relationships and dependencies for different forms. This paper describes an initial implementation of an interactive graphical tool to allow the handwritten fields of a form and their inter-relationships to be described and defined for automatic linking with appropriate OCR algorithms. Results indicate that the main requirement is for the operator to have a full understanding of the handwritten form and an ability to describe its contents.
在商业和商业的许多领域中,手工将手写的数据输入计算机是一项非常耗时的活动。所有广泛使用的表单都包含可以输入特定信息的离散字段。这些形式的自动识别可以使用现有的最先进的数字、字母、草书、签名和标记感知的OCR算法来实现,如果它们可以快速配置为不同形式的任何相互关系和依赖关系。本文描述了一个交互式图形工具的初始实现,该工具允许描述和定义表单的手写字段及其相互关系,以便与适当的OCR算法自动链接。结果表明,对操作员的主要要求是对手写表格有充分的理解和描述其内容的能力。
{"title":"Evaluation of an interactive tool for handwritten form description","authors":"G. Leedham, D. Monger","doi":"10.1109/ICDAR.1995.602135","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602135","url":null,"abstract":"A highly time-consuming activity in many areas of commerce and business is the manual entry into computer of data handwritten on forms. All forms in widespread use contain discrete fields where specific information can be entered. Automatic recognition of these forms could be achieved using existing state-of-the-art OCR algorithms for numerals, alphabetic characters, cursive words, signatures and mark sensing if they could be rapidly configured along with any inter-relationships and dependencies for different forms. This paper describes an initial implementation of an interactive graphical tool to allow the handwritten fields of a form and their inter-relationships to be described and defined for automatic linking with appropriate OCR algorithms. Results indicate that the main requirement is for the operator to have a full understanding of the handwritten form and an ability to describe its contents.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"111 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114235176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A probabilistic approach to automatic handwritten address reading 一种自动手写地址读取的概率方法
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599015
J. Bertille, M. Gilloux
To sort handwritten mail pieces, the French Postal Office Research Centre has developed a reading device. After recalling the different processing stages of our system, we focus on the problem of merging the decisions made by its different sub-modules (postal code and city name hypotheses) to produce a unique and reliable final decision. To achieve this goal we use a probabilistic modelling of the system behaviour taking into account the characteristics of all the processing stages involved in handwritten address recognition. This approach has been rested on large reference sets consisting of live mail handwritten envelopes and already produces very promising results.
为了对手写邮件进行分类,法国邮政研究中心开发了一种阅读设备。在回顾了我们系统的不同处理阶段之后,我们将重点放在合并其不同子模块(邮政编码和城市名称假设)做出的决策以产生唯一且可靠的最终决策的问题上。为了实现这一目标,我们使用系统行为的概率建模,考虑到手写地址识别中涉及的所有处理阶段的特征。这种方法已经建立在由实时邮件手写信封组成的大型参考集上,并且已经产生了非常有希望的结果。
{"title":"A probabilistic approach to automatic handwritten address reading","authors":"J. Bertille, M. Gilloux","doi":"10.1109/ICDAR.1995.599015","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599015","url":null,"abstract":"To sort handwritten mail pieces, the French Postal Office Research Centre has developed a reading device. After recalling the different processing stages of our system, we focus on the problem of merging the decisions made by its different sub-modules (postal code and city name hypotheses) to produce a unique and reliable final decision. To achieve this goal we use a probabilistic modelling of the system behaviour taking into account the characteristics of all the processing stages involved in handwritten address recognition. This approach has been rested on large reference sets consisting of live mail handwritten envelopes and already produces very promising results.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"144 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121696741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Four directional adjacency graphs (FDAG) and their application in locating fields in forms 四方向邻接图(FDAG)及其在表单字段定位中的应用
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602012
Jianxing Yuan, Y. Tang, C. Suen
A new non-hierarchical spatial data structure named four directional adjacency graphs (FDAG) is proposed. In the FDAG vertical and horizontal neighborhood relationship between rectangles is well represented so that structural information can be easily extracted. An application for structural analysis of forms is given, where experiments are conducted with positive results.
提出了一种新的非分层空间数据结构——四向邻接图。FDAG很好地表示了矩形之间的垂直和水平邻域关系,便于提取结构信息。给出了一种应用于形式结构分析的方法,并进行了实验,取得了积极的结果。
{"title":"Four directional adjacency graphs (FDAG) and their application in locating fields in forms","authors":"Jianxing Yuan, Y. Tang, C. Suen","doi":"10.1109/ICDAR.1995.602012","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602012","url":null,"abstract":"A new non-hierarchical spatial data structure named four directional adjacency graphs (FDAG) is proposed. In the FDAG vertical and horizontal neighborhood relationship between rectangles is well represented so that structural information can be easily extracted. An application for structural analysis of forms is given, where experiments are conducted with positive results.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124368685","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Multi-lingual, multi-font and multi-size large-set character recognition using self-organizing neural network 基于自组织神经网络的多语种、多字体、多字号大字符集字符识别
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598937
Seong-Whan Lee, Jongyeol Kim
We propose a practical scheme for multilingual multi font, and multi size large set character recognition using self organizing neural network. In order to improve the performance of the proposed scheme, a nonlinear shape normalization based on dot density and three kinds of hierarchical features are introduced. For coarse classification, two kinds of classifiers are proposed. One is a hierarchical tree classifier, and the other is a SOFM/LVQ based classifier which is composed of an adaptive SOFM coarse classifier and LVQ4 language classifiers. For fine classification, an LVQ4 classifier has been adopted. In order to evaluate the performance of the proposed scheme, recognition experiments with 3,367,200 characters having 7320 different classes have been carried out on a 486 DX-2 66 MHz PC. Experimental results reveal that the proposed scheme using an adaptive SOFM coarse classifier, LVQ4 language classifiers, and LVQ4 fine classifiers has a high recognition rate of over 98.27% and a fast execution time of more than 40 characters per second.
提出了一种基于自组织神经网络的多语言、多字体、多尺寸大字符集识别方案。为了提高算法的性能,引入了一种基于点密度和三种层次特征的非线性形状归一化方法。对于粗分类,提出了两种分类器。一种是层次树分类器,另一种是基于SOFM/LVQ的分类器,它由自适应SOFM粗分类器和LVQ4语言分类器组成。对于精细分类,采用LVQ4分类器。为了评估该方案的性能,在一台486 DX-2 66 MHz的PC机上进行了3367,200个字符、7320个不同类别的识别实验。实验结果表明,该方案采用自适应SOFM粗分类器、LVQ4语言分类器和LVQ4精细分类器,识别率达到98.27%以上,执行时间达到每秒40个字符以上。
{"title":"Multi-lingual, multi-font and multi-size large-set character recognition using self-organizing neural network","authors":"Seong-Whan Lee, Jongyeol Kim","doi":"10.1109/ICDAR.1995.598937","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598937","url":null,"abstract":"We propose a practical scheme for multilingual multi font, and multi size large set character recognition using self organizing neural network. In order to improve the performance of the proposed scheme, a nonlinear shape normalization based on dot density and three kinds of hierarchical features are introduced. For coarse classification, two kinds of classifiers are proposed. One is a hierarchical tree classifier, and the other is a SOFM/LVQ based classifier which is composed of an adaptive SOFM coarse classifier and LVQ4 language classifiers. For fine classification, an LVQ4 classifier has been adopted. In order to evaluate the performance of the proposed scheme, recognition experiments with 3,367,200 characters having 7320 different classes have been carried out on a 486 DX-2 66 MHz PC. Experimental results reveal that the proposed scheme using an adaptive SOFM coarse classifier, LVQ4 language classifiers, and LVQ4 fine classifiers has a high recognition rate of over 98.27% and a fast execution time of more than 40 characters per second.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"512 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127604269","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Construction of generic models of document structures using inference of tree grammars 使用树形语法推理构建文档结构的通用模型
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598977
O. Akindele, A. Belaïd
The use of generic model for a document class as the knowledge base in a Document Analysis System facilitates the analysis and understanding of documents belonging to this class. Nevertheless, absence of tools permitting the acquisition of this type of model is an hindrance to the conception of entirely automatic systems. In this paper, we present a method for acquiring the generic model for a document class from document samples belonging to this class. Our method is based on Inference of Tree Grammars and combination of ODA-like generic constructors. The method constructs specific physical structure for each sample and invites the user to assign logical labels to its components. From these logically labeled specific structures, it generates and modifies the generic model for the class under treatment.
在文档分析系统中,使用文档类的通用模型作为知识库,便于对属于该类的文档进行分析和理解。然而,缺乏允许获取这种类型模型的工具是完全自动化系统概念的障碍。本文提出了一种从属于文档类的文档样本中获取该类通用模型的方法。我们的方法是基于树语法推理和类oda泛型构造函数的组合。该方法为每个样本构建特定的物理结构,并请用户为其组件分配逻辑标签。从这些逻辑标记的特定结构中,它生成并修改所处理的类的通用模型。
{"title":"Construction of generic models of document structures using inference of tree grammars","authors":"O. Akindele, A. Belaïd","doi":"10.1109/ICDAR.1995.598977","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598977","url":null,"abstract":"The use of generic model for a document class as the knowledge base in a Document Analysis System facilitates the analysis and understanding of documents belonging to this class. Nevertheless, absence of tools permitting the acquisition of this type of model is an hindrance to the conception of entirely automatic systems. In this paper, we present a method for acquiring the generic model for a document class from document samples belonging to this class. Our method is based on Inference of Tree Grammars and combination of ODA-like generic constructors. The method constructs specific physical structure for each sample and invites the user to assign logical labels to its components. From these logically labeled specific structures, it generates and modifies the generic model for the class under treatment.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121719442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
The Delta LogNormal theory for the generation and modeling of cursive characters 草书字生成与建模的对数正态理论
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599042
W. Guerfali, R. Plamondon
We exploit the Delta LogNormal theory, a powerful tool for the generation and modeling of rapid movements to generate curvilinear strokes and constituting letters that respect both the dynamics and the appearance of movements made by a human. A theoretical analysis of the effects of the various parameters of the model is carried out: first, to reduce the size of the representation space of the letter models; and second, to select the parameters that constitute the optimal conditions for representing various symbols.
我们利用Delta LogNormal理论,这是一个强大的工具,用于快速运动的生成和建模,以生成曲线笔画,并构成尊重人类运动的动态和外观的字母。对模型各参数的影响进行了理论分析:首先,减小字母模型的表示空间大小;其次,选择构成各种符号表示的最优条件的参数。
{"title":"The Delta LogNormal theory for the generation and modeling of cursive characters","authors":"W. Guerfali, R. Plamondon","doi":"10.1109/ICDAR.1995.599042","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599042","url":null,"abstract":"We exploit the Delta LogNormal theory, a powerful tool for the generation and modeling of rapid movements to generate curvilinear strokes and constituting letters that respect both the dynamics and the appearance of movements made by a human. A theoretical analysis of the effects of the various parameters of the model is carried out: first, to reduce the size of the representation space of the letter models; and second, to select the parameters that constitute the optimal conditions for representing various symbols.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132523174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
A hidden Markov model extension of a neural predictive system for on-line character recognition 在线字符识别神经预测系统的隐马尔可夫模型扩展
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598942
S. Garcia-Salicetti, B. Dorizzi, P. Gallinari, A. Mellouk, D. Fanchon
The authors present a neural predictive system for on-line writer-independent character recognition. The data collection of each letter contains the pen trajectory information recorded by a digitizing tablet. Each letter is modeled by a fixed number of predictive neural networks (NN), so that a different multilayer NN models successive parts of a letter. The topology of each letter-model only permits transitions from each NN to itself or to its neighbors. In order to deal with the great variability proper to cursive handwriting in the omni-scriptor framework, they implement a holistic approach during both learning and recognition by performing adaptive segmentation. Also, the recognition step implements interactive recognition and segmentation. The approach compares neural techniques combined with dynamic programming to its extension to the hidden Markov model (HMM) framework. The first system gives quite good recognition rates on letter databases obtained from 10 different writers, and results improve considerably when one considers the extension of the first system to the durational HMM framework.
提出了一种用于在线字符识别的神经预测系统。每个字母的数据收集包含由数字化平板记录的笔轨迹信息。每个字母由固定数量的预测神经网络(NN)建模,以便不同的多层神经网络对字母的连续部分进行建模。每个字母模型的拓扑结构只允许从每个NN到它自己或它的邻居的转换。为了处理全脚本框架中草书笔迹的巨大可变性,他们通过执行自适应分割在学习和识别过程中实现了整体方法。识别步骤实现了交互式识别和分割。该方法将神经网络技术与动态规划相结合,将其扩展到隐马尔可夫模型(HMM)框架。第一个系统对来自10个不同写信人的字母数据库给出了相当好的识别率,当人们考虑将第一个系统扩展到持续HMM框架时,结果得到了很大的改善。
{"title":"A hidden Markov model extension of a neural predictive system for on-line character recognition","authors":"S. Garcia-Salicetti, B. Dorizzi, P. Gallinari, A. Mellouk, D. Fanchon","doi":"10.1109/ICDAR.1995.598942","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598942","url":null,"abstract":"The authors present a neural predictive system for on-line writer-independent character recognition. The data collection of each letter contains the pen trajectory information recorded by a digitizing tablet. Each letter is modeled by a fixed number of predictive neural networks (NN), so that a different multilayer NN models successive parts of a letter. The topology of each letter-model only permits transitions from each NN to itself or to its neighbors. In order to deal with the great variability proper to cursive handwriting in the omni-scriptor framework, they implement a holistic approach during both learning and recognition by performing adaptive segmentation. Also, the recognition step implements interactive recognition and segmentation. The approach compares neural techniques combined with dynamic programming to its extension to the hidden Markov model (HMM) framework. The first system gives quite good recognition rates on letter databases obtained from 10 different writers, and results improve considerably when one considers the extension of the first system to the durational HMM framework.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129963748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Power functions and their use in selecting distance functions for document degradation model validation 幂函数及其在选择用于文档退化模型验证的距离函数中的应用
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602007
T. Kanungo, R. Haralick, H. Baird
Two document degradation models that model the perturbations introduced during the document printing and scanning process were proposed recently. Although degradation models are very useful, it is very important that we validate these models by comparing the synthetically generated images against real images. In recent past, two different validation procedures have also been proposed to validate such document degradation models. These validation procedures are functions of sample size and various distance functions. In this paper we outline a statistical methodology to compare the various validation schemes that result by using different distance functions. This methodology is general enough to compare any two validation schemes.
最近提出了两个文档退化模型来模拟文档打印和扫描过程中引入的扰动。虽然退化模型非常有用,但通过将合成的图像与真实图像进行比较来验证这些模型是非常重要的。最近,人们还提出了两种不同的验证过程来验证此类文档退化模型。这些验证程序是样本量和各种距离函数的函数。在本文中,我们概述了一种统计方法来比较使用不同距离函数产生的各种验证方案。这种方法足够通用,可以比较任何两个验证方案。
{"title":"Power functions and their use in selecting distance functions for document degradation model validation","authors":"T. Kanungo, R. Haralick, H. Baird","doi":"10.1109/ICDAR.1995.602007","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602007","url":null,"abstract":"Two document degradation models that model the perturbations introduced during the document printing and scanning process were proposed recently. Although degradation models are very useful, it is very important that we validate these models by comparing the synthetically generated images against real images. In recent past, two different validation procedures have also been proposed to validate such document degradation models. These validation procedures are functions of sample size and various distance functions. In this paper we outline a statistical methodology to compare the various validation schemes that result by using different distance functions. This methodology is general enough to compare any two validation schemes.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"119 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133887331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Reading encrypted postal indicia 读取加密的邮政标志
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602075
Mark Cullen, L. Pintsov, Brian Romansky
The next generation of postal processing equipment will incorporate some method of verifying the postal revenue block (postal indicia) as a means of reducing postal fraud. The introduction of new digital printing technologies necessitates the encryption of revenue block information. This paper presents an approach for the verification process which includes algorithms for reading an encrypted postal indicia. In particular, postal indicia reading is tested for robustness against a variety of printing and media characteristics, and potential defects.
下一代邮政处理设备将采用一些验证邮政收入块(邮政标志)的方法,作为减少邮政欺诈的一种手段。新的数字印刷技术的引入需要对收入块信息进行加密。本文提出了一种验证过程的方法,其中包括读取加密邮政标志的算法。特别是,邮政标志阅读测试对各种印刷和媒体特性的稳健性,以及潜在的缺陷。
{"title":"Reading encrypted postal indicia","authors":"Mark Cullen, L. Pintsov, Brian Romansky","doi":"10.1109/ICDAR.1995.602075","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602075","url":null,"abstract":"The next generation of postal processing equipment will incorporate some method of verifying the postal revenue block (postal indicia) as a means of reducing postal fraud. The introduction of new digital printing technologies necessitates the encryption of revenue block information. This paper presents an approach for the verification process which includes algorithms for reading an encrypted postal indicia. In particular, postal indicia reading is tested for robustness against a variety of printing and media characteristics, and potential defects.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131645729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Column segmentation by white space pattern matching 通过空格模式匹配的列分割
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598960
M. Ozaki
Model-based column segmentation is described. Sequences of horizontal white space across a column are used as the basic features. Structures of columns in a specific publication are described by two levels of regular expressions: column expressions (CE) and element expressions (EE). Additional spatial constraints for element attributes can be described. A CE represents patterns of element sequences. An EE represents patterns of white space sequences for each element type. Segmentation is performed in three steps: element candidate extraction using EEs, column structure verification using the CE and ranking by comparison with statistical data. Experiments were performed on columns in two different scientific journals. More than 70% of the columns were correctly segmented as the top choice and more than 87% were in the top three choices. When spatial constraints were applied to element attributes, the rate was more than 90%.
描述了基于模型的列分割。横贯一列的水平留白序列被用作基本特征。特定发布中列的结构由两个级别的正则表达式描述:列表达式(CE)和元素表达式(EE)。可以描述元素属性的附加空间约束。CE表示元素序列的模式。EE表示每个元素类型的空白序列模式。分割分三个步骤进行:使用EEs提取候选元素,使用CE验证列结构,并通过与统计数据的比较进行排序。实验是在两种不同科学期刊的专栏上进行的。超过70%的列被正确分割为首选,超过87%的列在前三个选择中。当空间约束应用于元素属性时,成功率大于90%。
{"title":"Column segmentation by white space pattern matching","authors":"M. Ozaki","doi":"10.1109/ICDAR.1995.598960","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598960","url":null,"abstract":"Model-based column segmentation is described. Sequences of horizontal white space across a column are used as the basic features. Structures of columns in a specific publication are described by two levels of regular expressions: column expressions (CE) and element expressions (EE). Additional spatial constraints for element attributes can be described. A CE represents patterns of element sequences. An EE represents patterns of white space sequences for each element type. Segmentation is performed in three steps: element candidate extraction using EEs, column structure verification using the CE and ranking by comparison with statistical data. Experiments were performed on columns in two different scientific journals. More than 70% of the columns were correctly segmented as the top choice and more than 87% were in the top three choices. When spatial constraints were applied to element attributes, the rate was more than 90%.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131202398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Proceedings of 3rd International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1