首页 > 最新文献

Proceedings of 3rd International Conference on Document Analysis and Recognition最新文献

英文 中文
A Markovian random field approach to information retrieval 一种信息检索的马尔可夫随机场方法
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602070
D. Bouchaffra, J. Meunier
A Markovian random field approach is proposed for automatic information retrieval in full text documents. We draw up an analogy between a flow of queries/document images connections and statistical mechanics systems. The Markovian flow process machine (MFP) models the interaction between queries and document images as a dynamical system. The MFP machine searches to fit the user's queries by changing the set of descriptors contained in the document images. There is hence a constant transformation of the informational states of the fund. For each state, a certain degradation of the system is considered. We use simulated annealing algorithm to isolate low energy states: this corresponds to the best "matching" in some sense between queries and images.
提出了一种用于全文文档信息自动检索的马尔可夫随机场方法。我们将查询流/文档图像连接与统计力学系统进行类比。马尔可夫流处理机(MFP)将查询和文档图像之间的交互建模为一个动态系统。MFP机器通过更改文档图像中包含的描述符集来搜索以适应用户的查询。因此,国际货币基金组织的信息状态不断发生变化。对于每个状态,都考虑了系统的一定退化。我们使用模拟退火算法来隔离低能态:这在某种意义上对应于查询和图像之间的最佳“匹配”。
{"title":"A Markovian random field approach to information retrieval","authors":"D. Bouchaffra, J. Meunier","doi":"10.1109/ICDAR.1995.602070","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602070","url":null,"abstract":"A Markovian random field approach is proposed for automatic information retrieval in full text documents. We draw up an analogy between a flow of queries/document images connections and statistical mechanics systems. The Markovian flow process machine (MFP) models the interaction between queries and document images as a dynamical system. The MFP machine searches to fit the user's queries by changing the set of descriptors contained in the document images. There is hence a constant transformation of the informational states of the fund. For each state, a certain degradation of the system is considered. We use simulated annealing algorithm to isolate low energy states: this corresponds to the best \"matching\" in some sense between queries and images.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129005742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Extracting individual features from moments for Chinese writer identification 基于时刻特征的中文作者识别
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599030
Cheng-Lin Liu, Ru-Wei Dai, Ying-Jian Liu
To solve the problem of writer identification (WI) with indeterminate classes (writers) and objects (characters), it is a good way to extract individual features with clear physical meanings and small dynamic ranges. In this paper, a new method named Moment-Based Feature Method to identify Chinese writers is presented in which normalized individual features are derived from geometric moments of character images. The extracted features are invariant under translation, scaling, and stroke-width. They are explicitly corresponding to human perception of shape and distribute their values in small dynamic ranges. Experiments of writer recognition and verification are implemented to demonstrate the efficiency of this method and promising results have been achieved.
为了解决类(写作者)和对象(字符)不确定的写作者识别(WI)问题,提取物理意义明确、动态范围小的个体特征是一种很好的方法。本文提出了一种基于矩的汉字特征识别方法,该方法从汉字图像的几何矩中提取归一化的个体特征。提取的特征在平移、缩放和描边宽度下都是不变的。它们明确地对应于人类对形状的感知,并将它们的值分布在小的动态范围内。作者识别与验证实验验证了该方法的有效性,并取得了良好的效果。
{"title":"Extracting individual features from moments for Chinese writer identification","authors":"Cheng-Lin Liu, Ru-Wei Dai, Ying-Jian Liu","doi":"10.1109/ICDAR.1995.599030","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599030","url":null,"abstract":"To solve the problem of writer identification (WI) with indeterminate classes (writers) and objects (characters), it is a good way to extract individual features with clear physical meanings and small dynamic ranges. In this paper, a new method named Moment-Based Feature Method to identify Chinese writers is presented in which normalized individual features are derived from geometric moments of character images. The extracted features are invariant under translation, scaling, and stroke-width. They are explicitly corresponding to human perception of shape and distribute their values in small dynamic ranges. Experiments of writer recognition and verification are implemented to demonstrate the efficiency of this method and promising results have been achieved.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129411856","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Drawing capturing system using image enhancement 采用图像增强的绘图捕获系统
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.601980
Norio Nakamura, K. Hosaka, Masakazu Nagura
The paper describes the properties of Ueda's (1985) image enhancement method for line drawings and its merit for practical use. This method can remove the line discontinuities or mis-connections caused by scanning errors. The method is applied to simple images to evaluate its effect quantitatively. The authors confirm that it is more efficient than any other methods, and propose a drawing capturing system based on this method that can build up high quality drawing databases faster than any other system.
本文介绍了Ueda(1985)的线条图图像增强方法的特点及其在实际应用中的优点。这种方法可以消除由于扫描误差引起的线路不连续或误接。将该方法应用于简单图像,定量评价其效果。在此基础上,作者提出了一个基于该方法的图形捕获系统,该系统可以比其他系统更快地建立高质量的图形数据库。
{"title":"Drawing capturing system using image enhancement","authors":"Norio Nakamura, K. Hosaka, Masakazu Nagura","doi":"10.1109/ICDAR.1995.601980","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.601980","url":null,"abstract":"The paper describes the properties of Ueda's (1985) image enhancement method for line drawings and its merit for practical use. This method can remove the line discontinuities or mis-connections caused by scanning errors. The method is applied to simple images to evaluate its effect quantitatively. The authors confirm that it is more efficient than any other methods, and propose a drawing capturing system based on this method that can build up high quality drawing databases faster than any other system.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130755617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A simplified attributed graph grammar for high-level music recognition 用于高级音乐识别的简化属性图语法
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602096
S. Baumann
This paper describes a simplified attributed programmed graph grammar to represent and process a-priori knowledge about common music notation. The presented approach serves as a high-level recognition stage and is interlocked to previous low-level recognition phases in our entire optical music recognition system (DOREMIDI++). The implemented grammar rules and control diagrams describe a declarative knowledge base to drive a transformation algorithm. This transformation converts the results of symbol recognition stages to a symbolic representation of the musical score.
本文描述了一种简化的属性程序化图语法,用于表示和处理普通乐谱的先验知识。所提出的方法作为一个高级识别阶段,并在我们的整个光学音乐识别系统(doremid++)中与先前的低级识别阶段互锁。实现的语法规则和控制图描述了用于驱动转换算法的声明性知识库。这种转换将符号识别阶段的结果转换为乐谱的符号表示。
{"title":"A simplified attributed graph grammar for high-level music recognition","authors":"S. Baumann","doi":"10.1109/ICDAR.1995.602096","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602096","url":null,"abstract":"This paper describes a simplified attributed programmed graph grammar to represent and process a-priori knowledge about common music notation. The presented approach serves as a high-level recognition stage and is interlocked to previous low-level recognition phases in our entire optical music recognition system (DOREMIDI++). The implemented grammar rules and control diagrams describe a declarative knowledge base to drive a transformation algorithm. This transformation converts the results of symbol recognition stages to a symbolic representation of the musical score.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129835090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
A system for scanning and segmenting cursively handwritten words into basic strokes 一种扫描和分割草书手写文字为基本笔画的系统
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602082
C. Privitera, R. Plamondon
This paper presents a segmentation method that partly mimics the cognitive-behavioral process used by human subjects to recover motor-temporal information from the image of a handwritten word. The approach does not exploit any thinning procedure, but rather a different typology of information is manipulated concerning the curvature of the word contour. Starting from the maximum curvature points roughly corresponding to the beginning of a stroke, the algorithm scans the word, following the natural course of the line and attempts to repeat the same movement as executed by the writer during the generation of the word. At each maximum curvature point, the line is segmented and reconstructed by a smooth interpolation of the most interior points belonging to the line just covered. At the end of the scanning process, a temporal sequence of motor strokes is obtained which plausibly composes the original intended movement.
本文提出了一种部分模仿人类受试者从手写文字图像中恢复运动时间信息的认知行为过程的分割方法。该方法不利用任何细化过程,而是操纵关于单词轮廓曲率的不同类型的信息。该算法从大致对应笔画开始的最大曲率点开始扫描单词,遵循线条的自然轨迹,并尝试重复作者在生成单词时所执行的相同运动。在每个最大曲率点处,通过对属于刚刚覆盖的线的最内部点进行平滑插值来分割和重建该线。在扫描过程结束时,获得的时间序列的运动冲程似乎组成了原来的预期运动。
{"title":"A system for scanning and segmenting cursively handwritten words into basic strokes","authors":"C. Privitera, R. Plamondon","doi":"10.1109/ICDAR.1995.602082","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602082","url":null,"abstract":"This paper presents a segmentation method that partly mimics the cognitive-behavioral process used by human subjects to recover motor-temporal information from the image of a handwritten word. The approach does not exploit any thinning procedure, but rather a different typology of information is manipulated concerning the curvature of the word contour. Starting from the maximum curvature points roughly corresponding to the beginning of a stroke, the algorithm scans the word, following the natural course of the line and attempts to repeat the same movement as executed by the writer during the generation of the word. At each maximum curvature point, the line is segmented and reconstructed by a smooth interpolation of the most interior points belonging to the line just covered. At the end of the scanning process, a temporal sequence of motor strokes is obtained which plausibly composes the original intended movement.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128491004","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
False hits of tri-syllabic queries in a Chinese signature file 中文签名文件中三音节查询错误命中
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598966
Tyne Liang, Suh-Yin Lee, Wei-Pang Yang
In the application of the superimposed coding method to character-based Chinese text retrieval we find two kinds of false hits for a multi-syllabic (multicharacter) query. The first type is a random false hit (RFH) which is due to accidental setting of bits by irrelevant characters in a document signature. The other type is an adjacency false hit (AFH) which is due to the loss of character sequence information in signature creation. Since many query terms are proper nouns and Chinese names which often contain three characters (tri-syllabic), we derive a formula to estimate the RFH for trisyllabic queries. As for the AFH which cannot be reduced by single character (monogram) hashing method, a method which hashes consecutive character pairs (bigram) is designed to reduce both the AFH and the RFH. We find that there exists an optimal weight assignment for a minimal false hit rate in a combined scheme which encodes both monogram and bigram keys in document signatures.
在将叠加编码方法应用于基于字符的中文文本检索中,对一个多音节(多字符)查询发现了两种错误命中。第一种类型是随机错误命中(RFH),这是由于文档签名中不相关的字符偶然设置了位。另一种类型是邻接错误命中(AFH),这是由于签名创建过程中字符序列信息的丢失。由于许多查询词是专有名词和中文名称,通常包含三个字符(三音节),我们推导了一个公式来估计三音节查询的RFH。针对单字符(字母组合)哈希法无法减少的AFH,设计了一种对连续字符对(双字符)进行哈希的方法来同时减少AFH和RFH。我们发现在一个同时编码字母组合和双字母组合密钥的文件签名方案中存在一个最小错误命中率的最优权值分配。
{"title":"False hits of tri-syllabic queries in a Chinese signature file","authors":"Tyne Liang, Suh-Yin Lee, Wei-Pang Yang","doi":"10.1109/ICDAR.1995.598966","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598966","url":null,"abstract":"In the application of the superimposed coding method to character-based Chinese text retrieval we find two kinds of false hits for a multi-syllabic (multicharacter) query. The first type is a random false hit (RFH) which is due to accidental setting of bits by irrelevant characters in a document signature. The other type is an adjacency false hit (AFH) which is due to the loss of character sequence information in signature creation. Since many query terms are proper nouns and Chinese names which often contain three characters (tri-syllabic), we derive a formula to estimate the RFH for trisyllabic queries. As for the AFH which cannot be reduced by single character (monogram) hashing method, a method which hashes consecutive character pairs (bigram) is designed to reduce both the AFH and the RFH. We find that there exists an optimal weight assignment for a minimal false hit rate in a combined scheme which encodes both monogram and bigram keys in document signatures.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125858557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A high quality vectorization combining local quality measures and global constraints 结合局部质量度量和全局约束的高质量矢量化
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598986
M. Röösli, G. Monagan
We present a vectorization system to generate vector data which corresponds to the line structures of a raster image. The vector data consists of the primitives: "straight lane segment" and "circular arc". The vectorization system measures the quality of each primitive generated. Thus, the vectorization does not only produce high quality vector data, it also gives a precise description of the quality of the data generated. This is crucial if the requirements set by industrial applications are to be met. In order not to lose the quality of the vector data while constructing primitives into line objects, geometric constraints are incorporated already at the vectorization level: constraints like requiring segments to be parallel or perpendicular, circular arcs to be concentric, or tangents of the primitives to be equal at their connection point. After the constraints have been satisfied the resulting primitives still fulfil the quality requirements as before the constraints were imposed. The possibility to refit the generated vector data under adapted constraints allows for an efficient interactive postprocessing of the data.
我们提出了一个矢量化系统来生成与光栅图像的线结构相对应的矢量数据。向量数据由“直线段”和“圆弧”这两个原语组成。矢量化系统测量生成的每个原语的质量。因此,矢量化不仅可以产生高质量的矢量数据,还可以精确描述生成的数据的质量。如果要满足工业应用设定的要求,这一点至关重要。为了在将原语构建为线对象时不失去矢量数据的质量,几何约束已经在矢量化级别中加入:诸如要求线段平行或垂直,圆弧同心,或原语的切线在连接点相等的约束。在满足约束条件之后,结果原语仍然像施加约束条件之前一样满足质量要求。在适应的约束下重新构造生成的矢量数据的可能性允许对数据进行有效的交互式后处理。
{"title":"A high quality vectorization combining local quality measures and global constraints","authors":"M. Röösli, G. Monagan","doi":"10.1109/ICDAR.1995.598986","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598986","url":null,"abstract":"We present a vectorization system to generate vector data which corresponds to the line structures of a raster image. The vector data consists of the primitives: \"straight lane segment\" and \"circular arc\". The vectorization system measures the quality of each primitive generated. Thus, the vectorization does not only produce high quality vector data, it also gives a precise description of the quality of the data generated. This is crucial if the requirements set by industrial applications are to be met. In order not to lose the quality of the vector data while constructing primitives into line objects, geometric constraints are incorporated already at the vectorization level: constraints like requiring segments to be parallel or perpendicular, circular arcs to be concentric, or tangents of the primitives to be equal at their connection point. After the constraints have been satisfied the resulting primitives still fulfil the quality requirements as before the constraints were imposed. The possibility to refit the generated vector data under adapted constraints allows for an efficient interactive postprocessing of the data.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"435 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126104624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
ODIL: an SGML description language of the layout structure of documents ODIL:文档布局结构的SGML描述语言
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599040
P. Lefèvre, François Reynaud
This paper describes a coding format in SGML for the output of a document recognition prototype. Our proposal is a DTD named "ODIL"-Office Document Image description Language-that describes precisely the layout structure of a document after all recognition phases, including OCR. All layout objects of a document are defined in the form of SGML elements, and their characteristics are defined by SGML attributes. The basic objects are blocks, containing homogeneous information. Five types of information are supported by the ODIL language: texts, photos, line graphics, tables, mathematic formulas. The ODIL representation of the recognition results is well adapted to a further logical structure recognition. Starting from the ODIL DTD and using the RAINBOW transit DTD will permit to use SGML tools for the logical structure recognition which is viewed as an SGML up-conversion problem.
本文描述了一种用于文档识别原型输出的SGML编码格式。我们的建议是一个名为“ODIL”(office Document Image description language)的DTD,它精确地描述文档在所有识别阶段(包括OCR)之后的布局结构。文档的所有布局对象都以SGML元素的形式定义,它们的特征由SGML属性定义。基本对象是块,包含同构信息。ODIL语言支持五种类型的信息:文本、照片、线条图、表格、数学公式。识别结果的ODIL表示很好地适应于进一步的逻辑结构识别。从ODIL DTD开始并使用RAINBOW传输DTD将允许使用SGML工具进行逻辑结构识别,这被视为SGML上转换问题。
{"title":"ODIL: an SGML description language of the layout structure of documents","authors":"P. Lefèvre, François Reynaud","doi":"10.1109/ICDAR.1995.599040","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599040","url":null,"abstract":"This paper describes a coding format in SGML for the output of a document recognition prototype. Our proposal is a DTD named \"ODIL\"-Office Document Image description Language-that describes precisely the layout structure of a document after all recognition phases, including OCR. All layout objects of a document are defined in the form of SGML elements, and their characteristics are defined by SGML attributes. The basic objects are blocks, containing homogeneous information. Five types of information are supported by the ODIL language: texts, photos, line graphics, tables, mathematic formulas. The ODIL representation of the recognition results is well adapted to a further logical structure recognition. Starting from the ODIL DTD and using the RAINBOW transit DTD will permit to use SGML tools for the logical structure recognition which is viewed as an SGML up-conversion problem.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127730608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Description and recognition of form and automated form data entry 描述和识别表单和自动表单数据输入
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.601963
Jinhui Liu, Xiaoqing Ding, Youshou Wu
In this paper we present a form description method, in which frame lines are used to constitute a so-called frame template, which reflects the structure of a form either topologically or geometrically. Relevant item traversal algorithm is then proposed to locate and label form's items. We have also developed a robust and fast frame line detection method to make this form description practical for form recognition. Experimental results show our approach provides an effective way to convert printed forms into computerized format or collect information for database from printed forms.
在本文中,我们提出了一种形式描述方法,其中框架线被用来构成一个所谓的框架模板,它反映了一个形式的结构拓扑或几何。然后提出了相应的项遍历算法来定位和标记表单中的项。我们还开发了一种鲁棒和快速的帧线检测方法,使这种形式描述适用于形式识别。实验结果表明,该方法为将打印表单转换为计算机格式或将打印表单信息收集到数据库中提供了有效的方法。
{"title":"Description and recognition of form and automated form data entry","authors":"Jinhui Liu, Xiaoqing Ding, Youshou Wu","doi":"10.1109/ICDAR.1995.601963","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.601963","url":null,"abstract":"In this paper we present a form description method, in which frame lines are used to constitute a so-called frame template, which reflects the structure of a form either topologically or geometrically. Relevant item traversal algorithm is then proposed to locate and label form's items. We have also developed a robust and fast frame line detection method to make this form description practical for form recognition. Experimental results show our approach provides an effective way to convert printed forms into computerized format or collect information for database from printed forms.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114364483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Knowledge-based derivation of document logical structure 基于知识的文档逻辑结构派生
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599038
Debashish Niyogi, S. Srihari
The analysis of a document image to derive a symbolic description of its structure and contents involves using spatial domain knowledge to classify the different printed blocks (e.g., text paragraphs), group them into logical units (e.g., newspaper stories), and determine the reading order of the text blocks within each unit. These steps describe the conversion of the physical structure of a document into its logical structure. We have developed a computational model for document logical structure derivation, in which a rule-based control strategy utilizes the data obtained from analyzing a digitized document image, and makes inferences using a multi-level knowledge base of document layout rules. The knowledge-based document logical structure derivation system (DeLoS) based on this model consists of a hierarchical rule-based control system to guide the block classification, grouping and read-ordering operations; a global data structure to store the document image data and incremental inferences; and a domain knowledge base to encode the rules governing document layout.
对文档图像进行分析以获得其结构和内容的符号描述涉及到使用空间域知识对不同的印刷块(例如文本段落)进行分类,将它们分组为逻辑单元(例如报纸故事),并确定每个单元内文本块的阅读顺序。这些步骤描述了将文档的物理结构转换为其逻辑结构。我们建立了一个文档逻辑结构推导的计算模型,其中基于规则的控制策略利用从数字化文档图像中分析得到的数据,并使用多层次的文档布局规则知识库进行推理。基于该模型的基于知识的文档逻辑结构派生系统(DeLoS)包括一个分层的、基于规则的控制系统来指导块的分类、分组和读排序操作;用于存储文档图像数据和增量推断的全局数据结构;以及一个领域知识库,用于编码控制文档布局的规则。
{"title":"Knowledge-based derivation of document logical structure","authors":"Debashish Niyogi, S. Srihari","doi":"10.1109/ICDAR.1995.599038","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599038","url":null,"abstract":"The analysis of a document image to derive a symbolic description of its structure and contents involves using spatial domain knowledge to classify the different printed blocks (e.g., text paragraphs), group them into logical units (e.g., newspaper stories), and determine the reading order of the text blocks within each unit. These steps describe the conversion of the physical structure of a document into its logical structure. We have developed a computational model for document logical structure derivation, in which a rule-based control strategy utilizes the data obtained from analyzing a digitized document image, and makes inferences using a multi-level knowledge base of document layout rules. The knowledge-based document logical structure derivation system (DeLoS) based on this model consists of a hierarchical rule-based control system to guide the block classification, grouping and read-ordering operations; a global data structure to store the document image data and incremental inferences; and a domain knowledge base to encode the rules governing document layout.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"35 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114127740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
期刊
Proceedings of 3rd International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1