首页 > 最新文献

Proceedings of 3rd International Conference on Document Analysis and Recognition最新文献

英文 中文
Japanese document recognition based on interpolated n-gram model of character 基于字符插值n-图模型的日语文档识别
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598993
Hiroki Mori, Hirotomo Aso, S. Makino
N-gram model is widely applied to various pattern recognition system because it well represents local features of natural languages. In this paper, we describe a contextual postprocessing method using a trigram model of character for Japanese document recognition, and its advantage is revealed by practical experiments. The model is automatically obtained by statistical processing of training documents. The ability to reduce ambiguity is evaluated by the perplexity. In the processing, two smoothing methods are examined, and the predictive power of the deleted interpolation method is shown to be superior. For leading articles, the perplexity reduced to about 22 when using deleted interpolation. The output from OCR is processed very fast using a Viterbi algorithm. Experimental results of recognition for three kinds of documents show that the error correction rates are ranged from 75 to over 90 percent.
N-gram模型由于能很好地表征自然语言的局部特征,被广泛应用于各种模式识别系统中。本文提出了一种基于三字符模型的语境后处理方法,并通过实验验证了该方法的优越性。通过对训练文档进行统计处理,自动得到模型。减少歧义的能力是通过困惑度来评估的。在处理过程中,对两种平滑方法进行了比较,结果表明,删除插值方法的预测能力更强。对于主要文章,当使用删除插值时,困惑度减少到22左右。OCR的输出使用Viterbi算法处理得非常快。对三种文档的识别实验结果表明,该方法的错误率在75% ~ 90%以上。
{"title":"Japanese document recognition based on interpolated n-gram model of character","authors":"Hiroki Mori, Hirotomo Aso, S. Makino","doi":"10.1109/ICDAR.1995.598993","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598993","url":null,"abstract":"N-gram model is widely applied to various pattern recognition system because it well represents local features of natural languages. In this paper, we describe a contextual postprocessing method using a trigram model of character for Japanese document recognition, and its advantage is revealed by practical experiments. The model is automatically obtained by statistical processing of training documents. The ability to reduce ambiguity is evaluated by the perplexity. In the processing, two smoothing methods are examined, and the predictive power of the deleted interpolation method is shown to be superior. For leading articles, the perplexity reduced to about 22 when using deleted interpolation. The output from OCR is processed very fast using a Viterbi algorithm. Experimental results of recognition for three kinds of documents show that the error correction rates are ranged from 75 to over 90 percent.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126434114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A hypothesis testing approach to word recognition using an A* search algorithm 基于A*搜索算法的词识别假设检验方法
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599013
Chi Fang, J. Hull
An hypothesis testing approach for recognizing machine-printed words is presented in this paper. Based on knowledge of the document font and candidates for the identity of a word, this approach searches a tree of word decisions to generate and test hypotheses for character recognition and segmentation. The search starts at each sequential character position from both ends of a word image and proceeds inward. The accumulated cost of reaching a certain partial recognition decision is combined with the estimate of the potential cost to reach a goal state using an A* search algorithm. The proposed algorithm compensates for local degradations by relying on global characteristics of a word image. Tests of the algorithm show a recognition rate of 98.93% on degraded scanned document images with touching characters.
提出了一种机器打印单词识别的假设检验方法。基于文档字体的知识和候选词的身份,该方法搜索词决策树来生成和测试用于字符识别和分割的假设。搜索从单词图像两端的每个顺序字符位置开始,并向内进行。利用a *搜索算法,将达到某一局部识别决策的累积代价与达到目标状态的潜在代价相结合。该算法利用词图像的全局特征对局部退化进行补偿。实验表明,该算法对带有触摸字符的退化扫描文档图像的识别率达到98.93%。
{"title":"A hypothesis testing approach to word recognition using an A* search algorithm","authors":"Chi Fang, J. Hull","doi":"10.1109/ICDAR.1995.599013","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599013","url":null,"abstract":"An hypothesis testing approach for recognizing machine-printed words is presented in this paper. Based on knowledge of the document font and candidates for the identity of a word, this approach searches a tree of word decisions to generate and test hypotheses for character recognition and segmentation. The search starts at each sequential character position from both ends of a word image and proceeds inward. The accumulated cost of reaching a certain partial recognition decision is combined with the estimate of the potential cost to reach a goal state using an A* search algorithm. The proposed algorithm compensates for local degradations by relying on global characteristics of a word image. Tests of the algorithm show a recognition rate of 98.93% on degraded scanned document images with touching characters.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"2006 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125552697","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Numeral characters and capital letters segmentation recognition in mixed handwriting context 混合手写环境下的数字字符和大写字母分割识别
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602041
H. Wehbi, H. Oulhadj, J. Lemoine, É. Petit
For the analytic on-line recognition of handwriting, the range of pattern recognition problems can be described by the severity of letter segmentation required. More difficult problems require an interaction of letter segmentation and recognition. These problems include overlapping discretely written characters, pure cursive writing, and mixed cursive and discrete writing. To these problems concerning the letter segmentation, the word segmentation problems is added. Since a script can contain numbers, capital letters as well as lowercase letters, it is necessary to have a system able to recognize them. This paper describes an on-line system for identifying and recognizing numeral characters and capital letters in handwriting sentences. This system provides two segmentation modules: the first one is to isolate the word drawings within a sentence, and the other one is to separate numeral characters and capital letters from a mixed writing prior to their recognition.
对于笔迹的在线分析识别,模式识别问题的范围可以用字母分割的严重程度来描述。更困难的问题需要字母分割和识别的相互作用。这些问题包括重叠的离散字,纯草书,以及草书和离散字的混合。在字母分词问题的基础上,增加了分词问题。由于脚本可以包含数字、大写字母和小写字母,因此有必要让系统能够识别它们。本文介绍了一个手写句子中数字字符和大写字母的在线识别系统。该系统提供了两个分词模块:第一个模块是将句子中的单词图分离出来,另一个模块是将数字字符和大写字母从混合书写中分离出来,然后再进行识别。
{"title":"Numeral characters and capital letters segmentation recognition in mixed handwriting context","authors":"H. Wehbi, H. Oulhadj, J. Lemoine, É. Petit","doi":"10.1109/ICDAR.1995.602041","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602041","url":null,"abstract":"For the analytic on-line recognition of handwriting, the range of pattern recognition problems can be described by the severity of letter segmentation required. More difficult problems require an interaction of letter segmentation and recognition. These problems include overlapping discretely written characters, pure cursive writing, and mixed cursive and discrete writing. To these problems concerning the letter segmentation, the word segmentation problems is added. Since a script can contain numbers, capital letters as well as lowercase letters, it is necessary to have a system able to recognize them. This paper describes an on-line system for identifying and recognizing numeral characters and capital letters in handwriting sentences. This system provides two segmentation modules: the first one is to isolate the word drawings within a sentence, and the other one is to separate numeral characters and capital letters from a mixed writing prior to their recognition.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130449291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A fast algorithm for the minimum distance classifier and its application to Kanji character recognition 一种快速最小距离分类器算法及其在汉字识别中的应用
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598995
S. Senda, M. Minoh, I. Katsuo
A fast algorithm for the minimum distance classifier (MDC) is proposed. The MDC has been used in various areas of pattern recognition because it is simple and fast compared with other complicated classifiers. The algorithm proposed is much faster than the exhaustive one that calculates all the distances straighforwardly. Our algorithm, which produces the same output as the exhaustive, omits redundant calculations according to Karhunen-Loeve expansion. From the KL-expansion of the prototype patterns, we form a subspace of the feature space, in which the order of examining the prototypes is decided adaptive to a given unknown pattern. We have applied the algorithm to recognition of handprinted Kanji characters and measured its performance on the ETL9B database. As a result, the theoretical and practical speedups were 10-20 and 4-9, respectively.
提出了一种快速的最小距离分类器(MDC)算法。与其他复杂的分类器相比,MDC具有简单、快速的特点,已被广泛应用于模式识别的各个领域。该算法比直接计算所有距离的穷举算法要快得多。我们的算法根据Karhunen-Loeve展开省去了冗余计算,产生了与穷举式相同的输出。从原型模式的kl展开中,我们形成了特征空间的子空间,在该子空间中,根据给定的未知模式确定了检查原型的顺序。将该算法应用于手印汉字的识别,并在ETL9B数据库上进行了性能测试。结果,理论和实际加速分别为10-20和4-9。
{"title":"A fast algorithm for the minimum distance classifier and its application to Kanji character recognition","authors":"S. Senda, M. Minoh, I. Katsuo","doi":"10.1109/ICDAR.1995.598995","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598995","url":null,"abstract":"A fast algorithm for the minimum distance classifier (MDC) is proposed. The MDC has been used in various areas of pattern recognition because it is simple and fast compared with other complicated classifiers. The algorithm proposed is much faster than the exhaustive one that calculates all the distances straighforwardly. Our algorithm, which produces the same output as the exhaustive, omits redundant calculations according to Karhunen-Loeve expansion. From the KL-expansion of the prototype patterns, we form a subspace of the feature space, in which the order of examining the prototypes is decided adaptive to a given unknown pattern. We have applied the algorithm to recognition of handprinted Kanji characters and measured its performance on the ETL9B database. As a result, the theoretical and practical speedups were 10-20 and 4-9, respectively.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116599454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
An object-oriented model for drawing understanding and its ability of noise absorption 一种面向对象的绘图理解模型及其噪声吸收能力
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598990
Wei Wu, Wei Lu, M. Sakauchi
In this paper we propose a new framework of object-oriented model named MTDM (Matching Tree Driving Model) for drawing understanding and verify its ability of noise absorption. MTDM makes use of descriptions of object-oriented style and is an integration of static and dynamic description of recognition target. Static descriptions are for representation of abstract features so that description of structure and restriction become easier. At the same time static descriptions can be independent of matching procedures of recognition target. The dynamic descriptions are for matching control of recognition target in the form of tree structure named matching tree. Matching procedures for complex targets can be easily described with multiple matching trees. By application to several typical engineering drawings, particularly drawings with noises and distortions MTDM is proven to be suitable for multipurpose and multitarget platform.
本文提出了一种新的面向对象模型框架——匹配树驱动模型(Matching Tree Driving model, MTDM),用于绘图理解和验证其吸噪能力。MTDM采用面向对象的描述方式,是识别目标静态描述和动态描述的结合。静态描述用于表示抽象特征,以便更容易地描述结构和限制。同时,静态描述可以独立于识别目标的匹配过程。动态描述以树状结构的形式对识别目标进行匹配控制,称为匹配树。复杂目标的匹配过程可以用多个匹配树来描述。通过对几种典型工程图,特别是带有噪声和畸变的图的应用,证明了MTDM算法适用于多用途、多目标平台。
{"title":"An object-oriented model for drawing understanding and its ability of noise absorption","authors":"Wei Wu, Wei Lu, M. Sakauchi","doi":"10.1109/ICDAR.1995.598990","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598990","url":null,"abstract":"In this paper we propose a new framework of object-oriented model named MTDM (Matching Tree Driving Model) for drawing understanding and verify its ability of noise absorption. MTDM makes use of descriptions of object-oriented style and is an integration of static and dynamic description of recognition target. Static descriptions are for representation of abstract features so that description of structure and restriction become easier. At the same time static descriptions can be independent of matching procedures of recognition target. The dynamic descriptions are for matching control of recognition target in the form of tree structure named matching tree. Matching procedures for complex targets can be easily described with multiple matching trees. By application to several typical engineering drawings, particularly drawings with noises and distortions MTDM is proven to be suitable for multipurpose and multitarget platform.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"26 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120937177","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Document registration using projective geometry 文档注册使用射影几何
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602128
R. Safari, N. Narasimhamurthi, M. Shridhar, M. Ahmadi
In this paper, a technique for registering filled-in forms is presented. The technique determines the transformations that is required to convert a filled-in form to match a known master and then extracts filled-in information. This method involves determining corresponding points between the master and the filled-in form and using this correspondence to determine the appropriate transformation. The correspondence problem is solved using results from projective geometry.
本文介绍了一种注册已填写表单的技术。该技术确定将已填写的表单转换为与已知主表单匹配所需的转换,然后提取已填写的信息。此方法包括确定主表单和已填写表单之间的对应点,并使用此对应关系来确定适当的转换。用射影几何的结果解决了对应问题。
{"title":"Document registration using projective geometry","authors":"R. Safari, N. Narasimhamurthi, M. Shridhar, M. Ahmadi","doi":"10.1109/ICDAR.1995.602128","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602128","url":null,"abstract":"In this paper, a technique for registering filled-in forms is presented. The technique determines the transformations that is required to convert a filled-in form to match a known master and then extracts filled-in information. This method involves determining corresponding points between the master and the filled-in form and using this correspondence to determine the appropriate transformation. The correspondence problem is solved using results from projective geometry.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"157 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132909443","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Spatial sampling effects in optical character recognition 光学字符识别中的空间采样效应
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599001
D. Lopresti, Jiangying Zhou, G. Nagy, Prateek Sarkar
In this paper we examine the effects of random-phase spatial sampling on the optical character recognition process. We start by presenting a detailed analysis in the case of 1-dimensional patterns. Empirical data demonstrate that our model is accurate. We then give experimental results for more complex, 2-dimensional patterns (i.e. printed, scanned characters). Spatial sampling seems to account for a significant amount of the variability seen in practice.
本文研究了随机相位空间采样对光学字符识别过程的影响。我们首先对一维模式进行详细的分析。实证数据表明,我们的模型是准确的。然后,我们给出了更复杂的二维模式(即打印、扫描字符)的实验结果。空间采样似乎解释了实践中所见的大量可变性。
{"title":"Spatial sampling effects in optical character recognition","authors":"D. Lopresti, Jiangying Zhou, G. Nagy, Prateek Sarkar","doi":"10.1109/ICDAR.1995.599001","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599001","url":null,"abstract":"In this paper we examine the effects of random-phase spatial sampling on the optical character recognition process. We start by presenting a detailed analysis in the case of 1-dimensional patterns. Empirical data demonstrate that our model is accurate. We then give experimental results for more complex, 2-dimensional patterns (i.e. printed, scanned characters). Spatial sampling seems to account for a significant amount of the variability seen in practice.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134296729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A formal model for document processing of business forms 用于业务表单文档处理的正式模型
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598978
M. Cheriet, J. N. Said, C. Suen
We present a formal model for processing gray-scale images of business forms such as bank cheques. The formal model is based on a new hybrid-based approach namely the base lines. In fact, to segment handwritten and hand-printed data from bank cheques, knowledge rules and base lines will have important roles to segment and extract the information from bank cheques. The architectural design as well as the major components of the system is discussed in full detail. Moreover, the significant use of the morphological followed by the topological processing on gray-scale images is used as a major aspect to restore the lost information after the elimination of the background and the base lines from the gray-scale cheques.
我们提出了一个正式的模型,用于处理商业形式的灰度图像,如银行支票。正式模型是基于一种新的基于混合的方法,即基线。事实上,为了从银行支票中分割手写和手印数据,知识规则和基线对于从银行支票中分割和提取信息将起着重要的作用。详细讨论了系统的体系结构设计和主要组成部分。此外,对灰度图像进行形态学处理后再进行拓扑处理是在灰度检查中消除背景和基线后恢复丢失信息的主要方面。
{"title":"A formal model for document processing of business forms","authors":"M. Cheriet, J. N. Said, C. Suen","doi":"10.1109/ICDAR.1995.598978","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598978","url":null,"abstract":"We present a formal model for processing gray-scale images of business forms such as bank cheques. The formal model is based on a new hybrid-based approach namely the base lines. In fact, to segment handwritten and hand-printed data from bank cheques, knowledge rules and base lines will have important roles to segment and extract the information from bank cheques. The architectural design as well as the major components of the system is discussed in full detail. Moreover, the significant use of the morphological followed by the topological processing on gray-scale images is used as a major aspect to restore the lost information after the elimination of the background and the base lines from the gray-scale cheques.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"172 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132246066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Near-wordless document structure classification 近乎无词的文档结构分类
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599036
K. Summers
Automatic derivation of logical document structure from generic layout would enable the development of many highly flexible electronic document manipulation tools. This problem can be divided into the segmentation of text into pieces and the classification of these pieces as particular logical structures. This paper proposes an approach to the classification of logical document structures, according to their distance from predefined prototypes. The prototypes consider linguistic information minimally, thus relying minimally on the accuracy of OCR and decreasing language-dependence. Different classes of logical structures and the differences in the requisite information for classifying them are discussed. A prototype format is proposed, existing prototypes and a distance measurement are described, and performance results are provided.
从通用布局中自动派生逻辑文档结构将使开发许多高度灵活的电子文档操作工具成为可能。这个问题可以分为将文本分割成小块和将这些小块分类为特定的逻辑结构。本文提出了一种根据逻辑文档结构与预定义原型的距离对其进行分类的方法。原型对语言信息的考虑最少,因此对OCR准确性的依赖最小,减少了语言依赖性。讨论了不同类型的逻辑结构及其分类所需信息的差异。提出了一种原型格式,描述了现有的原型和距离测量,并提供了性能结果。
{"title":"Near-wordless document structure classification","authors":"K. Summers","doi":"10.1109/ICDAR.1995.599036","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599036","url":null,"abstract":"Automatic derivation of logical document structure from generic layout would enable the development of many highly flexible electronic document manipulation tools. This problem can be divided into the segmentation of text into pieces and the classification of these pieces as particular logical structures. This paper proposes an approach to the classification of logical document structures, according to their distance from predefined prototypes. The prototypes consider linguistic information minimally, thus relying minimally on the accuracy of OCR and decreasing language-dependence. Different classes of logical structures and the differences in the requisite information for classifying them are discussed. A prototype format is proposed, existing prototypes and a distance measurement are described, and performance results are provided.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"297 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132369823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Interactive acquisition of thematic information of Chinese verbs for judicial verdict document understanding using templates, syntactic clues, and heuristics 基于模板、句法线索和启发式的汉语动词主位信息交互习得研究
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.598998
K. H. Lin, Rey-Long Liu, V. Soo
The thematic knowledge can bridge the gap between semantic entities and syntactic constituents. In document understanding, the correctness and the efficiency could be improved if the thematic knowledge is available. In this paper, we propose a semi-automatic method to acquire thematic knowledge of Chinese verbs by exploiting syntactic clues. The syntactic clues, which may be collected by most existing syntactic processors, reduce the hypothesis space of the theta roles. The ambiguities may be further resolved by the evidences from a trainer. A set of heuristics based on linguistic constraints are employed to guide the ambiguity resolution process. To acquire thematic information for verbs, the argument structures of the verbs must be extracted first. A template matching method is used to extract the argument structure of verbs.
主题知识可以在语义实体和句法成分之间架起桥梁。在文档理解中,如果有主题知识,可以提高文档理解的正确性和效率。本文提出了一种利用句法线索半自动获取汉语动词主位知识的方法。句法线索可以被现有的大多数句法处理器收集,从而减少了theta角色的假设空间。培训师提供的证据可以进一步解决模糊性问题。采用一套基于语言约束的启发式方法来指导歧义消解过程。要获取动词的主位信息,首先必须提取动词的实元结构。采用模板匹配的方法提取动词的参数结构。
{"title":"Interactive acquisition of thematic information of Chinese verbs for judicial verdict document understanding using templates, syntactic clues, and heuristics","authors":"K. H. Lin, Rey-Long Liu, V. Soo","doi":"10.1109/ICDAR.1995.598998","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.598998","url":null,"abstract":"The thematic knowledge can bridge the gap between semantic entities and syntactic constituents. In document understanding, the correctness and the efficiency could be improved if the thematic knowledge is available. In this paper, we propose a semi-automatic method to acquire thematic knowledge of Chinese verbs by exploiting syntactic clues. The syntactic clues, which may be collected by most existing syntactic processors, reduce the hypothesis space of the theta roles. The ambiguities may be further resolved by the evidences from a trainer. A set of heuristics based on linguistic constraints are employed to guide the ambiguity resolution process. To acquire thematic information for verbs, the argument structures of the verbs must be extracted first. A template matching method is used to extract the argument structure of verbs.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114043561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of 3rd International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1