首页 > 最新文献

Proceedings of the Fourth International Conference on Document Analysis and Recognition最新文献

英文 中文
Low quality string recognition for factory automation 用于工厂自动化的低质量字符串识别
K. Sawa, S. Tsuruoka, T. Wakabayashi, F. Kimura, Y. Miyake
Describes a method for dot-printed character string recognition on a piece of steel for factory automation. Our scanned string consists of alphanumerics and the '-' character, and the number of characters is variable from 6 to 12 characters. We propose a new recognition procedure for low-quality strings. The procedure includes image emphasis with a Gaussian Laplacian filter, the extraction of the string subimage, segmentation-recognition with dynamic programming, and fine character recognition. We evaluated its accuracy on a UNIX workstation for 1036 images (8806 characters) scanned by a monochrome video camera in the actual production line at a steel-producing factory, and the average recognition rates were 99.2% for the character recognition and 91.6% for the string recognition.
描述用于工厂自动化的钢材上的点打印字符串识别方法。我们扫描的字符串由字母数字和'-'字符组成,字符数从6到12个字符不等。提出了一种新的低质量字符串识别方法。该程序包括高斯拉普拉斯滤波图像强调、字符串子图像提取、动态规划分割识别和精细字符识别。在UNIX工作站上对某钢厂实际生产线上单色摄像机扫描的1036张图像(8806个字符)进行了准确率评估,字符识别的平均识别率为99.2%,字符串识别的平均识别率为91.6%。
{"title":"Low quality string recognition for factory automation","authors":"K. Sawa, S. Tsuruoka, T. Wakabayashi, F. Kimura, Y. Miyake","doi":"10.1109/ICDAR.1997.620543","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620543","url":null,"abstract":"Describes a method for dot-printed character string recognition on a piece of steel for factory automation. Our scanned string consists of alphanumerics and the '-' character, and the number of characters is variable from 6 to 12 characters. We propose a new recognition procedure for low-quality strings. The procedure includes image emphasis with a Gaussian Laplacian filter, the extraction of the string subimage, segmentation-recognition with dynamic programming, and fine character recognition. We evaluated its accuracy on a UNIX workstation for 1036 images (8806 characters) scanned by a monochrome video camera in the actual production line at a steel-producing factory, and the average recognition rates were 99.2% for the character recognition and 91.6% for the string recognition.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132745175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
More versatile scientific documents 更多样化的科学文献
R. Fateman
The electronic representation of scientific documents (journals, technical reports, program documentation, laboratory notebooks, etc.) presents challenges in several distinct communities. We see five distinct groups who are concerned with electronic versions of scientific documents: (1) publishers of journals, texts and reference works, and their authors; (2) software publishers for OCR/document analysis and document formatting; (3) software publishers whose products access "contents semantics" from documents, including library keyword search programs, natural language search programs, database systems, visual presentation systems, mathematical computation systems, etc.; (4) institutions maintaining access to electronic libraries, which must be broadly construed to include data and programs of all sorts; and (5) individuals and programs acting as their agents who need to use these libraries to identify, locate and retrieve relevant documents. It would be good to have a convergence in design and standards for encoding new or pre-existing (typically paper-based) documents in order to meet the needs of all these groups. Various efforts, some loosely coordinated, but just as often competing, are trying to set standards and build tools. This paper discusses where we are headed.
科学文献(期刊、技术报告、程序文档、实验室笔记等)的电子表示在几个不同的群体中提出了挑战。我们看到有五个不同的群体关注科学文献的电子版本:(1)期刊、文本和参考文献的出版商及其作者;(2)用于OCR/文档分析和文档格式化的软件出版商;(三)其产品从文档中获取“内容语义”的软件发布者,包括图书馆关键字搜索程序、自然语言搜索程序、数据库系统、可视化呈现系统、数学计算系统等;(4)维护电子图书馆访问权限的机构,电子图书馆必须广义地理解为包括各种数据和程序;(5)需要使用这些库来识别、定位和检索相关文档的个人和程序。为了满足所有这些群体的需求,最好在编码新文档或已有文档(通常是基于纸张的)的设计和标准方面有一个统一。各种各样的努力,有些是松散协调的,但也经常是相互竞争的,都在试图设定标准和构建工具。本文讨论了我们的发展方向。
{"title":"More versatile scientific documents","authors":"R. Fateman","doi":"10.1109/ICDAR.1997.620680","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620680","url":null,"abstract":"The electronic representation of scientific documents (journals, technical reports, program documentation, laboratory notebooks, etc.) presents challenges in several distinct communities. We see five distinct groups who are concerned with electronic versions of scientific documents: (1) publishers of journals, texts and reference works, and their authors; (2) software publishers for OCR/document analysis and document formatting; (3) software publishers whose products access \"contents semantics\" from documents, including library keyword search programs, natural language search programs, database systems, visual presentation systems, mathematical computation systems, etc.; (4) institutions maintaining access to electronic libraries, which must be broadly construed to include data and programs of all sorts; and (5) individuals and programs acting as their agents who need to use these libraries to identify, locate and retrieve relevant documents. It would be good to have a convergence in design and standards for encoding new or pre-existing (typically paper-based) documents in order to meet the needs of all these groups. Various efforts, some loosely coordinated, but just as often competing, are trying to set standards and build tools. This paper discusses where we are headed.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133086589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Dynamic word based text compression 动态的基于单词的文本压缩
K. Ng, L. Cheng, C. H. Wong
We propose a dynamic text compression technique with a back searching algorithm and a new storage protocol. Codes being encoded are divided into three types namely copy, literal and hybrid codes. Multiple dictionaries are adopted and each of them has a linked sub-dictionary. Each dictionary has a portion of pre-defined words i.e. the most frequent words and the rest of the entries will depend on the message. A hashing function developed by Pearson (1990) is adopted. It serves two purposes. Firstly, it is used to initialize the dictionary. Secondly, it is used as a quick search to a particular word. By using this scheme, the spaces between words do not need to be considered. At the decoding side, a space character will be appended after each word is decoded. Therefore, the redundancy of space can also be compressed. The result shows that the original message will not be expanded even if we have poor dictionary design.
提出了一种基于反向搜索算法和新的存储协议的动态文本压缩技术。编码的代码分为三种类型,即复制代码、文字代码和混合代码。采用多个字典,每个字典都有一个链接的子字典。每个字典都有一部分预定义的单词,即最常见的单词,其余的条目将取决于消息。采用Pearson(1990)开发的哈希函数。它有两个目的。首先,它用于初始化字典。其次,它被用作对特定单词的快速搜索。通过使用这种方案,不需要考虑单词之间的空格。在解码端,每个字被解码后会附加一个空格字符。因此,空间的冗余也可以被压缩。结果表明,即使我们的字典设计很差,原始信息也不会被扩展。
{"title":"Dynamic word based text compression","authors":"K. Ng, L. Cheng, C. H. Wong","doi":"10.1109/ICDAR.1997.619880","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619880","url":null,"abstract":"We propose a dynamic text compression technique with a back searching algorithm and a new storage protocol. Codes being encoded are divided into three types namely copy, literal and hybrid codes. Multiple dictionaries are adopted and each of them has a linked sub-dictionary. Each dictionary has a portion of pre-defined words i.e. the most frequent words and the rest of the entries will depend on the message. A hashing function developed by Pearson (1990) is adopted. It serves two purposes. Firstly, it is used to initialize the dictionary. Secondly, it is used as a quick search to a particular word. By using this scheme, the spaces between words do not need to be considered. At the decoding side, a space character will be appended after each word is decoded. Therefore, the redundancy of space can also be compressed. The result shows that the original message will not be expanded even if we have poor dictionary design.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130954123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A new method for segmenting handwritten Chinese characters 一种手写汉字分词的新方法
L. Tseng, R. Chen
A new approach is proposed to segment off-line handwritten Chinese characters. Many papers have been published on the off-line recognition of Chinese characters, and almost all of them focus on the recognition of isolated Chinese characters. The segmentation of text into characters was rarely discussed. The segmentation is an important preprocess of the off-line Chinese character recognition because correct recognition of characters relies on correct segmentation of characters. In handwritten Chinese characters, characters may be written to touch each other or to overlap with each other, therefore, the segmentation problem is not an easy one. In this paper, we present a novel method which uses strokes to build stroke bounding boxes first. Then, the knowledge-based merging operations are used to merge those stroke bounding boxes and finally, a dynamic programming method is applied to find the best segmentation boundaries. A series of experiments show that our method is very effective for off-line handwritten Chinese character segmentation.
提出了一种离线手写体汉字分割的新方法。关于汉字的离线识别已经发表了很多论文,但几乎都集中在孤立汉字的识别上。文本分割成字符的问题很少被讨论。字符切分是离线汉字识别的重要预处理,正确的字符识别依赖于正确的字符切分。在手写汉字中,汉字可能会相互接触或相互重叠,因此,分割问题不是一个容易的问题。本文提出了一种利用笔画构建笔画边界盒的新方法。然后,采用基于知识的合并操作对边界框进行合并,最后采用动态规划方法寻找最佳分割边界。一系列实验表明,该方法对离线手写汉字分割是非常有效的。
{"title":"A new method for segmenting handwritten Chinese characters","authors":"L. Tseng, R. Chen","doi":"10.1109/ICDAR.1997.620565","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620565","url":null,"abstract":"A new approach is proposed to segment off-line handwritten Chinese characters. Many papers have been published on the off-line recognition of Chinese characters, and almost all of them focus on the recognition of isolated Chinese characters. The segmentation of text into characters was rarely discussed. The segmentation is an important preprocess of the off-line Chinese character recognition because correct recognition of characters relies on correct segmentation of characters. In handwritten Chinese characters, characters may be written to touch each other or to overlap with each other, therefore, the segmentation problem is not an easy one. In this paper, we present a novel method which uses strokes to build stroke bounding boxes first. Then, the knowledge-based merging operations are used to merge those stroke bounding boxes and finally, a dynamic programming method is applied to find the best segmentation boundaries. A series of experiments show that our method is very effective for off-line handwritten Chinese character segmentation.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134130155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Font recognition and contextual processing for more accurate text recognition 字体识别和上下文处理,更准确的文本识别
Hongwei Shi, T. Pavlidis
Font recognition and contextual processing are developed as two components that enhance the recognition accuracy of a text recognition system presented in a previous paper ((H. Shi and T. Pavlidis, 1996). Font information is extracted from two sources: one is the global page properties, and the other is the graph matching result of recognized short words such as a, it and of etc. Contextual processing is done by first composing word candidates from the recognition results and then checking each candidate with a dictionary through a spelling checker. Positional binary trigrams and word affixes are used to prune the search for word candidates.
在之前的一篇论文中,字体识别和上下文处理作为提高文本识别系统识别精度的两个组成部分被开发出来((H. Shi和T. Pavlidis, 1996)。字体信息的提取有两个来源:一个是全局页面属性,另一个是识别出的a、it、of等短词的图匹配结果。上下文处理是这样完成的:首先从识别结果中组合候选词,然后通过拼写检查器用字典检查每个候选词。使用位置二进制三元组和词缀来减少候选词的搜索。
{"title":"Font recognition and contextual processing for more accurate text recognition","authors":"Hongwei Shi, T. Pavlidis","doi":"10.1109/ICDAR.1997.619810","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619810","url":null,"abstract":"Font recognition and contextual processing are developed as two components that enhance the recognition accuracy of a text recognition system presented in a previous paper ((H. Shi and T. Pavlidis, 1996). Font information is extracted from two sources: one is the global page properties, and the other is the graph matching result of recognized short words such as a, it and of etc. Contextual processing is done by first composing word candidates from the recognition results and then checking each candidate with a dictionary through a spelling checker. Positional binary trigrams and word affixes are used to prune the search for word candidates.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133650078","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Rectangle labelling for an invoice understanding system 发票理解系统的矩形标签
F. Cesarini, E. Francesconi, M. Gori, S. Marinai, Jianqing Sheng, G. Soda
We present a method for the logical labelling of physical rectangles, extracted from invoices, based on a conceptual model which describes, as generally as possible, the invoice universe. This general knowledge is used in the semi automatic construction of a model for each class of invoices. Once the model is constructed, it can be applied to understand an invoice instance, whose class is univocally identified by its logo. This approach is used to design a flexible system which is able to learn, from a nucleus of general knowledge, a monotonic set of specific knowledge for each class of invoices (document models), in terms of physical coordinates for each rectangle and related semantic label.
我们提出了一种从发票中提取的物理矩形的逻辑标记方法,该方法基于一个概念模型,该模型尽可能地描述发票的范围。这种一般知识用于为每一类发票半自动构建模型。一旦构建了模型,就可以应用它来理解发票实例,发票实例的类由其徽标唯一标识。这种方法被用来设计一个灵活的系统,该系统能够从一般知识的核心中学习每一类发票(文档模型)的单调的特定知识集,根据每个矩形的物理坐标和相关的语义标签。
{"title":"Rectangle labelling for an invoice understanding system","authors":"F. Cesarini, E. Francesconi, M. Gori, S. Marinai, Jianqing Sheng, G. Soda","doi":"10.1109/ICDAR.1997.619865","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619865","url":null,"abstract":"We present a method for the logical labelling of physical rectangles, extracted from invoices, based on a conceptual model which describes, as generally as possible, the invoice universe. This general knowledge is used in the semi automatic construction of a model for each class of invoices. Once the model is constructed, it can be applied to understand an invoice instance, whose class is univocally identified by its logo. This approach is used to design a flexible system which is able to learn, from a nucleus of general knowledge, a monotonic set of specific knowledge for each class of invoices (document models), in terms of physical coordinates for each rectangle and related semantic label.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131998408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
An efficient fully parallel thinning algorithm 一种高效的全并行细化算法
N.H. Han, C. La, P. Rhee
The paper addresses an efficient parallel thinning algorithm based on weight-values. The weight-value of a black pixel is calculated by observing neighboring pixels, and it gives one an efficient way to decide whether the pixel is deleted or not. Owing to weight-values, the proposed algorithm uses only 3/spl times/3 templates. Furthermore, it examines only the elimination conditions corresponding the weight-value of boundary pixels, and all elimination conditions will not be searched as most other parallel iterative thinning algorithms. Thus, the execution time can be reduced a lot compared to that of previous approaches. The weight-value also allow one to deal with typical troublesome patterns efficiently. Without smoothing before thinning, the algorithm produces robust thinned images even in the presence of two pixel-width noises. The authors obtain encouraging results from extensive experiments.
提出了一种基于权值的并行细化算法。通过观察相邻像素来计算黑色像素的权重值,为判断该像素是否被删除提供了一种有效的方法。由于权重值的存在,该算法只使用了3/spl次/3个模板。此外,它只检查边界像素权值对应的消除条件,而不像大多数其他并行迭代细化算法那样搜索所有消除条件。因此,与以前的方法相比,执行时间可以减少很多。权重值还允许有效地处理典型的麻烦模式。该算法无需在细化前进行平滑处理,即使存在两个像素宽度的噪声,也能产生鲁棒的细化图像。作者从大量的实验中获得了令人鼓舞的结果。
{"title":"An efficient fully parallel thinning algorithm","authors":"N.H. Han, C. La, P. Rhee","doi":"10.1109/ICDAR.1997.619829","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619829","url":null,"abstract":"The paper addresses an efficient parallel thinning algorithm based on weight-values. The weight-value of a black pixel is calculated by observing neighboring pixels, and it gives one an efficient way to decide whether the pixel is deleted or not. Owing to weight-values, the proposed algorithm uses only 3/spl times/3 templates. Furthermore, it examines only the elimination conditions corresponding the weight-value of boundary pixels, and all elimination conditions will not be searched as most other parallel iterative thinning algorithms. Thus, the execution time can be reduced a lot compared to that of previous approaches. The weight-value also allow one to deal with typical troublesome patterns efficiently. Without smoothing before thinning, the algorithm produces robust thinned images even in the presence of two pixel-width noises. The authors obtain encouraging results from extensive experiments.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132223944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
The performance evaluation of thresholding algorithms for optical character recognition 光学字符识别的阈值算法性能评价
A. T. Abak, U. Baris, B. Sankur
The paper presents performance evaluation of thresholding algorithms in the context of document analysis and character recognition systems. Several thresholding algorithms are comparatively evaluated on the basis of the original bitmaps of characters. Different distance measures such as Hausdorff, Jaccard, and Yule are used to measure the similarity between thresholded bitmaps and original bitmaps of characters.
本文介绍了阈值算法在文档分析和字符识别系统中的性能评价。在原始字符位图的基础上,比较评价了几种阈值分割算法。使用不同的距离度量,如Hausdorff、Jaccard和Yule来度量阈值位图与字符原始位图之间的相似度。
{"title":"The performance evaluation of thresholding algorithms for optical character recognition","authors":"A. T. Abak, U. Baris, B. Sankur","doi":"10.1109/ICDAR.1997.620597","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620597","url":null,"abstract":"The paper presents performance evaluation of thresholding algorithms in the context of document analysis and character recognition systems. Several thresholding algorithms are comparatively evaluated on the basis of the original bitmaps of characters. Different distance measures such as Hausdorff, Jaccard, and Yule are used to measure the similarity between thresholded bitmaps and original bitmaps of characters.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"694 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132790979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Optical character recognition without segmentation 无分割的光学字符识别
M.A. Ozdil, F. Vural
A segmentation-free approach for off-line optical character recognition is presented. The proposed method performs the recognition by extracting the characters from the whole word, avoiding the segmentation process. A control point set which includes position and attribute vectors is selected for the features. In the training mode, each sample character is mapped to a set of control points and is stored in an archive which belongs to an alphabet. In the recognition mode, the control points of the input image are first extracted. Then, each control point is matched to the control points in the alphabet according to its attributes. During the matching process, a probability matrix is constructed which holds some matching measures (probabilities) for identifying the characters. Experimental results indicate that the proposed method is very robust in extracting the characters from a cursive script.
提出了一种无分割的离线光学字符识别方法。该方法通过从整个单词中提取字符来实现识别,避免了分割过程。为特征选择包含位置向量和属性向量的控制点集。在训练模式下,每个样本字符被映射到一组控制点,并存储在一个属于字母表的存档中。在识别模式中,首先提取输入图像的控制点。然后,将每个控制点根据其属性与字母表中的控制点进行匹配。在匹配过程中,构造一个概率矩阵,其中包含一些用于识别字符的匹配度量(概率)。实验结果表明,该方法对草书字符的提取具有较好的鲁棒性。
{"title":"Optical character recognition without segmentation","authors":"M.A. Ozdil, F. Vural","doi":"10.1109/ICDAR.1997.620545","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620545","url":null,"abstract":"A segmentation-free approach for off-line optical character recognition is presented. The proposed method performs the recognition by extracting the characters from the whole word, avoiding the segmentation process. A control point set which includes position and attribute vectors is selected for the features. In the training mode, each sample character is mapped to a set of control points and is stored in an archive which belongs to an alphabet. In the recognition mode, the control points of the input image are first extracted. Then, each control point is matched to the control points in the alphabet according to its attributes. During the matching process, a probability matrix is constructed which holds some matching measures (probabilities) for identifying the characters. Experimental results indicate that the proposed method is very robust in extracting the characters from a cursive script.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130026715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A method for connecting disappeared junction patterns on frame lines in form documents 一种连接表单文档中框架线上消失的连接图案的方法
Hiroshi Shinjo, Kazuki Nakashima, Masashi Koga, K. Marukawa, Y. Shima, Eiichi Hadano
Form document structure analysis is an essential technique for recognizing the positions of characters in general forms. However, it has a fundamental problem that interruptions of lines, as well as noise, lead to incorrect analysis. The paper focuses on a method for connecting junction patterns in which portions of the horizontal and vertical lines are not visible, referred to as "disappeared junction patterns". Our method has two key stages for making correct connections. The first is noise elimination, in which lines whose two end points meet no other lines and which are shorter than the minimum line length parameter, are eliminated. The second is object line selection, where only frame lines of tables are selected as object lines for connection. Experiments with 39 form images demonstrated the feasibility of this method.
表单文档结构分析是识别一般格式中字符位置的一项重要技术。然而,它有一个根本的问题,即线路中断,以及噪声,导致不正确的分析。本文的重点是一种方法,以连接的连接模式,其中部分的水平和垂直线是不可见的,称为“消失的连接模式”。我们的方法有两个关键阶段来建立正确的连接。首先是消噪,即消去两端不相交且小于最小线长参数的线。第二种是对象线选择,只选择表的框架线作为连接的对象线。39张表单图像的实验验证了该方法的可行性。
{"title":"A method for connecting disappeared junction patterns on frame lines in form documents","authors":"Hiroshi Shinjo, Kazuki Nakashima, Masashi Koga, K. Marukawa, Y. Shima, Eiichi Hadano","doi":"10.1109/ICDAR.1997.620590","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620590","url":null,"abstract":"Form document structure analysis is an essential technique for recognizing the positions of characters in general forms. However, it has a fundamental problem that interruptions of lines, as well as noise, lead to incorrect analysis. The paper focuses on a method for connecting junction patterns in which portions of the horizontal and vertical lines are not visible, referred to as \"disappeared junction patterns\". Our method has two key stages for making correct connections. The first is noise elimination, in which lines whose two end points meet no other lines and which are shorter than the minimum line length parameter, are eliminated. The second is object line selection, where only frame lines of tables are selected as object lines for connection. Experiments with 39 form images demonstrated the feasibility of this method.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130250597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
Proceedings of the Fourth International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1