首页 > 最新文献

Proceedings of 3rd International Conference on Document Analysis and Recognition最新文献

英文 中文
A language model based on semantically clustered words in a Chinese character recognition system 基于语义聚类词的汉字识别语言模型
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599033
Hsi-Jian Lee, Cheng-Huang Tung
This paper presents a new method for clustering the words in a dictionary into word groups, which are applied in a Chinese character recognition system with a language model to describe the contextual information. The Chinese synonym dictionary Tong2yi4ci2 ci2lin2 providing the semantic features is used to train the weights of the semantic attributes of the character-based word classes. The weights of the semantic attributes are next updated according to the words of the behavior dictionary, which has a rather complete word set. Then, the updated word classes are clustered into m groups according to the semantic measurement by a greedy method. The words in the behavior dictionary can finally be assigned into the m groups. The parameter space for bigram contextual information of the character recognition system is m/sup 2/. From the experimental results, the recognition system with the proposed model has shown better performance than that of a character-based bigram language model.
本文提出了一种将词典中的词聚类成词组的新方法,并将其应用于基于语言模型描述上下文信息的汉字识别系统中。使用提供语义特征的汉语同义词词典Tong2yi4ci2 ci2lin2来训练基于字符的词类的语义属性权值。然后根据行为字典中的单词更新语义属性的权重,该字典具有相当完整的单词集。然后,根据语义度量,采用贪心方法将更新后的词类聚为m组。行为字典中的单词最终可以分配到m组中。字符识别系统的双字母上下文信息的参数空间为m/sup 2/。实验结果表明,基于该模型的识别系统比基于字符的双字语言模型具有更好的识别性能。
{"title":"A language model based on semantically clustered words in a Chinese character recognition system","authors":"Hsi-Jian Lee, Cheng-Huang Tung","doi":"10.1109/ICDAR.1995.599033","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599033","url":null,"abstract":"This paper presents a new method for clustering the words in a dictionary into word groups, which are applied in a Chinese character recognition system with a language model to describe the contextual information. The Chinese synonym dictionary Tong2yi4ci2 ci2lin2 providing the semantic features is used to train the weights of the semantic attributes of the character-based word classes. The weights of the semantic attributes are next updated according to the words of the behavior dictionary, which has a rather complete word set. Then, the updated word classes are clustered into m groups according to the semantic measurement by a greedy method. The words in the behavior dictionary can finally be assigned into the m groups. The parameter space for bigram contextual information of the character recognition system is m/sup 2/. From the experimental results, the recognition system with the proposed model has shown better performance than that of a character-based bigram language model.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"172 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132794036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Gray scale filtering for line and word segmentation 灰度滤波的线和词分割
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.601979
Yi Lu, A. Tisler
The extraction of lines, words and characters from a digital document image are necessary computational steps preceding character recognition. Much has been discussed in character segmentation and recognition but little has been done in the area of line and word segmentation. The authors present two special filters, minimum difference filters (MDF) and average difference filters (ADF) to facilitate line and word segmentation. They discuss how to select the scales of these filters dynamically and how to use the filters to eliminate crossing lines from a text image.
从数字文档图像中提取行、词和字符是字符识别的必要计算步骤。在字符分割和识别方面已经讨论了很多,但在线分和词分方面却做得很少。作者提出了两种特殊的滤波器,最小差分滤波器(MDF)和平均差分滤波器(ADF),以方便线和词的分割。他们讨论了如何动态地选择这些过滤器的尺度,以及如何使用过滤器从文本图像中消除交叉线。
{"title":"Gray scale filtering for line and word segmentation","authors":"Yi Lu, A. Tisler","doi":"10.1109/ICDAR.1995.601979","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.601979","url":null,"abstract":"The extraction of lines, words and characters from a digital document image are necessary computational steps preceding character recognition. Much has been discussed in character segmentation and recognition but little has been done in the area of line and word segmentation. The authors present two special filters, minimum difference filters (MDF) and average difference filters (ADF) to facilitate line and word segmentation. They discuss how to select the scales of these filters dynamically and how to use the filters to eliminate crossing lines from a text image.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133641941","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Segmentation of numeric strings 数字字符串的分割
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602080
G. Congedo, G. Dimauro, S. Impedovo, G. Pirlo
This paper presents a complete procedure for the segmentation of handwritten numeric strings. The procedure uses an hypothesis-then-verification strategy in which multiple segmentation algorithms based on contiguous row partition work sequentially on the binary image until an acceptable segmentation is obtained. At this purpose a new set of algorithms simulating a "drop falling" process is introduced. The experimental tests demonstrate the effectiveness of the new algorithms in obtaining high-confidence segmentation hypotheses.
本文给出了一个完整的手写数字字符串分割程序。该方法采用假设-验证策略,其中基于连续行分割的多个分割算法依次对二值图像进行分割,直到获得可接受的分割。为此,本文引入了一套新的模拟“滴落”过程的算法。实验验证了新算法在获得高置信度分割假设方面的有效性。
{"title":"Segmentation of numeric strings","authors":"G. Congedo, G. Dimauro, S. Impedovo, G. Pirlo","doi":"10.1109/ICDAR.1995.602080","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602080","url":null,"abstract":"This paper presents a complete procedure for the segmentation of handwritten numeric strings. The procedure uses an hypothesis-then-verification strategy in which multiple segmentation algorithms based on contiguous row partition work sequentially on the binary image until an acceptable segmentation is obtained. At this purpose a new set of algorithms simulating a \"drop falling\" process is introduced. The experimental tests demonstrate the effectiveness of the new algorithms in obtaining high-confidence segmentation hypotheses.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117101132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
Mathematics recognition using graph rewriting 使用图形重写的数学识别
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599026
Ann Grbavec, D. Blostein
This paper investigates graph rewriting as a tool for high-level recognition of two-dimensional mathematical notation. "High-level recognition" is the process of determining the meaning of a diagram from the output of a symbol recognizer. Characteristic problems of high-level mathematics recognition include: determining the groupings of symbols into recursive subexpressions and resolving ambiguities that depend upon global context. Our graph-rewriting approach uses knowledge of the notational conventions of mathematics, such as operator precedence and operator range, more effectively than syntactic or previous structural methods. Graph rewriting offers a flexible formalism with a strong theoretical foundation for manipulating two-dimensional patterns. It has been shown to be a useful technique for high-level recognition of circuit diagrams and musical scores. By demonstrating a graph-rewriting strategy for mathematics recognition, this paper provides further evidence for graph rewriting as a general tool for diagram recognition, and identifies some of the issues that must be considered as this potential is explored.
本文研究了图形重写作为二维数学符号高级识别的一种工具。“高级识别”是根据符号识别器的输出确定图表含义的过程。高级数学识别的特征问题包括:确定符号的递归子表达式的分组和解决依赖于全局上下文的歧义。我们的图重写方法使用了数学符号约定的知识,例如运算符优先级和运算符范围,比语法方法或以前的结构方法更有效。图形重写为操作二维模式提供了灵活的形式和强大的理论基础。它已被证明是一种有用的技术,用于高级识别电路图和乐谱。通过展示用于数学识别的图形重写策略,本文为图形重写作为图形识别的通用工具提供了进一步的证据,并确定了在探索这种潜力时必须考虑的一些问题。
{"title":"Mathematics recognition using graph rewriting","authors":"Ann Grbavec, D. Blostein","doi":"10.1109/ICDAR.1995.599026","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599026","url":null,"abstract":"This paper investigates graph rewriting as a tool for high-level recognition of two-dimensional mathematical notation. \"High-level recognition\" is the process of determining the meaning of a diagram from the output of a symbol recognizer. Characteristic problems of high-level mathematics recognition include: determining the groupings of symbols into recursive subexpressions and resolving ambiguities that depend upon global context. Our graph-rewriting approach uses knowledge of the notational conventions of mathematics, such as operator precedence and operator range, more effectively than syntactic or previous structural methods. Graph rewriting offers a flexible formalism with a strong theoretical foundation for manipulating two-dimensional patterns. It has been shown to be a useful technique for high-level recognition of circuit diagrams and musical scores. By demonstrating a graph-rewriting strategy for mathematics recognition, this paper provides further evidence for graph rewriting as a general tool for diagram recognition, and identifies some of the issues that must be considered as this potential is explored.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116510208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 82
Faxed image restoration using Kalman filtering 基于卡尔曼滤波的传真图像恢复
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.601987
Myoung-Young Yoon, Seong-Whan Lee, Ju-Sung Kim
We present a new scheme for the restoration of faxed images degraded by both salient noise and additive white noise. We consider two fundamental aspects of faxed image restoration: modeling and the restoration algorithm. First, a model of a faxed image is presented. The model is based on an autoregressive Gauss-Markov random field and includes vertical and horizontal overlap effects. In particular, we concentrate on the nonsymmetric half plane causality. Second, the restoration of faxed images degraded by salient noise and additive white noise is considered by 2D Kalman filtering which provides an efficient recursive procedure. In order to illustrate the effectiveness of the proposed scheme, we present experimental results on the restoration of a faxed image.
我们提出了一种新的方案来恢复被显著噪声和加性白噪声退化的传真图像。我们考虑了传真图像恢复的两个基本方面:建模和恢复算法。首先,提出了一种传真图像的模型。该模型基于自回归高斯-马尔可夫随机场,包括垂直和水平重叠效应。特别地,我们集中于非对称半平面的因果关系。其次,采用二维卡尔曼滤波方法对显著噪声和加性白噪声退化的传真图像进行恢复,提供了一种有效的递归算法。为了说明所提方案的有效性,我们给出了对传真图像恢复的实验结果。
{"title":"Faxed image restoration using Kalman filtering","authors":"Myoung-Young Yoon, Seong-Whan Lee, Ju-Sung Kim","doi":"10.1109/ICDAR.1995.601987","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.601987","url":null,"abstract":"We present a new scheme for the restoration of faxed images degraded by both salient noise and additive white noise. We consider two fundamental aspects of faxed image restoration: modeling and the restoration algorithm. First, a model of a faxed image is presented. The model is based on an autoregressive Gauss-Markov random field and includes vertical and horizontal overlap effects. In particular, we concentrate on the nonsymmetric half plane causality. Second, the restoration of faxed images degraded by salient noise and additive white noise is considered by 2D Kalman filtering which provides an efficient recursive procedure. In order to illustrate the effectiveness of the proposed scheme, we present experimental results on the restoration of a faxed image.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115675755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
A microprocessor-based optical character recognition check reader 一种基于微处理器的光学字符识别校验器
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602066
Francis Y. L. Chin, Francis Wu
Magnetic Ink Character Recognition (MICR) technology has widely been used for processing bank checks. Since the MICR character set is a special type font, and the ink is also readable by human being, optical approach can also be used. This report will describe the design of a low-cost, but highly accurate, microprocessor-based optical character recognition (OCR) check reader. The performance of our OCR reader is affected by a number of factors, mainly the noise generated by the lens system and the colour image at the check background. In this paper we describe how our software solution can alleviate these problems. As speed is another concern, special attention is paid to the design of recognition algorithm, such as the avoidance of floating point arithmetics, hardware limitations, etc.
磁墨字符识别(MICR)技术已广泛应用于银行支票处理。由于MICR字符集是一种特殊的类型字体,而且墨水也是人类可读的,因此也可以采用光学方法。本报告将描述一种低成本,但高精度,基于微处理器的光学字符识别(OCR)检查阅读器的设计。我们的OCR阅读器的性能受到许多因素的影响,主要是镜头系统产生的噪声和检查背景下的彩色图像。在本文中,我们描述了我们的软件解决方案如何缓解这些问题。由于速度是另一个需要考虑的问题,因此在识别算法的设计上需要特别注意,如避免浮点运算、硬件限制等。
{"title":"A microprocessor-based optical character recognition check reader","authors":"Francis Y. L. Chin, Francis Wu","doi":"10.1109/ICDAR.1995.602066","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602066","url":null,"abstract":"Magnetic Ink Character Recognition (MICR) technology has widely been used for processing bank checks. Since the MICR character set is a special type font, and the ink is also readable by human being, optical approach can also be used. This report will describe the design of a low-cost, but highly accurate, microprocessor-based optical character recognition (OCR) check reader. The performance of our OCR reader is affected by a number of factors, mainly the noise generated by the lens system and the colour image at the check background. In this paper we describe how our software solution can alleviate these problems. As speed is another concern, special attention is paid to the design of recognition algorithm, such as the avoidance of floating point arithmetics, hardware limitations, etc.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121754232","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Extracting characters and character lines in multi-agent scheme 多智能体方案中的字符和字符行提取
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.599000
K. Gyohten, Tomoko Sumiya, N. Babaguchi, K. Kakusho, T. Kitahashi
In this paper, we present COCE (COordinative Character Extractor), a new method for extracting printed Japanese characters from an unformatted document image. This research aims to exploit knowledge independent of the layouts. COCE is based on a multiagent scheme where each agent is assigned to a single character line and tries to extract characters by making use of the knowledge about features of a character line as well as shapes and arrangement of characters. Moreover, the agents communicate with each other to keep consistency between their tasks. We have favourable results for the effectiveness of this method.
本文提出了一种从未格式化文档图像中提取打印日文字符的新方法COCE (coordinated Character Extractor)。本研究旨在开发与布局无关的知识。COCE基于多智能体方案,每个智能体被分配到单个字符线,并试图利用有关字符线的特征以及字符的形状和排列的知识来提取字符。此外,代理之间相互通信以保持其任务之间的一致性。我们对这种方法的有效性有良好的结果。
{"title":"Extracting characters and character lines in multi-agent scheme","authors":"K. Gyohten, Tomoko Sumiya, N. Babaguchi, K. Kakusho, T. Kitahashi","doi":"10.1109/ICDAR.1995.599000","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.599000","url":null,"abstract":"In this paper, we present COCE (COordinative Character Extractor), a new method for extracting printed Japanese characters from an unformatted document image. This research aims to exploit knowledge independent of the layouts. COCE is based on a multiagent scheme where each agent is assigned to a single character line and tries to extract characters by making use of the knowledge about features of a character line as well as shapes and arrangement of characters. Moreover, the agents communicate with each other to keep consistency between their tasks. We have favourable results for the effectiveness of this method.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123198255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
On-line cursive script recognition using an island-driven search technique 使用岛驱动搜索技术的在线草书识别
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602043
Seung-Ho Lee, Hyunkyu Lee, J. H. Kim
A new approach for on-line cursive script recognition that combines a letter spotting technique with an island-driven lattice search algorithm is presented. Initially, all plausible letter components within an input pattern are detected, using a letter spotting technique based on hidden Markov models. A word hypothesis lattice is generated as a result of the letter spotting. Then an island-driven search algorithm is performed to find the optimal path on the word hypothesis lattice, which corresponds to the most probable word among the dictionary words. The results of this experiment suggest that the proposed method works effectively in recognizing English cursive words. In a word recognition test, the average 85.4% word accuracy was obtained.
提出了一种结合字母识别技术和岛驱动格搜索算法的在线草书识别新方法。最初,使用基于隐马尔可夫模型的字母识别技术,检测输入模式中所有可能的字母组件。单词假设格是由字母点阵生成的。然后使用岛驱动搜索算法在单词假设格上寻找最优路径,该路径对应于字典单词中最可能的单词。实验结果表明,该方法能够有效地识别英文草书单词。在单词识别测试中,平均准确率达到85.4%。
{"title":"On-line cursive script recognition using an island-driven search technique","authors":"Seung-Ho Lee, Hyunkyu Lee, J. H. Kim","doi":"10.1109/ICDAR.1995.602043","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602043","url":null,"abstract":"A new approach for on-line cursive script recognition that combines a letter spotting technique with an island-driven lattice search algorithm is presented. Initially, all plausible letter components within an input pattern are detected, using a letter spotting technique based on hidden Markov models. A word hypothesis lattice is generated as a result of the letter spotting. Then an island-driven search algorithm is performed to find the optimal path on the word hypothesis lattice, which corresponds to the most probable word among the dictionary words. The results of this experiment suggest that the proposed method works effectively in recognizing English cursive words. In a word recognition test, the average 85.4% word accuracy was obtained.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124377857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Improved binarization algorithm for document image by histogram and edge detection 基于直方图和边缘检测的改进文档图像二值化算法
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.601976
Moon-Soo Chang, S. Kang, Woo-Sik Rho, Heok-Gu Kim, Duck-Jin Kim
A binarization method is presented to counter the stroke connectivity problems of characters arising from mid-level-quality binary image scanning systems. In the output of a binary image scanning system, separate strokes may look connected if the point size is small and the character strokes are complex while strokes may lose connectivity if they are generated at low intensity. Also, erroneous recognition may result if a blemished document surface distorts the image. To counter these problems and to further enhance the quality of character recognition, the authors have developed an integrated binarization scheme, exploiting synergistic use of an adaptive thresholding technique and variable histogram equalization. This algorithm is composed of two components. The first removes background noise via gray level histogram equalization while the second enhances the gray level of characters over and above the surrounding background via an edge image composition technique.
针对中等质量二值图像扫描系统中出现的字符笔画连通性问题,提出了一种二值化方法。在二值图像扫描系统的输出中,如果点尺寸小,字符笔画复杂,则单独的笔画看起来可能是相连的,而如果笔画在低强度下生成,则可能失去连通性。此外,如果有瑕疵的文档表面扭曲了图像,则可能导致错误识别。为了解决这些问题并进一步提高字符识别的质量,作者开发了一个集成的二值化方案,利用自适应阈值技术和变量直方图均衡的协同使用。该算法由两个部分组成。第一种方法是通过灰度直方图均衡化去除背景噪声,第二种方法是通过边缘图像合成技术增强字符在周围背景之上的灰度。
{"title":"Improved binarization algorithm for document image by histogram and edge detection","authors":"Moon-Soo Chang, S. Kang, Woo-Sik Rho, Heok-Gu Kim, Duck-Jin Kim","doi":"10.1109/ICDAR.1995.601976","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.601976","url":null,"abstract":"A binarization method is presented to counter the stroke connectivity problems of characters arising from mid-level-quality binary image scanning systems. In the output of a binary image scanning system, separate strokes may look connected if the point size is small and the character strokes are complex while strokes may lose connectivity if they are generated at low intensity. Also, erroneous recognition may result if a blemished document surface distorts the image. To counter these problems and to further enhance the quality of character recognition, the authors have developed an integrated binarization scheme, exploiting synergistic use of an adaptive thresholding technique and variable histogram equalization. This algorithm is composed of two components. The first removes background noise via gray level histogram equalization while the second enhances the gray level of characters over and above the surrounding background via an edge image composition technique.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128540205","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
Automatic text skew estimation in document images 文档图像中的自动文本倾斜估计
Pub Date : 1995-08-14 DOI: 10.1109/ICDAR.1995.602126
Su S. Chen, R. Haralick, I. T. Phillips
This paper describes an algorithm to estimate the text skew angle in a document image. The algorithm utilizes the recursive morphological transforms and yields accurate estimates of text skew angles on a large document image data set. The algorithm computes the optimal parameter settings on the fly without any human interaction. In this automatic mode, experimental results indicate that the algorithm generates estimated text skew angles within 0.5/spl deg/ of the true text skew angles with a probability of 99%. To process a 300 dpi document image, the algorithm takes 10 seconds on SUN Sparc 10 machines.
本文描述了一种估计文档图像中文本倾斜角度的算法。该算法利用递归形态学变换,在大型文档图像数据集上产生准确的文本倾斜角度估计。该算法在没有任何人工干预的情况下动态计算最佳参数设置。在这种自动模式下,实验结果表明,算法生成的估计文本偏斜角度在真实文本偏斜角度的0.5/spl度/范围内,概率为99%。要处理一张300 dpi的文档图像,该算法在SUN Sparc 10机器上需要10秒。
{"title":"Automatic text skew estimation in document images","authors":"Su S. Chen, R. Haralick, I. T. Phillips","doi":"10.1109/ICDAR.1995.602126","DOIUrl":"https://doi.org/10.1109/ICDAR.1995.602126","url":null,"abstract":"This paper describes an algorithm to estimate the text skew angle in a document image. The algorithm utilizes the recursive morphological transforms and yields accurate estimates of text skew angles on a large document image data set. The algorithm computes the optimal parameter settings on the fly without any human interaction. In this automatic mode, experimental results indicate that the algorithm generates estimated text skew angles within 0.5/spl deg/ of the true text skew angles with a probability of 99%. To process a 300 dpi document image, the algorithm takes 10 seconds on SUN Sparc 10 machines.","PeriodicalId":273519,"journal":{"name":"Proceedings of 3rd International Conference on Document Analysis and Recognition","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128912150","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
期刊
Proceedings of 3rd International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1