首页 > 最新文献

Proceedings of the Fourth International Conference on Document Analysis and Recognition最新文献

英文 中文
Recognising letters in on-line handwriting using hierarchical fuzzy inference 基于层次模糊推理的在线手写字母识别
A. Hennig, N. Sherkat, R. Whitrow
The recognition of unconstrained handwriting has to cope with the ambiguity and variability of cursive script. Preprocessing techniques are often applied to on-line data before representing the script as basic primitives, resulting in the propagation of errors introduced during pre-processing. This paper therefore combines pre-processing of the data (i.e. tangential smoothing) and encoding into primitives (Partial Strokes) in a single step. Finding the correct character at the correct place (i.e. letter spotting) is the main problem in non-holistic recognition approaches. Many cursive letters are composed of common shapes of varying complexity that can in turn consist of other subshapes. In this paper, we present a production rule system using Hierarchical Fuzzy Inference in order to exploit this hierarchical property of cursive script. Shapes of increasing complexity are found on a page of handwriting until letters are finally spotted. Zoning is then applied to verify their vertical position. The performance of letter spotting is compared with an alternative method.
无约束笔迹的识别必须处理草书的歧义性和可变性。在将脚本表示为基本原语之前,通常对在线数据应用预处理技术,导致预处理过程中引入的错误传播。因此,本文将数据预处理(即切向平滑)和编码成原语(部分笔画)在一个步骤中结合起来。在正确的位置找到正确的字符(即字母定位)是非整体识别方法的主要问题。许多草书字母由不同复杂程度的普通形状组成,这些形状又可以由其他子形状组成。为了充分利用草书的这种层次性,本文提出了一种基于层次模糊推理的生成规则系统。在一页手写体上发现越来越复杂的形状,直到最后发现字母。然后应用分区来验证它们的垂直位置。将字母识别的性能与另一种方法进行了比较。
{"title":"Recognising letters in on-line handwriting using hierarchical fuzzy inference","authors":"A. Hennig, N. Sherkat, R. Whitrow","doi":"10.1109/ICDAR.1997.620648","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620648","url":null,"abstract":"The recognition of unconstrained handwriting has to cope with the ambiguity and variability of cursive script. Preprocessing techniques are often applied to on-line data before representing the script as basic primitives, resulting in the propagation of errors introduced during pre-processing. This paper therefore combines pre-processing of the data (i.e. tangential smoothing) and encoding into primitives (Partial Strokes) in a single step. Finding the correct character at the correct place (i.e. letter spotting) is the main problem in non-holistic recognition approaches. Many cursive letters are composed of common shapes of varying complexity that can in turn consist of other subshapes. In this paper, we present a production rule system using Hierarchical Fuzzy Inference in order to exploit this hierarchical property of cursive script. Shapes of increasing complexity are found on a page of handwriting until letters are finally spotted. Zoning is then applied to verify their vertical position. The performance of letter spotting is compared with an alternative method.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123487547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Finding straight lines in drawings 在图纸中寻找直线
Juan F. Arias, A. K. Chhabra, Vishal Misra
We have developed an efficient method to extract straight lines at any orientation from a line drawing. The method works by extracting the horizontal and vertical lines using the FAST method, detecting the angles of the other lines and applying the FAST method again while the image is rotated to each corresponding angle. The method is efficient because it is based on very efficient line finding, transposition, and rotation operations which work over the run-length representation of the line drawing.
我们已经开发了一种有效的方法,可以从线条图中提取任意方向的直线。该方法的工作原理是使用FAST方法提取水平线和垂直线,检测其他线的角度,并在图像旋转到每个相应角度时再次应用FAST方法。这种方法是有效的,因为它基于非常有效的寻线、换位和旋转操作,这些操作在线条绘制的运行长度表示上工作。
{"title":"Finding straight lines in drawings","authors":"Juan F. Arias, A. K. Chhabra, Vishal Misra","doi":"10.1109/ICDAR.1997.620618","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620618","url":null,"abstract":"We have developed an efficient method to extract straight lines at any orientation from a line drawing. The method works by extracting the horizontal and vertical lines using the FAST method, detecting the angles of the other lines and applying the FAST method again while the image is rotated to each corresponding angle. The method is efficient because it is based on very efficient line finding, transposition, and rotation operations which work over the run-length representation of the line drawing.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122035375","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Logo and word matching using a general approach to signal registration 采用通用的标志与文字匹配方法进行信号注册
P. Suda, C. Bridoux, B. Kämmerer, G. Maderlechner
The paper presents work in the field of logo and word recognition. The approach is based on a general theory for signal registration and is thus applicable to a broad variety of signal processing domains. It has been fruitfully applied to solve speech and handwriting recognition as well as tasks in the field of document analysis.
本文介绍了在标识和文字识别领域的工作。该方法基于信号配准的一般理论,因此适用于各种信号处理领域。它已成功地应用于解决语音和手写识别以及文档分析领域的任务。
{"title":"Logo and word matching using a general approach to signal registration","authors":"P. Suda, C. Bridoux, B. Kämmerer, G. Maderlechner","doi":"10.1109/ICDAR.1997.619814","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619814","url":null,"abstract":"The paper presents work in the field of logo and word recognition. The approach is based on a general theory for signal registration and is thus applicable to a broad variety of signal processing domains. It has been fruitfully applied to solve speech and handwriting recognition as well as tasks in the field of document analysis.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124448982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
The function of documents 文件的功能
D. Doermann, E. Rivlin, A. Rosenfeld
The purpose of a document is to facilitate the transfer of information from its author to its readers. It is the author's job to design the document so that the information it contains can be interpreted accurately and efficiently. To do this, the author can make use of a set of stylistic tools. In this paper, we introduce the concept of document functionality, which attempts to describe the roles of documents and their components in the process of transferring information. A functional description of a document provides insight into the type of the document, into its intended uses, and into strategies for automatic document interpretation and retrieval. To demonstrate these ideas, we define a taxonomy of functional document components and show how functional descriptions can be used to reverse-engineer the intentions of the author, to navigate in document space, and to provide important contextual information to aid in interpretation.
文档的目的是方便信息从作者传递给读者。作者的工作是设计文档,使其包含的信息能够准确有效地解释。为了做到这一点,作者可以使用一套风格工具。本文引入了文档功能的概念,试图描述文档及其组成部分在信息传递过程中的作用。文档的功能描述提供了对文档类型、预期用途以及自动文档解释和检索策略的深入了解。为了演示这些思想,我们定义了功能性文档组件的分类法,并展示了如何使用功能性描述来逆向工程作者的意图,在文档空间中导航,以及提供重要的上下文信息来帮助解释。
{"title":"The function of documents","authors":"D. Doermann, E. Rivlin, A. Rosenfeld","doi":"10.1109/ICDAR.1997.620674","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620674","url":null,"abstract":"The purpose of a document is to facilitate the transfer of information from its author to its readers. It is the author's job to design the document so that the information it contains can be interpreted accurately and efficiently. To do this, the author can make use of a set of stylistic tools. In this paper, we introduce the concept of document functionality, which attempts to describe the roles of documents and their components in the process of transferring information. A functional description of a document provides insight into the type of the document, into its intended uses, and into strategies for automatic document interpretation and retrieval. To demonstrate these ideas, we define a taxonomy of functional document components and show how functional descriptions can be used to reverse-engineer the intentions of the author, to navigate in document space, and to provide important contextual information to aid in interpretation.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126161990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
The HOVER system for rapid holistic verification of off-line handwritten phrases HOVER系统用于离线手写短语的快速整体验证
S. Madhvanath, Evelyn Kleinberg, V. Govindaraju, S. Srihari
The authors describe ongoing research on a system for rapid verification of unconstrained off-line handwritten phrases using perceptual holistic features of the handwritten phrase image. The system is used to verify handwritten street names automatically extracted from live US mail against recognition results of analytical classifiers. The system rejects errors with 98% accuracy at the 30% accept level, while consuming approximately 20 msec per image on the average on a 150 MHz SPARC 10.
作者描述了正在进行的一项系统研究,该系统使用手写短语图像的感知整体特征来快速验证无约束离线手写短语。该系统用于根据分析分类器的识别结果,验证从美国实时邮件中自动提取的手写街道名称。系统在30%接受水平下以98%的准确率拒绝错误,同时在150 MHz SPARC 10上平均每张图像消耗约20毫秒。
{"title":"The HOVER system for rapid holistic verification of off-line handwritten phrases","authors":"S. Madhvanath, Evelyn Kleinberg, V. Govindaraju, S. Srihari","doi":"10.1109/ICDAR.1997.620633","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620633","url":null,"abstract":"The authors describe ongoing research on a system for rapid verification of unconstrained off-line handwritten phrases using perceptual holistic features of the handwritten phrase image. The system is used to verify handwritten street names automatically extracted from live US mail against recognition results of analytical classifiers. The system rejects errors with 98% accuracy at the 30% accept level, while consuming approximately 20 msec per image on the average on a 150 MHz SPARC 10.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127535138","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Markov model order optimization for text recognition 文本识别的马尔可夫模型顺序优化
C. Olivier, F. Jouzel, M. Avila
Markov models are currently used for printed or handwritten word recognition. The order k is a very important parameter of these models. The aim of this paper is to use model selection criteria in order to estimate the order of a Markov model. Akaike (1973) suggested the AIC criterion for the estimation of the order k of a parameterized statistical model, including the term k as penalization of the likelihood function. Yet, selection according to this criterion leads asymptotically to a strict overestimation of the order. That is why we suggest the use of other consistent criteria in a Markovian case: the Bayesian and the Hannan and Quinn information criteria (BIC and /spl rho/, respectively). The performance of the criteria are analysed on simulated data and on a real case: a handwritten word description. We discuss the limit of these methods in relation to the number of states in the model.
马尔可夫模型目前用于打印或手写单词识别。k阶是这些模型的一个非常重要的参数。本文的目的是利用模型选择准则来估计马尔可夫模型的阶数。Akaike(1973)提出了估算参数化统计模型k阶的AIC准则,其中包括k项作为似然函数的惩罚项。然而,根据这一标准进行选择会逐渐导致对顺序的严格高估。这就是为什么我们建议在马尔可夫情况下使用其他一致的标准:贝叶斯和Hannan和Quinn信息标准(分别为BIC和/spl rho/)。在模拟数据和一个真实案例上分析了这些标准的性能:一个手写的单词描述。我们讨论了这些方法与模型中状态数的关系。
{"title":"Markov model order optimization for text recognition","authors":"C. Olivier, F. Jouzel, M. Avila","doi":"10.1109/ICDAR.1997.620560","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620560","url":null,"abstract":"Markov models are currently used for printed or handwritten word recognition. The order k is a very important parameter of these models. The aim of this paper is to use model selection criteria in order to estimate the order of a Markov model. Akaike (1973) suggested the AIC criterion for the estimation of the order k of a parameterized statistical model, including the term k as penalization of the likelihood function. Yet, selection according to this criterion leads asymptotically to a strict overestimation of the order. That is why we suggest the use of other consistent criteria in a Markovian case: the Bayesian and the Hannan and Quinn information criteria (BIC and /spl rho/, respectively). The performance of the criteria are analysed on simulated data and on a real case: a handwritten word description. We discuss the limit of these methods in relation to the number of states in the model.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127254671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
High speed rough classification for handwritten characters using hierarchical learning vector quantization 基于层次学习向量量化的手写体字符高速粗分类
Yuji Waizumi, N. Kato, Kazuki Saruta, Y. Nemoto
Today, high accuracy of character recognition is attainable using a neural network for problems with a relatively small number of categories. But for large categories, like Chinese characters, it is difficult to reach the neural network convergence because of the "local minima problem" and a large number of calculations. Studies are being done to solve the problem by splitting the neural network into some small modules. The effectiveness of the combination of learning vector quantization (LVQ) and back propagation (BP) has been reported. LVQ is used for rough classification and BP is used for fine recognition. It is difficult to obtain high accuracy for rough classification by LVQ itself. To deal with this problem, we propose hierarchical learning vector quantization (HLVQ). HLVQ divides categories in feature space hierarchically in the learning procedure. The adjacent feature spaces overlap each other near the borders. HLVQ possesses both classification speed and accuracy due to the hierarchical architecture and the overlapping technique. In the experiment using ETL9B, the largest database of handwritten characters in Japan, (includes 3036 categories, 607,200 samples), the effectiveness of HLVQ was verified.
目前,使用神经网络对相对较少类别的问题可以实现高精度的字符识别。但对于像汉字这样的大类别,由于“局部极小问题”和大量的计算,很难达到神经网络的收敛性。为了解决这个问题,人们正在进行研究,把神经网络分成一些小模块。已经报道了学习向量量化(LVQ)和反向传播(BP)相结合的有效性。LVQ用于粗分类,BP用于精细识别。仅靠LVQ本身很难获得较高的粗分类精度。为了解决这个问题,我们提出了层次学习向量量化(HLVQ)。HLVQ在学习过程中对特征空间中的类别进行分层划分。相邻的特征空间在边界附近相互重叠。由于层次结构和重叠技术,HLVQ具有快速和准确的分类能力。在使用日本最大的手写体数据库ETL9B(包含3036个类别,607,200个样本)的实验中,验证了HLVQ的有效性。
{"title":"High speed rough classification for handwritten characters using hierarchical learning vector quantization","authors":"Yuji Waizumi, N. Kato, Kazuki Saruta, Y. Nemoto","doi":"10.1109/ICDAR.1997.619807","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619807","url":null,"abstract":"Today, high accuracy of character recognition is attainable using a neural network for problems with a relatively small number of categories. But for large categories, like Chinese characters, it is difficult to reach the neural network convergence because of the \"local minima problem\" and a large number of calculations. Studies are being done to solve the problem by splitting the neural network into some small modules. The effectiveness of the combination of learning vector quantization (LVQ) and back propagation (BP) has been reported. LVQ is used for rough classification and BP is used for fine recognition. It is difficult to obtain high accuracy for rough classification by LVQ itself. To deal with this problem, we propose hierarchical learning vector quantization (HLVQ). HLVQ divides categories in feature space hierarchically in the learning procedure. The adjacent feature spaces overlap each other near the borders. HLVQ possesses both classification speed and accuracy due to the hierarchical architecture and the overlapping technique. In the experiment using ETL9B, the largest database of handwritten characters in Japan, (includes 3036 categories, 607,200 samples), the effectiveness of HLVQ was verified.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130586946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Automatic separation of words in multi-lingual multi-script Indian documents 自动分词在多语言多脚本印度文件
U. Pal, B. B. Chaudhuri
In a multi-lingual country like India, a document may contain more than one script forms. For such a document it is necessary to separate different script forms before feeding them to OCRs of individual script. In this paper an automatic word segmentation approach is described which can separate Roman, Bangla and Devnagari scripts present in a single document. The approach has a tree structure where at first Roman script words are separated using the 'headline' feature. The headline is common in Bangla and Devnagari but absent in Roman. Next, Bangla and Devnagari words are separated using some finer characteristics of the character set although recognition of individual character is avoided. At present, the system has an overall accuracy of 96.09%.
在像印度这样的多语言国家,一个文档可能包含不止一种脚本形式。对于这样的文档,有必要在将不同的脚本表单提供给单个脚本的ocr之前分离它们。本文描述了一种自动分词方法,该方法可以将单个文档中的罗马语、孟加拉语和德文加里语三种文字分离出来。该方法有一个树状结构,首先罗马文字使用“标题”特征分开。这个标题在孟加拉语和德文加里语中很常见,但在罗马语中却没有。接下来,使用字符集的一些更精细的特征将孟加拉语和德文加里语分开,尽管避免了对单个字符的识别。目前,该系统的总体准确率为96.09%。
{"title":"Automatic separation of words in multi-lingual multi-script Indian documents","authors":"U. Pal, B. B. Chaudhuri","doi":"10.1109/ICDAR.1997.620567","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620567","url":null,"abstract":"In a multi-lingual country like India, a document may contain more than one script forms. For such a document it is necessary to separate different script forms before feeding them to OCRs of individual script. In this paper an automatic word segmentation approach is described which can separate Roman, Bangla and Devnagari scripts present in a single document. The approach has a tree structure where at first Roman script words are separated using the 'headline' feature. The headline is common in Bangla and Devnagari but absent in Roman. Next, Bangla and Devnagari words are separated using some finer characteristics of the character set although recognition of individual character is avoided. At present, the system has an overall accuracy of 96.09%.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132057149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Recognition of printed Arabic text using neural networks 用神经网络识别印刷阿拉伯文本
A. Amin, W. Mansoor
The main theme of the paper is the automatic recognition of Arabic printed text using artificial neural networks in addition to conventional techniques. This approach has a number of advantages: it combines rule based (structural) and classification tests; feature extraction is inexpensive; and execution time is independent of character font and size. The technique can be divided into three major steps: The first step is preprocessing in which the original image is transformed into a binary image utilizing a 300 dpi scanner and then forming the connected component. Second, global features of the input Arabic word are then extracted such as number of subwords, number of peaks within the subword, number and position of the complementary character, etc. Finally, an artificial neural network is used for character classification. The algorithm was implemented on a powerful MS-DOS microcomputer and written in C.
本文的主题是在传统技术的基础上,利用人工神经网络对阿拉伯语印刷文本进行自动识别。这种方法有很多优点:它结合了基于规则的(结构)和分类测试;特征提取成本低;并且执行时间与字符字体和大小无关。该技术可分为三个主要步骤:第一步是预处理,其中原始图像转换成二值图像利用300 dpi扫描仪,然后形成连接的组件。其次,提取输入的阿拉伯语单词的全局特征,如子词数、子词内的峰数、互补字符的数量和位置等;最后,利用人工神经网络对字符进行分类。该算法在功能强大的MS-DOS微机上实现,并用C语言编写。
{"title":"Recognition of printed Arabic text using neural networks","authors":"A. Amin, W. Mansoor","doi":"10.1109/ICDAR.1997.620576","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.620576","url":null,"abstract":"The main theme of the paper is the automatic recognition of Arabic printed text using artificial neural networks in addition to conventional techniques. This approach has a number of advantages: it combines rule based (structural) and classification tests; feature extraction is inexpensive; and execution time is independent of character font and size. The technique can be divided into three major steps: The first step is preprocessing in which the original image is transformed into a binary image utilizing a 300 dpi scanner and then forming the connected component. Second, global features of the input Arabic word are then extracted such as number of subwords, number of peaks within the subword, number and position of the complementary character, etc. Finally, an artificial neural network is used for character classification. The algorithm was implemented on a powerful MS-DOS microcomputer and written in C.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132200786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Constraints on handwriting Korean characters to improve the machine readability 对手写韩文汉字的限制以提高机器可读性
Choi Baek Young, S. Bang
Realizing that the availability of a practical recognition system for Korean handwritten characters without any constraints has a long way to go, we have attempted to find a set of writing constraints which significantly improves the machine-readability. Based on our observation that the majority of the misrecognitions reported are caused by ambiguous characters, we have developed a set of writing constraints which maximally disambiguate those characters. Through experiments, we have confirmed that the recognition rate of those handwritten data could be improved significantly by applying the proposed set of constraints.
意识到一个没有任何约束的实用的韩文手写字符识别系统的可用性还有很长的路要走,我们试图找到一组显著提高机器可读性的书写约束。根据我们的观察,报告的大多数错误识别是由歧义字符引起的,我们开发了一套最大限度地消除这些字符歧义的书写约束。通过实验,我们证实了应用所提出的约束集可以显著提高这些手写数据的识别率。
{"title":"Constraints on handwriting Korean characters to improve the machine readability","authors":"Choi Baek Young, S. Bang","doi":"10.1109/ICDAR.1997.619840","DOIUrl":"https://doi.org/10.1109/ICDAR.1997.619840","url":null,"abstract":"Realizing that the availability of a practical recognition system for Korean handwritten characters without any constraints has a long way to go, we have attempted to find a set of writing constraints which significantly improves the machine-readability. Based on our observation that the majority of the misrecognitions reported are caused by ambiguous characters, we have developed a set of writing constraints which maximally disambiguate those characters. Through experiments, we have confirmed that the recognition rate of those handwritten data could be improved significantly by applying the proposed set of constraints.","PeriodicalId":435320,"journal":{"name":"Proceedings of the Fourth International Conference on Document Analysis and Recognition","volume":"113 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126706590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
Proceedings of the Fourth International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1