首页 > 最新文献

Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.最新文献

英文 中文
N-gram and N-class models for on line handwriting recognition 在线手写识别的N-gram和N-class模型
Freddy Perraud, C. Viard-Gaudin, E. Morin, P. Lallican
This paper highlights the interest of a language modelin increasing the performances of on-line handwritingrecognition systems. Models based on statisticalapproaches, trained on written corpora, have beeninvestigated. Two kinds of models have been studied: n-grammodels and n-class models. In the latter case, theclasses result either from a syntactic criteria or acontextual criteria. In order to integrate it into smallcapacity systems (mobile device), an n-class model hasbeen designed by combining these criteria. It outperformsbulkier models based on n-gram. Integration into an on-linehandwriting recognition system demonstrates asubstantial performance improvement due to the languagemodel.
本文强调了语言模型在提高在线手写识别系统性能方面的作用。基于统计方法的模型,在书面语料库上训练,已经被调查。研究了两种模型:n-gram模型和n-class模型。在后一种情况下,类要么来自语法标准,要么来自上下文标准。为了将其集成到小容量系统(移动设备)中,结合这些标准设计了一个n级模型。它优于基于n-gram的笨重模型。集成到在线手写识别系统中,由于该语言模型,性能得到了实质性的改善。
{"title":"N-gram and N-class models for on line handwriting recognition","authors":"Freddy Perraud, C. Viard-Gaudin, E. Morin, P. Lallican","doi":"10.1109/ICDAR.2003.1227818","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227818","url":null,"abstract":"This paper highlights the interest of a language modelin increasing the performances of on-line handwritingrecognition systems. Models based on statisticalapproaches, trained on written corpora, have beeninvestigated. Two kinds of models have been studied: n-grammodels and n-class models. In the latter case, theclasses result either from a syntactic criteria or acontextual criteria. In order to integrate it into smallcapacity systems (mobile device), an n-class model hasbeen designed by combining these criteria. It outperformsbulkier models based on n-gram. Integration into an on-linehandwriting recognition system demonstrates asubstantial performance improvement due to the languagemodel.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130460461","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
A vector approach for automatic interpretation of the French cadastral map 矢量法自动解译法国地籍图
Jean-Marc Viglino, M. Pierrot-Deseilligny
This paper deals with cadastral maps interpretation device. The challenge is to propose a complete reconstruction of the parcel's areas and buildings to use with geographic information systems. It is based on a low level primitives extraction and classification. As this low level may be quite noisy, an interpretation process classifies medium level objects and manages convenient processes to the particular extracted shape. Then, a reconstruction step is used to label the parcels areas and determine the final land partition. We present at first the vectorization strategy in our particular context then we will discuss the different tools used to reach the higher level.
本文研究地籍图解释装置。挑战是提出一个完整的重建包裹的区域和建筑,以使用地理信息系统。它基于低级原语的提取和分类。由于这种低层次可能相当嘈杂,解释过程对中等层次对象进行分类,并管理对特定提取形状的方便处理。然后,重建步骤用于标记地块区域并确定最终的土地分区。我们首先在我们的特定环境中提出矢量化策略,然后我们将讨论用于达到更高级别的不同工具。
{"title":"A vector approach for automatic interpretation of the French cadastral map","authors":"Jean-Marc Viglino, M. Pierrot-Deseilligny","doi":"10.1109/ICDAR.2003.1227678","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227678","url":null,"abstract":"This paper deals with cadastral maps interpretation device. The challenge is to propose a complete reconstruction of the parcel's areas and buildings to use with geographic information systems. It is based on a low level primitives extraction and classification. As this low level may be quite noisy, an interpretation process classifies medium level objects and manages convenient processes to the particular extracted shape. Then, a reconstruction step is used to label the parcels areas and determine the final land partition. We present at first the vectorization strategy in our particular context then we will discuss the different tools used to reach the higher level.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121432170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Combining model-based and discriminative classifiers: application to handwritten character recognition 结合基于模型和判别分类器:在手写体字符识别中的应用
L. Prevost, C. Michel-Sendis, A. Moises, L. Oudot, M. Milgram
Handwriting recognition is such a complex classification problem that it is quite usual now to make co-operate several classification methods at the pre-processing stage or at the classification stage. In this paper, we present an original two stages recognizer. The first stage is a model-based classifier that stores an exhaustive set of character models. The second stage is a discriminative classifier that separates the most ambiguous pairs of classes. This hybrid architecture is based on the idea that the correct class almost systematically belongs to the two more relevant classes found by the first classifier. Experiments on the Unipen database show a 30% improvement on a 62 class recognition problem.
手写识别是一个非常复杂的分类问题,通常在预处理阶段或分类阶段采用多种分类方法进行协作。本文提出了一种新颖的两阶段识别器。第一阶段是基于模型的分类器,它存储一组详尽的字符模型。第二阶段是判别分类器,分离最模糊的类对。这种混合体系结构基于这样一种思想,即正确的类几乎系统地属于第一个分类器发现的两个更相关的类。在Unipen数据库上的实验表明,在62个类的识别问题上提高了30%。
{"title":"Combining model-based and discriminative classifiers: application to handwritten character recognition","authors":"L. Prevost, C. Michel-Sendis, A. Moises, L. Oudot, M. Milgram","doi":"10.1109/ICDAR.2003.1227623","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227623","url":null,"abstract":"Handwriting recognition is such a complex classification problem that it is quite usual now to make co-operate several classification methods at the pre-processing stage or at the classification stage. In this paper, we present an original two stages recognizer. The first stage is a model-based classifier that stores an exhaustive set of character models. The second stage is a discriminative classifier that separates the most ambiguous pairs of classes. This hybrid architecture is based on the idea that the correct class almost systematically belongs to the two more relevant classes found by the first classifier. Experiments on the Unipen database show a 30% improvement on a 62 class recognition problem.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123085334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Fast lexicon-based word recognition in noisy index card images 在噪声索引卡图像中快速基于词典的单词识别
S. Lucas, Gregory Patoulas, A. Downton
This paper describes a complete system for reading type-written lexicon words in noisy images - in this case museum index cards. The system is conceptually simple, and straightforward to implement. It involves three stages of processing. The first stage extracts row-regions from the image, where each row is a hypothesized line of text. The next stage scans an OCR classifier over each row image, creating a character hypothesis graph in the process. This graph is then searched using a priority-queue based algorithm for the best matches with a set of words (lexicon). Performance evaluation on a set of museum archive cards indicates competitive accuracy and also reasonable throughput. The priority queue algorithm is over two hundred times faster than using flat dynamic programming on these graphs.
本文描述了一个完整的系统,用于在嘈杂的图像中阅读打字的词汇-在本例中是博物馆索引卡。该系统在概念上很简单,并且易于实现。它包括三个处理阶段。第一阶段从图像中提取行区域,其中每一行都是假设的文本行。下一阶段扫描每个行图像的OCR分类器,在此过程中创建一个字符假设图。然后使用基于优先级队列的算法搜索此图,以寻找与一组单词(词典)的最佳匹配。对一套博物馆档案卡的性能评估表明具有竞争力的准确性和合理的吞吐量。优先级队列算法比在这些图上使用平面动态规划快200多倍。
{"title":"Fast lexicon-based word recognition in noisy index card images","authors":"S. Lucas, Gregory Patoulas, A. Downton","doi":"10.1109/ICDAR.2003.1227708","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227708","url":null,"abstract":"This paper describes a complete system for reading type-written lexicon words in noisy images - in this case museum index cards. The system is conceptually simple, and straightforward to implement. It involves three stages of processing. The first stage extracts row-regions from the image, where each row is a hypothesized line of text. The next stage scans an OCR classifier over each row image, creating a character hypothesis graph in the process. This graph is then searched using a priority-queue based algorithm for the best matches with a set of words (lexicon). Performance evaluation on a set of museum archive cards indicates competitive accuracy and also reasonable throughput. The priority queue algorithm is over two hundred times faster than using flat dynamic programming on these graphs.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"94 17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126056467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Detection of matrices and segmentation of matrix elements in scanned images of scientific documents 科学文献扫描图像中矩阵的检测与矩阵元素的分割
T. Kanahori, M. Suzuki
We proposed a method for recognizing matrices which contain abbreviation symbols, and a format for representing the structure of matrices, and reported experimental results in our paper (2002). The method consisted of 4 processes: detection of matrices, segmentation of elements, construction of networks and analysis of the matrix structure. In the paper, our work is described with a focus on the construction of networks and the analysis of the matrix structure. However, we concluded that improvements in the other two processes were very important for obtaining a high accuracy rate for recognition. In this paper, we describe the two improved processes, the detection of matrices and the segmentation of elements, and we report the experimental results.
我们提出了一种识别包含缩写符号的矩阵的方法,以及一种表示矩阵结构的格式,并在我们的论文(2002)中报告了实验结果。该方法包括4个步骤:矩阵检测、元素分割、网络构建和矩阵结构分析。在本文中,我们的工作重点是网络的构建和矩阵结构的分析。然而,我们得出结论,其他两个过程的改进对于获得较高的识别准确率非常重要。本文描述了矩阵检测和元素分割这两个改进过程,并给出了实验结果。
{"title":"Detection of matrices and segmentation of matrix elements in scanned images of scientific documents","authors":"T. Kanahori, M. Suzuki","doi":"10.1109/ICDAR.2003.1227704","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227704","url":null,"abstract":"We proposed a method for recognizing matrices which contain abbreviation symbols, and a format for representing the structure of matrices, and reported experimental results in our paper (2002). The method consisted of 4 processes: detection of matrices, segmentation of elements, construction of networks and analysis of the matrix structure. In the paper, our work is described with a focus on the construction of networks and the analysis of the matrix structure. However, we concluded that improvements in the other two processes were very important for obtaining a high accuracy rate for recognition. In this paper, we describe the two improved processes, the detection of matrices and the segmentation of elements, and we report the experimental results.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121118519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Postal envelope segmentation by 2-D histogram clustering through watershed transform 基于分水岭变换的二维直方图聚类邮政信封分割
Eduardo Akira Yonekura, J. Facon
In this paper we present a new postal envelope segmentation method based on 2-D histogram clustering and watershed transform. Segmentation task consists in detecting the modes associated with homogeneous regions in envelope images such as handwritten address block, postmarks, stamps and background. The homogeneous modes in 2-D histogram are segmented through the morphological watershed transform. Our approach is applied to complex Brazilian postal envelopes. Very little a priori knowledge of the envelope images is required. The advantages of this approach will be described and illustrated with tests carried out on 300 different images where there are no fixed position for the handwritten address block, postmarks and stamps.
本文提出了一种基于二维直方图聚类和分水岭变换的邮政信封分割方法。分割任务包括检测信封图像中手写地址块、邮戳、邮票和背景等同质区域的相关模式。通过形态学分水岭变换对二维直方图中的均匀模式进行分割。我们的方法适用于复杂的巴西邮政信封。很少先验知识的包络图像是必需的。本文将描述和说明这种方法的优点,并对300张不同的图像进行测试,其中手写地址块、邮戳和邮票没有固定的位置。
{"title":"Postal envelope segmentation by 2-D histogram clustering through watershed transform","authors":"Eduardo Akira Yonekura, J. Facon","doi":"10.1109/ICDAR.2003.1227685","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227685","url":null,"abstract":"In this paper we present a new postal envelope segmentation method based on 2-D histogram clustering and watershed transform. Segmentation task consists in detecting the modes associated with homogeneous regions in envelope images such as handwritten address block, postmarks, stamps and background. The homogeneous modes in 2-D histogram are segmented through the morphological watershed transform. Our approach is applied to complex Brazilian postal envelopes. Very little a priori knowledge of the envelope images is required. The advantages of this approach will be described and illustrated with tests carried out on 300 different images where there are no fixed position for the handwritten address block, postmarks and stamps.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131807875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Best practices for convolutional neural networks applied to visual document analysis 卷积神经网络应用于可视化文档分析的最佳实践
P. Simard, David Steinkraus, John C. Platt
Neural networks are a powerful technology forclassification of visual inputs arising from documents.However, there is a confusing plethora of different neuralnetwork methods that are used in the literature and inindustry. This paper describes a set of concrete bestpractices that document analysis researchers can use toget good results with neural networks. The mostimportant practice is getting a training set as large aspossible: we expand the training set by adding a newform of distorted data. The next most important practiceis that convolutional neural networks are better suited forvisual document tasks than fully connected networks. Wepropose that a simple "do-it-yourself" implementation ofconvolution with a flexible architecture is suitable formany visual document problems. This simpleconvolutional neural network does not require complexmethods, such as momentum, weight decay, structure-dependentlearning rates, averaging layers, tangent prop,or even finely-tuning the architecture. The end result is avery simple yet general architecture which can yieldstate-of-the-art performance for document analysis. Weillustrate our claims on the MNIST set of English digitimages.
神经网络是一种强大的技术,用于分类来自文档的视觉输入。然而,在文献和工业中使用的不同的神经网络方法令人困惑。本文描述了一组具体的最佳实践,文件分析研究人员可以使用神经网络获得良好的结果。最重要的实践是获得尽可能大的训练集:我们通过添加新形式的扭曲数据来扩展训练集。下一个最重要的实践是,卷积神经网络比完全连接的网络更适合于视觉文档任务。我们提出一个简单的“自己动手”的卷积实现,具有灵活的架构,适用于许多可视化文档问题。这个简单的卷积神经网络不需要复杂的方法,比如动量、权重衰减、结构相关学习率、平均层、切线支撑,甚至微调架构。最终的结果是非常简单而通用的架构,可以为文档分析提供最先进的性能。我们用MNIST的英语数字图像集来说明我们的主张。
{"title":"Best practices for convolutional neural networks applied to visual document analysis","authors":"P. Simard, David Steinkraus, John C. Platt","doi":"10.1109/ICDAR.2003.1227801","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227801","url":null,"abstract":"Neural networks are a powerful technology forclassification of visual inputs arising from documents.However, there is a confusing plethora of different neuralnetwork methods that are used in the literature and inindustry. This paper describes a set of concrete bestpractices that document analysis researchers can use toget good results with neural networks. The mostimportant practice is getting a training set as large aspossible: we expand the training set by adding a newform of distorted data. The next most important practiceis that convolutional neural networks are better suited forvisual document tasks than fully connected networks. Wepropose that a simple \"do-it-yourself\" implementation ofconvolution with a flexible architecture is suitable formany visual document problems. This simpleconvolutional neural network does not require complexmethods, such as momentum, weight decay, structure-dependentlearning rates, averaging layers, tangent prop,or even finely-tuning the architecture. The end result is avery simple yet general architecture which can yieldstate-of-the-art performance for document analysis. Weillustrate our claims on the MNIST set of English digitimages.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"4 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131505750","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2755
Directional wavelet approach to remove document image interference 方向小波法去除文档图像干扰
Qian Wang, Tao Xia, C. Tan, Lida Li
In this paper, we propose a directional wavelet approachto remove images of interfering strokes coming from theback of a historical handwritten document due to seepingof ink during long period of storage. Our previous workrequired mapping of both sides of the document in orderto identify the interfering strokes to be eliminated. Perfectmapping, however, is difficult due to document skews,differing resolutions, non-availability of the reverseside and warped pages during scanning. The newapproach does not require double-sided mapping butinstead uses a directional wavelet transformto distinguish the foreground and reverse side strokes.Experiments have shown that the directional waveletoperation effectively removes the interfering strokes.
在本文中,我们提出了一种定向小波方法来去除由于长时间存储期间墨水渗出而来自历史手写文件背面的干扰笔画图像。我们之前的工作需要对文档的两面进行映射,以便识别要消除的干扰笔画。然而,由于文档倾斜、不同的分辨率、不可用的反面和扫描时扭曲的页面,完美的映射是困难的。新方法不需要双面映射,而是使用方向小波变换来区分前景和反面笔画。实验表明,定向小波运算能有效地去除干扰笔画。
{"title":"Directional wavelet approach to remove document image interference","authors":"Qian Wang, Tao Xia, C. Tan, Lida Li","doi":"10.1109/ICDAR.2003.1227759","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227759","url":null,"abstract":"In this paper, we propose a directional wavelet approachto remove images of interfering strokes coming from theback of a historical handwritten document due to seepingof ink during long period of storage. Our previous workrequired mapping of both sides of the document in orderto identify the interfering strokes to be eliminated. Perfectmapping, however, is difficult due to document skews,differing resolutions, non-availability of the reverseside and warped pages during scanning. The newapproach does not require double-sided mapping butinstead uses a directional wavelet transformto distinguish the foreground and reverse side strokes.Experiments have shown that the directional waveletoperation effectively removes the interfering strokes.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132558670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Writer identification based on the fractal construction of a reference base 基于分形构造的作家识别参考库
A. Seropian, M. Grimaldi, N. Vincent
Our aim is to achieve writer identification processthanks to a fractal analysis of handwriting style. For eachwriter, a set of characteristics is extracted. They arespecific to the writer. Advantage is taken from theautosimilarity properties that are present in one'shandwriting. In order to do that, some invariant patternscharacterizing the writing are extracted. During thetraining step these invariant patterns appear along afractal compression process, then they are organized in areference base that can be associated with the writer.This base allows to analyze an unknown writing thewriter of which has to be identified. A Pattern Matchingprocess is performed using all the reference basessuccessively. The results of this analyze are estimatedthrough the signal to noise ratio. Thus, the signal to noiseratio according to a set of bases identifies the unknowntext's writer.
我们的目标是通过对笔迹风格的分形分析来实现作者的识别过程。对于每个writer,提取一组特征。它们是特定于作者的。优点是从一个人的笔迹中呈现的自相似特性中获得的。为了做到这一点,提取了一些具有书写特征的不变模式。在训练阶段,这些不变模式沿着分形压缩过程出现,然后将它们组织在可以与编写器相关联的参考库中。这个基础允许分析一个未知的写作,其作者必须确定。一个模式匹配过程使用所有的引用库依次执行。通过信噪比对分析结果进行估计。因此,根据一组碱基的信噪比来识别未知文本的作者。
{"title":"Writer identification based on the fractal construction of a reference base","authors":"A. Seropian, M. Grimaldi, N. Vincent","doi":"10.1109/ICDAR.2003.1227840","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227840","url":null,"abstract":"Our aim is to achieve writer identification processthanks to a fractal analysis of handwriting style. For eachwriter, a set of characteristics is extracted. They arespecific to the writer. Advantage is taken from theautosimilarity properties that are present in one'shandwriting. In order to do that, some invariant patternscharacterizing the writing are extracted. During thetraining step these invariant patterns appear along afractal compression process, then they are organized in areference base that can be associated with the writer.This base allows to analyze an unknown writing thewriter of which has to be identified. A Pattern Matchingprocess is performed using all the reference basessuccessively. The results of this analyze are estimatedthrough the signal to noise ratio. Thus, the signal to noiseratio according to a set of bases identifies the unknowntext's writer.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"208 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115903434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Model length adaptation of an HMM based cursive word recognition system 基于HMM的草书词识别系统模型长度自适应
M. Schambach
On the basis of a well accepted, HMM-based cursive script recognition system, an algorithm which automatically adapts the length of the models representing the letter writing variants is proposed. An average improvement in recognition performance of about 2.72 percent could be obtained. Two initialization methods for the algorithm have been tested, which show quite different behaviors; both prove to be useful in different application areas. To get a deeper insight into the functioning of the algorithm a method for the visualization of letter HMMs is developed. It shows the plausibility of most results, but also the limitations of the proposed method. However, these are mostly due to given restrictions of the training and recognition method of the underlying system.
在一个公认的基于hmm的草书识别系统的基础上,提出了一种自动适应代表字母书写变体的模型长度的算法。识别性能的平均提高约为2.72%。对该算法的两种初始化方法进行了测试,结果表明两种方法的行为完全不同;两者都被证明在不同的应用领域是有用的。为了更深入地了解该算法的功能,开发了一种字母hmm的可视化方法。它表明了大多数结果的合理性,但也表明了所提出方法的局限性。然而,这主要是由于底层系统的训练和识别方法受到一定的限制。
{"title":"Model length adaptation of an HMM based cursive word recognition system","authors":"M. Schambach","doi":"10.1109/ICDAR.2003.1227642","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227642","url":null,"abstract":"On the basis of a well accepted, HMM-based cursive script recognition system, an algorithm which automatically adapts the length of the models representing the letter writing variants is proposed. An average improvement in recognition performance of about 2.72 percent could be obtained. Two initialization methods for the algorithm have been tested, which show quite different behaviors; both prove to be useful in different application areas. To get a deeper insight into the functioning of the algorithm a method for the visualization of letter HMMs is developed. It shows the plausibility of most results, but also the limitations of the proposed method. However, these are mostly due to given restrictions of the training and recognition method of the underlying system.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131995947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
期刊
Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1