首页 > 最新文献

Proceedings of Sixth International Conference on Document Analysis and Recognition最新文献

英文 中文
Handwritten numeral recognition using flexible matching based on learning of stroke statistics 基于笔画统计学习的灵活匹配手写数字识别
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953862
Takashi Kobayashi, Kaori Nakamura, Hirokazu Muramatsu, Takahiro Sugiyama, K. Abe
The purpose of this study is to learn shapes and structures of a given learning set of handwritten numerals and to develop a flexible matching method for recognition based on the learning. First, this paper proposes a method of how to obtain a set of standard character patterns and the ranges of variations varying statistically from the given learning character samples. Then the recognition is made as follows: each standard pattern is deformed to match with the input character; and the matching is evaluated by the energy of deformation; and the closeness of the standard pattern to the input.
本研究的目的是学习给定的手写数字学习集的形状和结构,并在学习的基础上开发一种灵活的匹配识别方法。首先,本文提出了一种如何从给定的学习字符样本中获得一组标准字符模式和统计变化范围的方法。然后进行识别:对每个标准模式进行变形,使其与输入字符匹配;用变形能量来评价匹配度;以及标准模式与输入的接近程度。
{"title":"Handwritten numeral recognition using flexible matching based on learning of stroke statistics","authors":"Takashi Kobayashi, Kaori Nakamura, Hirokazu Muramatsu, Takahiro Sugiyama, K. Abe","doi":"10.1109/ICDAR.2001.953862","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953862","url":null,"abstract":"The purpose of this study is to learn shapes and structures of a given learning set of handwritten numerals and to develop a flexible matching method for recognition based on the learning. First, this paper proposes a method of how to obtain a set of standard character patterns and the ranges of variations varying statistically from the given learning character samples. Then the recognition is made as follows: each standard pattern is deformed to match with the input character; and the matching is evaluated by the energy of deformation; and the closeness of the standard pattern to the input.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128170920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Adaptive N-best-list handwritten word recognition 自适应n最佳列表手写单词识别
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953777
T. Kwok, M. Perrone
We investigate a novel method for adaptively improving the machine recognition of handwritten words by applying a k-nearest neighbor (k-NN) classifier to the N-best word-hypothesis lists generated by a writer-independent hidden Markov model (HMM). Each new N-best list from the HMM is compared to the N-best lists in the k-NN classifier. A decision module is used to select between the output of the HMM and the matches found by the k-NN classifier. The N-best list chosen by the decision module can be automatically added to the k-NN classifier if it is not already in the k-NN classifier. This dynamic update of the k-NN classifier enables the system to adapt to new data without retraining. On a writer-independent set of 1158 handwritten words, this method reduces the error rate by approximately 30%. This method is fast and memory-efficient, and lends itself to many interesting generalizations.
我们研究了一种自适应改进手写单词机器识别的新方法,该方法将k-最近邻(k-NN)分类器应用于由作者独立的隐马尔可夫模型(HMM)生成的n个最佳单词假设列表。HMM中的每个新的n个最佳列表与k-NN分类器中的n个最佳列表进行比较。决策模块用于在HMM的输出和k-NN分类器找到的匹配之间进行选择。决策模块选择的n个最佳列表如果不在k-NN分类器中,则可以自动添加到k-NN分类器中。这种k-NN分类器的动态更新使系统无需重新训练即可适应新数据。在1158个独立于写作者的手写单词集上,该方法将错误率降低了大约30%。这种方法速度快,内存效率高,可以进行许多有趣的推广。
{"title":"Adaptive N-best-list handwritten word recognition","authors":"T. Kwok, M. Perrone","doi":"10.1109/ICDAR.2001.953777","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953777","url":null,"abstract":"We investigate a novel method for adaptively improving the machine recognition of handwritten words by applying a k-nearest neighbor (k-NN) classifier to the N-best word-hypothesis lists generated by a writer-independent hidden Markov model (HMM). Each new N-best list from the HMM is compared to the N-best lists in the k-NN classifier. A decision module is used to select between the output of the HMM and the matches found by the k-NN classifier. The N-best list chosen by the decision module can be automatically added to the k-NN classifier if it is not already in the k-NN classifier. This dynamic update of the k-NN classifier enables the system to adapt to new data without retraining. On a writer-independent set of 1158 handwritten words, this method reduces the error rate by approximately 30%. This method is fast and memory-efficient, and lends itself to many interesting generalizations.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125454021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Text extraction from color documents-clustering approaches in three and four dimensions 彩色文档的文本提取——三维和四维聚类方法
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953923
T. Perroud, K. Sobottka, H. Bunke, L. Hall
Colored paper documents often contain important text information. For automating the retrieval process, identification of text elements is essential. In order to reduce the number of colors in a scanned document, color clustering is usually done first. In this article two histogram-based color clustering algorithms are investigated. The first is based on the RGB color space exclusively, while the second takes spatial information into account, in addition to the colors. Experimental results have shown that the use of spatial information in the clustering algorithm has a positive impact. Thus the automatic retrieval of text information can be improved. The proposed methods for clustering are not restricted to document images. They can also be used for processing Web or video images, for example.
彩色纸文档通常包含重要的文本信息。为了使检索过程自动化,文本元素的识别是必不可少的。为了减少扫描文档中的颜色数量,通常首先进行颜色聚类。本文研究了两种基于直方图的颜色聚类算法。第一种方法是完全基于RGB色彩空间,而第二种方法除了色彩之外还考虑了空间信息。实验结果表明,在聚类算法中使用空间信息具有积极的影响。从而提高文本信息的自动检索能力。本文提出的聚类方法并不局限于文档图像。例如,它们还可以用于处理网络或视频图像。
{"title":"Text extraction from color documents-clustering approaches in three and four dimensions","authors":"T. Perroud, K. Sobottka, H. Bunke, L. Hall","doi":"10.1109/ICDAR.2001.953923","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953923","url":null,"abstract":"Colored paper documents often contain important text information. For automating the retrieval process, identification of text elements is essential. In order to reduce the number of colors in a scanned document, color clustering is usually done first. In this article two histogram-based color clustering algorithms are investigated. The first is based on the RGB color space exclusively, while the second takes spatial information into account, in addition to the colors. Experimental results have shown that the use of spatial information in the clustering algorithm has a positive impact. Thus the automatic retrieval of text information can be improved. The proposed methods for clustering are not restricted to document images. They can also be used for processing Web or video images, for example.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"01 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121623493","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Robust feature extraction based on run-length compensation for degraded handwritten character recognition 基于游程补偿的退化手写字符识别鲁棒特征提取
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953870
M. Mori, M. Sawaki, N. Hagita, H. Murase, N. Mukawa
Conventional features are robust for recognizing either deformed or degraded characters. This paper proposes a feature extraction method that is robust for both of them. Run-length compensation is introduced for extracting approximate directional run-lengths of strokes from degraded handwritten characters. This technique is applied to the conventional feature vector based on directional run-lengths. Experiments for handwritten characters with additive or subtractive noise show that the proposed feature is superior to conventional ones over a wide range of the degree of noise.
传统特征对于识别变形或退化的字符都是鲁棒的。本文提出了一种对两者都具有鲁棒性的特征提取方法。为从退化的手写字符中提取笔画的大致方向行长,引入了行长补偿。将该技术应用于基于方向行程长度的传统特征向量。对带有加性和减性噪声的手写字符进行的实验表明,在较大的噪声范围内,所提出的特征优于传统特征。
{"title":"Robust feature extraction based on run-length compensation for degraded handwritten character recognition","authors":"M. Mori, M. Sawaki, N. Hagita, H. Murase, N. Mukawa","doi":"10.1109/ICDAR.2001.953870","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953870","url":null,"abstract":"Conventional features are robust for recognizing either deformed or degraded characters. This paper proposes a feature extraction method that is robust for both of them. Run-length compensation is introduced for extracting approximate directional run-lengths of strokes from degraded handwritten characters. This technique is applied to the conventional feature vector based on directional run-lengths. Experiments for handwritten characters with additive or subtractive noise show that the proposed feature is superior to conventional ones over a wide range of the degree of noise.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132045130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Text line segmentation and word recognition in a system for general writer independent handwriting recognition 文本行分割和词识别系统中一般写作者独立的手写识别
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953775
Urs-Viktor Marti, H. Bunke
We present a system for recognizing unconstrained English handwritten text based on a large vocabulary. We describe the three main components of the system, which are preprocessing, feature extraction and recognition. In the preprocessing phase the handwritten texts are first segmented into lines. Then each line of text is normalized with respect to of skew, slant, vertical position and width. After these steps, text lines are segmented into single words. For this purpose distances between connected components are measured. Using a threshold, the distances are divided into distances within a word and distances between different words. A line of text is segmented at positions where the distances are larger than the chosen threshold. From each image representing a single word, a sequence of features is extracted. These features are input to a recognition procedure which is based on hidden Markov models. To investigate the stability of the segmentation algorithm the threshold that separates intra- and inter-word distances from each other is varied. If the threshold is small many errors are caused by over-segmentation, while for large thresholds under-segmentation errors occur. The best segmentation performance is 95.56% correctly segmented words, tested on 541 text lines containing 3899 words. Given a correct segmentation rate of 95.56%, a recognition rate of 73.45% on the word level is achieved.
提出了一种基于大词汇量的无约束英语手写文本识别系统。介绍了该系统的三个主要组成部分:预处理、特征提取和识别。在预处理阶段,手写文本首先被分割成行。然后,对每一行文本的倾斜、倾斜、垂直位置和宽度进行归一化。在这些步骤之后,文本行被分割成单个单词。为此,测量连接组件之间的距离。使用阈值,将距离分为单词内的距离和不同单词之间的距离。在距离大于所选阈值的位置上分割一行文本。从每个代表单个单词的图像中提取一系列特征。这些特征输入到基于隐马尔可夫模型的识别过程中。为了研究分割算法的稳定性,我们改变了分割词内和词间距离的阈值。当阈值较小时,许多错误是由过分割引起的,而当阈值较大时,则会出现分段不足错误。在包含3899个单词的541行文本上测试,最佳分割性能是95.56%的正确分割单词。在正确分割率为95.56%的情况下,在词级上的识别率为73.45%。
{"title":"Text line segmentation and word recognition in a system for general writer independent handwriting recognition","authors":"Urs-Viktor Marti, H. Bunke","doi":"10.1109/ICDAR.2001.953775","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953775","url":null,"abstract":"We present a system for recognizing unconstrained English handwritten text based on a large vocabulary. We describe the three main components of the system, which are preprocessing, feature extraction and recognition. In the preprocessing phase the handwritten texts are first segmented into lines. Then each line of text is normalized with respect to of skew, slant, vertical position and width. After these steps, text lines are segmented into single words. For this purpose distances between connected components are measured. Using a threshold, the distances are divided into distances within a word and distances between different words. A line of text is segmented at positions where the distances are larger than the chosen threshold. From each image representing a single word, a sequence of features is extracted. These features are input to a recognition procedure which is based on hidden Markov models. To investigate the stability of the segmentation algorithm the threshold that separates intra- and inter-word distances from each other is varied. If the threshold is small many errors are caused by over-segmentation, while for large thresholds under-segmentation errors occur. The best segmentation performance is 95.56% correctly segmented words, tested on 541 text lines containing 3899 words. Given a correct segmentation rate of 95.56%, a recognition rate of 73.45% on the word level is achieved.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132367340","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 95
An OCR system for Telugu 泰卢固语的OCR系统
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953958
A. Negi, C. Bhagvati, B. Krishna
Telugu is the language spoken by more than 100 million people of South India. Telugu has a complex orthography with a large number of distinct character shapes (estimated to be of the order of 10,000) composed of simple and compound characters formed from 16 vowels (called achchus) and 36 consonants (called hallus). We present an efficient and practical approach to Telugu OCR which limits the number of templates to be recognized to just 370, avoiding issues of classifier design for thousands of shapes or very complex glyph segmentation. A compositional approach using connected components and fringe distance template matching was tested to give a raw OCR accuracy of about 92%. Several experiments across varying fonts and resolutions showed the approach to be satisfactory.
泰卢固语是印度南部超过1亿人使用的语言。泰卢固语有一个复杂的正字法,有大量不同的字符形状(估计有10,000个),由16个元音(称为achchus)和36个辅音(称为hallus)组成的简单和复合字符组成。我们提出了一种高效实用的泰卢固语OCR方法,该方法将要识别的模板数量限制在370个,避免了为数千个形状或非常复杂的字形分割设计分类器的问题。采用连通分量和条纹距离模板匹配的组合方法进行了测试,原始OCR精度约为92%。几个不同字体和分辨率的实验表明,这种方法是令人满意的。
{"title":"An OCR system for Telugu","authors":"A. Negi, C. Bhagvati, B. Krishna","doi":"10.1109/ICDAR.2001.953958","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953958","url":null,"abstract":"Telugu is the language spoken by more than 100 million people of South India. Telugu has a complex orthography with a large number of distinct character shapes (estimated to be of the order of 10,000) composed of simple and compound characters formed from 16 vowels (called achchus) and 36 consonants (called hallus). We present an efficient and practical approach to Telugu OCR which limits the number of templates to be recognized to just 370, avoiding issues of classifier design for thousands of shapes or very complex glyph segmentation. A compositional approach using connected components and fringe distance template matching was tested to give a raw OCR accuracy of about 92%. Several experiments across varying fonts and resolutions showed the approach to be satisfactory.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"171 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132530670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 143
Discrimination of Oriental and Euramerican scripts using fractal feature 基于分形特征的东西方文字辨析
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953959
Yu Tao, Y. Tang
This paper presents a new approach based on modified fractal signatures (MFS) and modified fractal features (MFF)for the discrimination of Oriental and Euramerican scripts. These methods will be useful in the measurement and classification of patterns. MFS do not need iterative breaking or merging, and can divide a document into blocks in a single step. MFF is also used in the identification and classification of a selected set of texture images with good results. It is anticipated that this approach could be widely used to process various types of documents, even including some with high geometrical complexity.
提出了一种基于改进分形特征(MFS)和改进分形特征(MFF)的东西方文字识别新方法。这些方法将有助于模式的测量和分类。MFS不需要迭代分解或合并,并且可以在单个步骤中将文档分成块。将MFF算法应用于选定的一组纹理图像的识别和分类,取得了较好的效果。预期这种方法可以广泛用于处理各种类型的文件,甚至包括一些几何复杂度很高的文件。
{"title":"Discrimination of Oriental and Euramerican scripts using fractal feature","authors":"Yu Tao, Y. Tang","doi":"10.1109/ICDAR.2001.953959","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953959","url":null,"abstract":"This paper presents a new approach based on modified fractal signatures (MFS) and modified fractal features (MFF)for the discrimination of Oriental and Euramerican scripts. These methods will be useful in the measurement and classification of patterns. MFS do not need iterative breaking or merging, and can divide a document into blocks in a single step. MFF is also used in the identification and classification of a selected set of texture images with good results. It is anticipated that this approach could be widely used to process various types of documents, even including some with high geometrical complexity.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129993465","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Creating generic text summaries 创建通用文本摘要
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953917
Yihong Gong, Xin Liu
We propose two generic text summarization methods that create text summaries by ranking and extracting sentences from the original documents. The first method uses standard information retrieval methods to rank sentence relevances, while the second method uses the latent semantic analysis technique to identify semantically important sentences, for summary creations. Both methods strive to select sentences that are highly ranked and different from each other. This is an attempt to create a summary with a wider coverage of the document's main content and less redundancy. Performance evaluations on the two summarization methods are conducted by comparing their summarization outputs with the manual summaries generated by three independent human evaluators.
我们提出了两种通用的文本摘要方法,通过从原始文档中排序和提取句子来创建文本摘要。第一种方法使用标准的信息检索方法对句子相关度进行排序,第二种方法使用潜在语义分析技术识别语义重要的句子,以便创建摘要。这两种方法都努力选择排名较高且彼此不同的句子。这是为了创建一个更广泛地涵盖文档主要内容和减少冗余的摘要。通过将两种总结方法的总结输出与三位独立的人工评估者生成的手动总结进行比较,对两种总结方法进行绩效评估。
{"title":"Creating generic text summaries","authors":"Yihong Gong, Xin Liu","doi":"10.1109/ICDAR.2001.953917","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953917","url":null,"abstract":"We propose two generic text summarization methods that create text summaries by ranking and extracting sentences from the original documents. The first method uses standard information retrieval methods to rank sentence relevances, while the second method uses the latent semantic analysis technique to identify semantically important sentences, for summary creations. Both methods strive to select sentences that are highly ranked and different from each other. This is an attempt to create a summary with a wider coverage of the document's main content and less redundancy. Performance evaluations on the two summarization methods are conducted by comparing their summarization outputs with the manual summaries generated by three independent human evaluators.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130111131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
A class-modularity for character recognition 用于字符识别的类模块化
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953756
Il-Seok Oh, Jin-Seon Lee, C. Suen
A class-modular classifier can be characterized by two prominent features: low classifier complexity and independence of classes. While conventional character recognition systems adopting the class modularity are faithful to the first feature, they do not investigate the second one. Since a class can be handled independently of the other classes, the class-specific feature set and classifier architecture can be optimally designed for a specific class Here we propose a general framework for the class modularity that exploits fully both features and present four types of class-modular architecture. The neural network classifier is used for testing the framework A simultaneous selection of the feature set and network architecture is performed by the genetic algorithm. The effectiveness of the class-specific features and classifier architectures is confirmed by experimental results on the recognition of handwritten numerals.
类模块化分类器具有两个显著特征:分类器复杂度低和类的独立性。采用类模块化的传统字符识别系统忠实于第一个特征,而不研究第二个特征。由于类可以独立于其他类进行处理,因此可以针对特定类优化设计特定于类的特征集和分类器体系结构。在这里,我们提出了一个类模块化的通用框架,该框架充分利用了这两种特征,并提出了四种类型的类模块化体系结构。采用神经网络分类器对框架进行测试,同时采用遗传算法对特征集和网络结构进行选择。手写数字识别的实验结果证实了分类特征和分类器结构的有效性。
{"title":"A class-modularity for character recognition","authors":"Il-Seok Oh, Jin-Seon Lee, C. Suen","doi":"10.1109/ICDAR.2001.953756","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953756","url":null,"abstract":"A class-modular classifier can be characterized by two prominent features: low classifier complexity and independence of classes. While conventional character recognition systems adopting the class modularity are faithful to the first feature, they do not investigate the second one. Since a class can be handled independently of the other classes, the class-specific feature set and classifier architecture can be optimally designed for a specific class Here we propose a general framework for the class modularity that exploits fully both features and present four types of class-modular architecture. The neural network classifier is used for testing the framework A simultaneous selection of the feature set and network architecture is performed by the genetic algorithm. The effectiveness of the class-specific features and classifier architectures is confirmed by experimental results on the recognition of handwritten numerals.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130471240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Word discrimination based on bigram co-occurrences 基于重字共现的词辨别
Pub Date : 2001-09-10 DOI: 10.1109/ICDAR.2001.953773
A. El-Nasan, S. Veeramachaneni, G. Nagy
Very few pairs of English words share exactly the same letter bigrams. This linguistic property can be exploited to bring lexical context into the classification stage of a word recognition system. The lexical n-gram matches between every word in a lexicon and a subset of reference words can be precomputed. If a match function can detect matching segments of at least n-gram length from the feature representation of words, then an unknown word can be recognized by determining the subset of reference words having an n-gram match at the feature level with the unknown word. We show that with a reasonable number of reference words, bigrams represent the best compromise between the recall ability of single letters and the precision of trigrams. Our simulations indicate that using a longer reference list can compensate errors in feature extraction. The algorithm is fast enough, even with a slow processor, for human-computer interaction.
很少有英语单词对具有完全相同的字母组合。这一语言特性可用于将词汇语境引入词识别系统的分类阶段。词汇库中每个单词与参考单词子集之间的词汇n-gram匹配可以预先计算。如果匹配函数可以从单词的特征表示中检测到长度至少为n个gram的匹配片段,则可以通过确定在特征级别上与未知单词具有n个gram匹配的参考单词子集来识别未知单词。我们表明,在合理数量的参考词下,双字母代表了单字母记忆能力和三字母记忆精度之间的最佳折衷。仿真结果表明,使用较长的参考列表可以弥补特征提取中的误差。即使处理器速度较慢,该算法对于人机交互来说也足够快。
{"title":"Word discrimination based on bigram co-occurrences","authors":"A. El-Nasan, S. Veeramachaneni, G. Nagy","doi":"10.1109/ICDAR.2001.953773","DOIUrl":"https://doi.org/10.1109/ICDAR.2001.953773","url":null,"abstract":"Very few pairs of English words share exactly the same letter bigrams. This linguistic property can be exploited to bring lexical context into the classification stage of a word recognition system. The lexical n-gram matches between every word in a lexicon and a subset of reference words can be precomputed. If a match function can detect matching segments of at least n-gram length from the feature representation of words, then an unknown word can be recognized by determining the subset of reference words having an n-gram match at the feature level with the unknown word. We show that with a reasonable number of reference words, bigrams represent the best compromise between the recall ability of single letters and the precision of trigrams. Our simulations indicate that using a longer reference list can compensate errors in feature extraction. The algorithm is fast enough, even with a slow processor, for human-computer interaction.","PeriodicalId":277816,"journal":{"name":"Proceedings of Sixth International Conference on Document Analysis and Recognition","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133317152","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
Proceedings of Sixth International Conference on Document Analysis and Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1