首页 > 最新文献

Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.最新文献

英文 中文
Progress in camera-based document image analysis 基于摄像机的文档图像分析研究进展
D. Doermann, Jian Liang, Huiping Li
The increasing availability of high performance, low priced, portable digital imaging devices has created a tremendous opportunity for supplementing traditional scanning for document image acquisition. Digital cameras attached to cellular phones, PDAs, or as standalone still or video devices are highly mobile and easy to use; they can capture images of any kind of document including very thick books, historical pages too fragile to touch, and text in scenes; and they are much more versatile than desktop scanners. Should robust solutions to the analysis of documents captured with such devices become available, there is clearly a demand from many domains. Traditional scanner-based document analysis techniques provide us with a good reference and starting point, but they cannot be used directly on camera-captured images. Camera captured images can suffer from low resolution, blur, and perspective distortion, as well as complex layout and interaction of the content and background. In this paper we present a survey of application domains, technical challenges and solutions for recognizing documents captured by digital cameras. We begin by describing typical imaging devices and the imaging process. We discuss document analysis from a single camera-captured image as well as multiple frames and highlight some sample applications under development and feasible ideas for future development.
高性能、低价、便携式数字成像设备的日益普及,为补充传统的文档图像采集扫描创造了巨大的机会。数码相机连接到移动电话,pda,或作为独立的静态或视频设备是高度移动和易于使用;它们可以捕捉任何类型文件的图像,包括非常厚的书、太脆弱而不能触摸的历史页面和场景中的文本;而且它们比桌面扫描器更通用。如果对用这些设备捕获的文档进行分析的健壮的解决方案可用,那么显然会有来自许多领域的需求。传统的基于扫描仪的文档分析技术为我们提供了一个很好的参考和起点,但它们不能直接用于相机捕获的图像。相机拍摄的图像可能会出现低分辨率、模糊和视角失真,以及内容和背景的复杂布局和交互。在本文中,我们提出了应用领域的调查,技术挑战和解决方案,以识别由数码相机捕获的文件。我们首先描述典型的成像设备和成像过程。我们讨论了单帧和多帧的文档分析,并重点介绍了一些正在开发的示例应用和未来发展的可行思路。
{"title":"Progress in camera-based document image analysis","authors":"D. Doermann, Jian Liang, Huiping Li","doi":"10.1109/ICDAR.2003.1227735","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227735","url":null,"abstract":"The increasing availability of high performance, low priced, portable digital imaging devices has created a tremendous opportunity for supplementing traditional scanning for document image acquisition. Digital cameras attached to cellular phones, PDAs, or as standalone still or video devices are highly mobile and easy to use; they can capture images of any kind of document including very thick books, historical pages too fragile to touch, and text in scenes; and they are much more versatile than desktop scanners. Should robust solutions to the analysis of documents captured with such devices become available, there is clearly a demand from many domains. Traditional scanner-based document analysis techniques provide us with a good reference and starting point, but they cannot be used directly on camera-captured images. Camera captured images can suffer from low resolution, blur, and perspective distortion, as well as complex layout and interaction of the content and background. In this paper we present a survey of application domains, technical challenges and solutions for recognizing documents captured by digital cameras. We begin by describing typical imaging devices and the imaging process. We discuss document analysis from a single camera-captured image as well as multiple frames and highlight some sample applications under development and feasible ideas for future development.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131056141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 222
Confidence evaluation for combining diverse classifiers 多分类器组合置信度评价
Hongwei Hao, Cheng-Lin Liu, H. Sako
For combining classifiers at measurement level, thediverse outputs of classifiers should be transformed touniform measures that represent the confidence ofdecision, hopefully, the class probability or likelihood.This paper presents our experimental results of classifiercombination using confidence evaluation. We test threetypes of confidences: log-likelihood, exponential andsigmoid. For re-scaling the classifier outputs, we usethree scaling functions based on global normalizationand Gaussian density estimation. Experimental results inhandwritten digit recognition show that via confidenceevaluation, superior classification performance can beobtained using simple combination rules.
为了在度量水平上组合分类器,分类器的不同输出应该转换为代表决策置信度的统一度量,希望是类概率或似然。本文给出了基于置信度评价的分类器组合的实验结果。我们测试了三种类型的信心:对数似然,指数和s型。为了重新缩放分类器输出,我们使用了基于全局归一化和高斯密度估计的三个缩放函数。手写体数字识别的实验结果表明,通过置信度评价,使用简单的组合规则可以获得较好的分类性能。
{"title":"Confidence evaluation for combining diverse classifiers","authors":"Hongwei Hao, Cheng-Lin Liu, H. Sako","doi":"10.1109/ICDAR.2003.1227764","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227764","url":null,"abstract":"For combining classifiers at measurement level, thediverse outputs of classifiers should be transformed touniform measures that represent the confidence ofdecision, hopefully, the class probability or likelihood.This paper presents our experimental results of classifiercombination using confidence evaluation. We test threetypes of confidences: log-likelihood, exponential andsigmoid. For re-scaling the classifier outputs, we usethree scaling functions based on global normalizationand Gaussian density estimation. Experimental results inhandwritten digit recognition show that via confidenceevaluation, superior classification performance can beobtained using simple combination rules.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131160839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A scalable solution for integrating illustrated parts drawings into a Class IV Interactive Electronic Technical Manual 一个可扩展的解决方案,用于将插图零件图集成到IV类交互式电子技术手册中
Molly L. Boose, D. B. Shema, Lawrence S. Baum
This paper discusses a scalable solution for integrating legacy illustrated parts drawings into a Class IV Interactive Electronic Technical Manual (IETM) (1995). An IETM is an interactive electronic version of a system's technical manuals such as for a commercial airplane or a military helicopter. It contains the information a technician needs to do her job including troubleshooting, vehicle maintenance and repair procedures. A Class IV IETM is an IETM that is authored and managed directly via a database. The end-user system optimizes viewing and navigation, minimizing the need for users to browse and search through large volumes of data. The Boeing Company has hundreds of thousands of illustrated parts drawings for both commercial and military vehicles. As Boeing migrates to Class IV IETM systems, it is necessary to incorporate existing illustrated parts drawings into the new systems. Manually re-authoring the drawings to bring them up to the level of a Class IV IETM is prohibitively expensive. Our solution is to provide a batch-processing system that performs the required modifications to the raster images and automatically updates the IETM database.
本文讨论了一种可扩展的解决方案,将遗留的图解零件图集成到IV类交互式电子技术手册(IETM)(1995)中。IETM是系统技术手册的交互式电子版本,例如用于商用飞机或军用直升机。它包含了技术人员工作所需的信息,包括故障排除、车辆维护和维修程序。IV类IETM是直接通过数据库编写和管理的IETM。最终用户系统优化了查看和导航,最大限度地减少了用户浏览和搜索大量数据的需要。波音公司有成千上万的商用和军用车辆的图解零件图。随着波音向IV类IETM系统的迁移,有必要将现有的图解部件图纸合并到新系统中。手动重新编写图纸以使其达到IV类IETM的水平是非常昂贵的。我们的解决方案是提供一个批处理系统,该系统对光栅图像执行所需的修改并自动更新IETM数据库。
{"title":"A scalable solution for integrating illustrated parts drawings into a Class IV Interactive Electronic Technical Manual","authors":"Molly L. Boose, D. B. Shema, Lawrence S. Baum","doi":"10.1109/ICDAR.2003.1227679","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227679","url":null,"abstract":"This paper discusses a scalable solution for integrating legacy illustrated parts drawings into a Class IV Interactive Electronic Technical Manual (IETM) (1995). An IETM is an interactive electronic version of a system's technical manuals such as for a commercial airplane or a military helicopter. It contains the information a technician needs to do her job including troubleshooting, vehicle maintenance and repair procedures. A Class IV IETM is an IETM that is authored and managed directly via a database. The end-user system optimizes viewing and navigation, minimizing the need for users to browse and search through large volumes of data. The Boeing Company has hundreds of thousands of illustrated parts drawings for both commercial and military vehicles. As Boeing migrates to Class IV IETM systems, it is necessary to incorporate existing illustrated parts drawings into the new systems. Manually re-authoring the drawings to bring them up to the level of a Class IV IETM is prohibitively expensive. Our solution is to provide a batch-processing system that performs the required modifications to the raster images and automatically updates the IETM database.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129323292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
A character recognizer for Turkish language 土耳其语字符识别器
Sait Ulas Korkmaz, G. Kirçiçegi, Y. Akinci, V. Atalay
This paper presents particularly a contextual postprocessing subsystem for a Turkish machine printedcharacter recognition system. The contextual postprocessing subsystem is based on positional binary 3-gram statistics for Turkish language, an error correctorparser and a lexicon, which contains root words and theinflected forms of the root words. Error corrector parseris used for correcting CR alternatives using TurkishMorphology.
本文详细介绍了一个土耳其语机器印刷字符识别系统的上下文后处理子系统。上下文后处理子系统基于土耳其语的位置二进制3-gram统计,一个错误校正分析器和一个词典,其中包含词根和词根的屈折形式。错误校正分析器用于使用TurkishMorphology纠正CR选项。
{"title":"A character recognizer for Turkish language","authors":"Sait Ulas Korkmaz, G. Kirçiçegi, Y. Akinci, V. Atalay","doi":"10.1109/ICDAR.2003.1227855","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227855","url":null,"abstract":"This paper presents particularly a contextual postprocessing subsystem for a Turkish machine printedcharacter recognition system. The contextual postprocessing subsystem is based on positional binary 3-gram statistics for Turkish language, an error correctorparser and a lexicon, which contains root words and theinflected forms of the root words. Error corrector parseris used for correcting CR alternatives using TurkishMorphology.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128204457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Reference line extraction from form documents with complicated backgrounds 从复杂背景的表单文档中提取参考线
Dihua Xi, Seong-Whan Lee
Form document analysis is one of the most essential tasksin document analysis and recognition. One of the most fundamentaland crucial tasks is the extraction of the referencelines which are contained in almost all form documents.This paper presents an efficient methodology for the complicatedgrey-level form image processing. We construct anon-orthogonal wavelet with adjustable rectangle supportsand offer algorithms for the extraction of the reference linesbased on the strip growth method using the multiresolutionwavelet sub images. We have compared this system with thepopular Hough transform (HT) based and the novel orthogonalwavelet based methods. As shown in the experiments,the proposed algorithmdemonstrates high performance andfast speed for the complicated form images. This system isalso effective for the form images with slight skew.
格式文件分析是文件分析与识别中最重要的任务之一。最基本和最关键的任务之一是提取几乎所有表单文件中包含的参考资料。本文提出了一种处理复杂灰度形式图像的有效方法。我们构造了具有可调矩形支撑的非正交小波,并利用多分辨率小波子图像提出了基于条带生长法提取参考线的算法。我们将该系统与流行的基于霍夫变换(HT)和新的基于正交小波的方法进行了比较。实验结果表明,该算法在处理复杂形状图像时具有较高的性能和速度。该系统对于有轻微倾斜的表单图像也很有效。
{"title":"Reference line extraction from form documents with complicated backgrounds","authors":"Dihua Xi, Seong-Whan Lee","doi":"10.1109/ICDAR.2003.1227823","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227823","url":null,"abstract":"Form document analysis is one of the most essential tasksin document analysis and recognition. One of the most fundamentaland crucial tasks is the extraction of the referencelines which are contained in almost all form documents.This paper presents an efficient methodology for the complicatedgrey-level form image processing. We construct anon-orthogonal wavelet with adjustable rectangle supportsand offer algorithms for the extraction of the reference linesbased on the strip growth method using the multiresolutionwavelet sub images. We have compared this system with thepopular Hough transform (HT) based and the novel orthogonalwavelet based methods. As shown in the experiments,the proposed algorithmdemonstrates high performance andfast speed for the complicated form images. This system isalso effective for the form images with slight skew.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127173645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
String extraction from color airline coupon image using statistical approach 基于统计方法的彩色航空优惠券图像字符串提取
Yi Li, Zhiyan Wang, Haizan Zeng
A novel technique is presented in this paper to extract strings in color images of both business settlement plan (BSP) and non-BSP airline coupon. The essential concept is to remove non-text pixels from complex coupon images, rather than extract strings directly. First we transfer color images from RGB to HSV space, which is approximate uniformed, and then remove the black component of images using the property of HSV space. A statistical approach called principal components analysis (PCA) is applied to extract strings by removing the background decorative pattern based on priori environment. Finally, a method to validate and improve performance is present.
本文提出了一种从商业结算计划(BSP)和非BSP航空优惠券彩色图像中提取字符串的新方法。其基本概念是从复杂的优惠券图像中去除非文本像素,而不是直接提取字符串。首先将彩色图像从RGB转移到近似均匀的HSV空间,然后利用HSV空间的特性去除图像中的黑色成分。采用主成分分析(PCA)的统计方法,根据先验环境去除背景装饰图案,提取字符串。最后,提出了一种验证和改进性能的方法。
{"title":"String extraction from color airline coupon image using statistical approach","authors":"Yi Li, Zhiyan Wang, Haizan Zeng","doi":"10.1109/ICDAR.2003.1227675","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227675","url":null,"abstract":"A novel technique is presented in this paper to extract strings in color images of both business settlement plan (BSP) and non-BSP airline coupon. The essential concept is to remove non-text pixels from complex coupon images, rather than extract strings directly. First we transfer color images from RGB to HSV space, which is approximate uniformed, and then remove the black component of images using the property of HSV space. A statistical approach called principal components analysis (PCA) is applied to extract strings by removing the background decorative pattern based on priori environment. Finally, a method to validate and improve performance is present.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114296911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Text identification in noisy document images using Markov random model 基于马尔可夫随机模型的噪声文档图像文本识别
Yefeng Zheng, Huiping Li, D. Doermann
In this paper we address the problem of the identification of text from noisy documents. We segment and identify handwriting from machine printed text because 1) handwriting in a document often indicates corrections, additions or other supplemental information that should be treated differently from the main body or body content, and 2) the segmentation and recognition techniques for machine printed text and handwriting are significantly different. Our novelty is that we treat noise as a separate class and model noise based on selected features. Trained Fisher classifiers are used to identify machine printed text and handwriting from noise. We further exploit context to refine the classification. A Markov random field (MRF) based approach is used to model the geometrical structure of the printed text, handwriting and noise to rectify the mis-classification. Experimental results show our approach is promising and robust, and can significantly improve the page segmentation results in noise documents.
在本文中,我们解决了从噪声文档中识别文本的问题。我们从机器打印文本中分割和识别手写,因为1)文档中的手写通常表示更正、添加或其他补充信息,这些信息应该与主体或主体内容区别对待,2)机器打印文本和手写的分割和识别技术有很大不同。我们的新颖之处在于,我们将噪声视为一个单独的类别,并基于选定的特征对噪声进行建模。经过训练的Fisher分类器用于从噪声中识别机器打印的文本和手写。我们进一步利用上下文来改进分类。基于马尔可夫随机场(MRF)的方法对打印文本、手写和噪声的几何结构进行建模,以纠正错误分类。实验结果表明,该方法具有良好的鲁棒性,可以显著改善噪声文档中的页面分割效果。
{"title":"Text identification in noisy document images using Markov random model","authors":"Yefeng Zheng, Huiping Li, D. Doermann","doi":"10.1109/ICDAR.2003.1227734","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227734","url":null,"abstract":"In this paper we address the problem of the identification of text from noisy documents. We segment and identify handwriting from machine printed text because 1) handwriting in a document often indicates corrections, additions or other supplemental information that should be treated differently from the main body or body content, and 2) the segmentation and recognition techniques for machine printed text and handwriting are significantly different. Our novelty is that we treat noise as a separate class and model noise based on selected features. Trained Fisher classifiers are used to identify machine printed text and handwriting from noise. We further exploit context to refine the classification. A Markov random field (MRF) based approach is used to model the geometrical structure of the printed text, handwriting and noise to rectify the mis-classification. Experimental results show our approach is promising and robust, and can significantly improve the page segmentation results in noise documents.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121871123","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Detection of text marks on moving vehicles 移动车辆上的文本标记检测
R. Kasturi
Vehicle text marks are unique features which are useful for identifying vehicles in video surveillance applications. We propose a method for finding such text marks. An existing text detection algorithm is modified such that detection is increased and made more robust to outdoor conditions. False alarm is reduced by introducing a binary image test which remove detections that are not likely to be caused by text. The method is tested on a captured video of a typical street scene.
车辆文本标记是视频监控应用中用于识别车辆的独特功能。我们提出了一种查找此类文本标记的方法。对现有的文本检测算法进行了改进,从而增加了检测并使其对室外条件更具鲁棒性。通过引入二值图像测试来减少误报,该测试可以去除不太可能由文本引起的检测。该方法在一个典型街景的视频上进行了测试。
{"title":"Detection of text marks on moving vehicles","authors":"R. Kasturi","doi":"10.1109/ICDAR.2003.1227696","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227696","url":null,"abstract":"Vehicle text marks are unique features which are useful for identifying vehicles in video surveillance applications. We propose a method for finding such text marks. An existing text detection algorithm is modified such that detection is increased and made more robust to outdoor conditions. False alarm is reduced by introducing a binary image test which remove detections that are not likely to be caused by text. The method is tested on a captured video of a typical street scene.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"250 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121880873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Individuality of numerals 数字的个性
S. Srihari, C. Tomai, Bin Zhang, Sangjik Lee
The analysis of handwritten documents from the view-pointof determining their writership has great bearing onthe criminal justice system. In many cases, only a limitedamount of handwriting is available and sometimes it consistsof only numerals. Using a large number of handwrittennumeral images extracted from about 3000 samples writtenby 1000 writers, a study of the individuality of numerals foridentification/verification purposes was conducted. The individualityof numerals was studied using cluster analysis.Numerals discriminability was measured for writer verification.The study shows that some numerals present a higherdiscriminatory power and that their performances for theverification/identification tasks are very different.
从笔迹鉴定的角度对手写文书进行分析,对刑事司法制度有着重要的影响。在许多情况下,只有有限数量的笔迹可用,有时它只包含数字。利用从1000位写作者所写的约3000个样本中提取的大量手写数字图像,对数字的个性进行了识别/验证目的的研究。采用聚类分析对数字的个性进行了研究。测量了数字的可判别性,以供作者验证。研究表明,一些数字具有较高的歧视性,它们在验证/识别任务中的表现差异很大。
{"title":"Individuality of numerals","authors":"S. Srihari, C. Tomai, Bin Zhang, Sangjik Lee","doi":"10.1109/ICDAR.2003.1227826","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227826","url":null,"abstract":"The analysis of handwritten documents from the view-pointof determining their writership has great bearing onthe criminal justice system. In many cases, only a limitedamount of handwriting is available and sometimes it consistsof only numerals. Using a large number of handwrittennumeral images extracted from about 3000 samples writtenby 1000 writers, a study of the individuality of numerals foridentification/verification purposes was conducted. The individualityof numerals was studied using cluster analysis.Numerals discriminability was measured for writer verification.The study shows that some numerals present a higherdiscriminatory power and that their performances for theverification/identification tasks are very different.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"2012 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129982815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Unsupervised feature selection using multi-objective genetic algorithms for handwritten word recognition 基于多目标遗传算法的无监督特征选择手写体单词识别
M. Morita, R. Sabourin, F. Bortolozzi, C. Y. Suen
In this paper a methodology for feature selection in unsupervisedlearning is proposed. It makes use of a multi-objectivegenetic algorithm where the minimization of thenumber of features and a validity index that measures thequality of clusters have been used to guide the search towardsthe more discriminant features and the best numberof clusters. The proposed strategy is evaluated usingtwo synthetic data sets and then it is applied to handwrittenmonth word recognition. Comprehensive experimentsdemonstrate the feasibility and efficiency of the proposedmethodology.
本文提出了一种无监督学习中特征选择的方法。它利用多目标遗传算法,其中特征数量的最小化和衡量聚类质量的有效性指标被用来指导搜索更有区别的特征和最佳数量的聚类。使用两个合成数据集对所提出的策略进行了评估,然后将其应用于手写月词识别。综合实验证明了该方法的可行性和有效性。
{"title":"Unsupervised feature selection using multi-objective genetic algorithms for handwritten word recognition","authors":"M. Morita, R. Sabourin, F. Bortolozzi, C. Y. Suen","doi":"10.1109/ICDAR.2003.1227746","DOIUrl":"https://doi.org/10.1109/ICDAR.2003.1227746","url":null,"abstract":"In this paper a methodology for feature selection in unsupervisedlearning is proposed. It makes use of a multi-objectivegenetic algorithm where the minimization of thenumber of features and a validity index that measures thequality of clusters have been used to guide the search towardsthe more discriminant features and the best numberof clusters. The proposed strategy is evaluated usingtwo synthetic data sets and then it is applied to handwrittenmonth word recognition. Comprehensive experimentsdemonstrate the feasibility and efficiency of the proposedmethodology.","PeriodicalId":249193,"journal":{"name":"Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130088080","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1