首页 > 最新文献

2008 The Eighth IAPR International Workshop on Document Analysis Systems最新文献

英文 中文
A Hybrid System for Text Detection in Video Frames 视频帧文本检测混合系统
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.72
M. Anthimopoulos, B. Gatos, I. Pratikakis
This paper proposes a hybrid system for text detection in video frames. The system consists of two main stages. In the first stage text regions are detected based on the edge map of the image leading in a high recall rate with minimum computation requirements. In the sequel, a refinement stage uses an SVM classifier trained on features obtained by a new local binary pattern based operator which results in diminishing false alarms. Experimental results show the overall performance of the system that proves the discriminating ability of the proposed feature set.
本文提出了一种用于视频帧中文本检测的混合系统。该系统由两个主要阶段组成。在第一阶段,根据图像的边缘图检测文字区域,从而以最低的计算要求获得较高的召回率。在后一阶段,细化阶段使用 SVM 分类器,该分类器根据基于新的局部二进制模式算子获得的特征进行训练,从而减少误报。实验结果显示了系统的整体性能,证明了所提出的特征集的辨别能力。
{"title":"A Hybrid System for Text Detection in Video Frames","authors":"M. Anthimopoulos, B. Gatos, I. Pratikakis","doi":"10.1109/DAS.2008.72","DOIUrl":"https://doi.org/10.1109/DAS.2008.72","url":null,"abstract":"This paper proposes a hybrid system for text detection in video frames. The system consists of two main stages. In the first stage text regions are detected based on the edge map of the image leading in a high recall rate with minimum computation requirements. In the sequel, a refinement stage uses an SVM classifier trained on features obtained by a new local binary pattern based operator which results in diminishing false alarms. Experimental results show the overall performance of the system that proves the discriminating ability of the proposed feature set.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131192132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Kanji Character Detection from Complex Real Scene Images based on Character Properties 基于字符属性的复杂真实场景图像汉字字符检测
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.34
Lianli Xu, H. Nagayoshi, H. Sako
Character recognition in complex real scene images is a very challenging undertaking. The most popular approach is to segment the text area using some extra pre-knowledge, such as "characters are in a signboard'', etc. This approach makes it possible to construct a very time-consuming method, but generality is still a problem. In this paper, we propose a more general method by utilizing only character features. Our algorithm consists of five steps: pre-processing to extract connected components, initial classification using primitive rules, strong classification using AdaBoost, Markov random field (MRF) clustering to combine connected components with similar properties, and post-processing using optical character recognition (OCR) results. The results of experiments using 11 images containing 1691 characters (including characters in bad condition) indicated the effectiveness of the proposed system, namely, that 52.9% of characters were extracted correctly with 625 noise components extracted as characters.
复杂的真实场景图像中的字符识别是一项非常具有挑战性的工作。最流行的方法是使用一些额外的预先知识来分割文本区域,例如“字符位于广告牌中”等。这种方法使得构造一个非常耗时的方法成为可能,但是通用性仍然是一个问题。在本文中,我们提出了一种更通用的方法,即仅利用字符特征。我们的算法包括五个步骤:预处理提取连接成分,使用原始规则进行初始分类,使用AdaBoost进行强分类,使用马尔可夫随机场(MRF)聚类对具有相似属性的连接成分进行组合,以及使用光学字符识别(OCR)结果进行后处理。对11幅包含1691个字符(含不良字符)的图像进行了实验,结果表明了该系统的有效性,即以625个噪声分量作为字符提取的字符正确率为52.9%。
{"title":"Kanji Character Detection from Complex Real Scene Images based on Character Properties","authors":"Lianli Xu, H. Nagayoshi, H. Sako","doi":"10.1109/DAS.2008.34","DOIUrl":"https://doi.org/10.1109/DAS.2008.34","url":null,"abstract":"Character recognition in complex real scene images is a very challenging undertaking. The most popular approach is to segment the text area using some extra pre-knowledge, such as \"characters are in a signboard'', etc. This approach makes it possible to construct a very time-consuming method, but generality is still a problem. In this paper, we propose a more general method by utilizing only character features. Our algorithm consists of five steps: pre-processing to extract connected components, initial classification using primitive rules, strong classification using AdaBoost, Markov random field (MRF) clustering to combine connected components with similar properties, and post-processing using optical character recognition (OCR) results. The results of experiments using 11 images containing 1691 characters (including characters in bad condition) indicated the effectiveness of the proposed system, namely, that 52.9% of characters were extracted correctly with 625 noise components extracted as characters.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134533071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
End-to-End Trainable Thai OCR System Using Hidden Markov Models 基于隐马尔可夫模型的端到端可训练泰国OCR系统
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.76
K. Krstovski, Ehry MacRostie, R. Prasad, P. Natarajan
In this paper we present an end-to-end trainable optical character recognition (OCR) system for recognizing machine-printed text in Thai documents. The end-to-end OCR system is based on a script-independent methodology using hidden Markov models. Our system provides an integrated workflow beginning with annotation and transcription of training images to performing OCR on new images with models trained on transcribed training images. The efficacy of our end-to-end OCR system is demonstrated by rapidly configuring our OCR engine for the Thai script. We present experimental results on Thai documents to highlight the specific challenges posed by the Thai script and analyze the recognition performance as a function of amount of training data.
在本文中,我们提出了一个端到端可训练的光学字符识别(OCR)系统,用于识别机器打印文本的泰国文档。端到端OCR系统基于使用隐马尔可夫模型的脚本独立方法。我们的系统提供了一个集成的工作流程,从对训练图像的注释和转录开始,到使用在转录的训练图像上训练的模型对新图像执行OCR。通过为泰语脚本快速配置OCR引擎,我们的端到端OCR系统的有效性得到了验证。我们展示了泰语文档的实验结果,以突出泰语脚本带来的具体挑战,并分析了识别性能作为训练数据量的函数。
{"title":"End-to-End Trainable Thai OCR System Using Hidden Markov Models","authors":"K. Krstovski, Ehry MacRostie, R. Prasad, P. Natarajan","doi":"10.1109/DAS.2008.76","DOIUrl":"https://doi.org/10.1109/DAS.2008.76","url":null,"abstract":"In this paper we present an end-to-end trainable optical character recognition (OCR) system for recognizing machine-printed text in Thai documents. The end-to-end OCR system is based on a script-independent methodology using hidden Markov models. Our system provides an integrated workflow beginning with annotation and transcription of training images to performing OCR on new images with models trained on transcribed training images. The efficacy of our end-to-end OCR system is demonstrated by rapidly configuring our OCR engine for the Thai script. We present experimental results on Thai documents to highlight the specific challenges posed by the Thai script and analyze the recognition performance as a function of amount of training data.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132395641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Robust System to Detect and Localize Texts in Natural Scene Images 一种鲁棒的自然场景图像文本检测与定位系统
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.42
Yi-Feng Pan, Xinwen Hou, Cheng-Lin Liu
In this paper, we present a robust system to accurately detect and localize texts in natural scene images. For text detection, a region-based method utilizing multiple features and cascade AdaBoost classifier is adopted. For text localization, a window grouping method integrating text line competition analysis is used to generate text lines. Then within each text line, local binarization is used to extract candidate connected components (CCs) and non-text CCs are filtered out by Markov Random Fields (MRF) model, through which text line can be localized accurately. Experiments on the public benchmark ICDAR 2003 Robust Reading and Text Locating Dataset show that our system is comparable to the best existing methods both in accuracy and speed.
在本文中,我们提出了一个鲁棒的系统来准确地检测和定位自然场景图像中的文本。文本检测采用基于区域的多特征级联AdaBoost分类器。在文本定位方面,采用结合文本行竞争分析的窗口分组方法生成文本行。然后在每条文本行内,采用局部二值化方法提取候选连通分量(cc),利用马尔科夫随机场(MRF)模型过滤非文本连通分量,实现文本行精确定位。在公共基准ICDAR 2003鲁棒阅读和文本定位数据集上的实验表明,我们的系统在准确性和速度上与现有的最佳方法相当。
{"title":"A Robust System to Detect and Localize Texts in Natural Scene Images","authors":"Yi-Feng Pan, Xinwen Hou, Cheng-Lin Liu","doi":"10.1109/DAS.2008.42","DOIUrl":"https://doi.org/10.1109/DAS.2008.42","url":null,"abstract":"In this paper, we present a robust system to accurately detect and localize texts in natural scene images. For text detection, a region-based method utilizing multiple features and cascade AdaBoost classifier is adopted. For text localization, a window grouping method integrating text line competition analysis is used to generate text lines. Then within each text line, local binarization is used to extract candidate connected components (CCs) and non-text CCs are filtered out by Markov Random Fields (MRF) model, through which text line can be localized accurately. Experiments on the public benchmark ICDAR 2003 Robust Reading and Text Locating Dataset show that our system is comparable to the best existing methods both in accuracy and speed.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122160738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 98
Towards Whole-Book Recognition 迈向全本识别
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.50
Pingping Xiu, H. Baird
We describe experimental results for unsupervised recognition of the textual contents of book-images using fully automatic mutual-entropy-based model adaptation. Each experiment starts with approximate iconic and linguistic models---derived from (generally errorful) OCR results and (generally incomplete) dictionaries---and then runs a fully automatic adaptation algorithm which, guided entirely by evidence internal to the test set, attempts to correct the models for improved accuracy. The iconic model describes image formation and determines the behavior of a character-image classifier. The linguistic model describes word-occurrence probabilities. Our adaptation algorithm detects disagreements between the models by analyzing mutual entropy between (1) the a posteriori probability distribution of character classes (the recognition results from image classification alone), and (2) the a posteriori probability distribution of word classes (the recognition results from image classification combined with linguistic constraints). Disagreements identify candidates for automatic model corrections. We report experiments on 40 textlines in which word error rates fall monotonicaly with passage lengths. We also report experiments on an enhanced algorithm which can cope with character-segmentation errors (a single split, or a single merge, per word). In order to scale up experiments, soon, to whole book images, we have revised data structures and implemented speed enhancements. For this algorithm, we report results on three increasingly long passage lengths: (a) one full page, (b) five pages, and (b) ten pages. We observe that error rates on long words fall monotonically with passage lengths.
我们描述了使用全自动基于互熵的模型自适应对图书图像文本内容进行无监督识别的实验结果。每个实验都从近似的符号和语言模型开始,这些模型来自(通常错误的)OCR结果和(通常不完整的)字典,然后运行一个全自动的自适应算法,该算法完全由测试集内部的证据指导,试图纠正模型以提高准确性。图标模型描述图像的形成,并决定字符图像分类器的行为。语言模型描述单词出现的概率。我们的自适应算法通过分析(1)字符类的后验概率分布(仅图像分类的识别结果)和(2)词类的后验概率分布(结合语言约束的图像分类的识别结果)之间的互熵来检测模型之间的分歧。分歧确定自动模型修正的候选者。我们报告了40个文本行的实验,其中单词错误率随段落长度单调下降。我们还报告了一种增强算法的实验,该算法可以处理字符分割错误(每个单词的单个分割或单个合并)。为了将实验扩展到整本书图像,我们修改了数据结构并实现了速度增强。对于这个算法,我们报告了三个越来越长的段落长度的结果:(a)一整页,(b)五页,(b)十页。我们观察到长词的错误率随着段落长度单调下降。
{"title":"Towards Whole-Book Recognition","authors":"Pingping Xiu, H. Baird","doi":"10.1109/DAS.2008.50","DOIUrl":"https://doi.org/10.1109/DAS.2008.50","url":null,"abstract":"We describe experimental results for unsupervised recognition of the textual contents of book-images using fully automatic mutual-entropy-based model adaptation. Each experiment starts with approximate iconic and linguistic models---derived from (generally errorful) OCR results and (generally incomplete) dictionaries---and then runs a fully automatic adaptation algorithm which, guided entirely by evidence internal to the test set, attempts to correct the models for improved accuracy. The iconic model describes image formation and determines the behavior of a character-image classifier. The linguistic model describes word-occurrence probabilities. Our adaptation algorithm detects disagreements between the models by analyzing mutual entropy between (1) the a posteriori probability distribution of character classes (the recognition results from image classification alone), and (2) the a posteriori probability distribution of word classes (the recognition results from image classification combined with linguistic constraints). Disagreements identify candidates for automatic model corrections. We report experiments on 40 textlines in which word error rates fall monotonicaly with passage lengths. We also report experiments on an enhanced algorithm which can cope with character-segmentation errors (a single split, or a single merge, per word). In order to scale up experiments, soon, to whole book images, we have revised data structures and implemented speed enhancements. For this algorithm, we report results on three increasingly long passage lengths: (a) one full page, (b) five pages, and (b) ten pages. We observe that error rates on long words fall monotonically with passage lengths.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122355633","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Object Extraction from Colour Cadastral Maps 彩色地籍图中物体的提取
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.9
R. Raveaux, J. Burie, J. Ogier
In this paper, an object extraction method from ancient colour maps is proposed. It consists on the localization of quarters inside a given cadastral map. The colour aspect is exploited thanks to a colour restoration algorithm and the selection of a relevant hybrid colour model. Objects composing the map are located using a multi-components gradient. To identify quarters, a peeling the onion method is adopted. This selective method starts by separated text and graphics. On the graphic layer, a connected component analysis is carried out through the use of a neighbourhood graph. This graph is smartly pruned to consider only significant areas. Consequently, the quarter boundaries are found using a snake which is a computer-generated curve that moves within an image to fit a given object. The performance of our method is measured up in two steps: Firstly, the colour space selection is assessed according to the colour distinction capacity while being robust to variations/noise then the automatic extraction approach is compared to the user ground truth. Results show the good behaviour of the whole system.
提出了一种从古代彩色地图中提取目标的方法。它包括在给定的地籍地图内的区域定位。通过颜色恢复算法和相关混合颜色模型的选择,对颜色方面进行了开发。组成地图的对象使用多分量梯度进行定位。为了识别四分之一,采用了剥洋葱的方法。这种选择性方法首先将文本和图形分开。在图形层,通过使用邻域图进行连通成分分析。这张图被巧妙地修剪为只考虑重要区域。因此,使用蛇来找到四分之一边界,蛇是计算机生成的曲线,在图像中移动以适应给定对象。首先,根据颜色区分能力对颜色空间选择进行评估,同时对变化/噪声具有鲁棒性,然后将自动提取方法与用户地面真值进行比较。结果表明,整个系统性能良好。
{"title":"Object Extraction from Colour Cadastral Maps","authors":"R. Raveaux, J. Burie, J. Ogier","doi":"10.1109/DAS.2008.9","DOIUrl":"https://doi.org/10.1109/DAS.2008.9","url":null,"abstract":"In this paper, an object extraction method from ancient colour maps is proposed. It consists on the localization of quarters inside a given cadastral map. The colour aspect is exploited thanks to a colour restoration algorithm and the selection of a relevant hybrid colour model. Objects composing the map are located using a multi-components gradient. To identify quarters, a peeling the onion method is adopted. This selective method starts by separated text and graphics. On the graphic layer, a connected component analysis is carried out through the use of a neighbourhood graph. This graph is smartly pruned to consider only significant areas. Consequently, the quarter boundaries are found using a snake which is a computer-generated curve that moves within an image to fit a given object. The performance of our method is measured up in two steps: Firstly, the colour space selection is assessed according to the colour distinction capacity while being robust to variations/noise then the automatic extraction approach is compared to the user ground truth. Results show the good behaviour of the whole system.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125156422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Writer-Dependent Recognition of Handwritten Whiteboard Notes in Smart Meeting Room Environments 智能会议室环境下手写白板笔记的写作者依赖识别
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.8
M. Liwicki, A. Schlapbach, H. Bunke
In this paper we present a writer-dependent handwriting recognition system based on hidden Markov models (HMMs). This system, which has been developed in the context of research on smart meeting rooms, operates in two stages. First, a Gaussian mixture model (GMM)-based writer identification system developed for smart meeting rooms identifies the person writing on the whiteboard. Then a recognition system adapted to the individual writer is applied. Two different methods for obtaining writer-dependent recognizers are proposed. The first method uses the available writer-specific data to train an individual recognition system for each writer from scratch, while the second method takes a writer-independent recognizer and adapts it with the data from the considered writer. The experiments have been performed on the IAM-OnDB. In the first stage,the writer identification system produces a perfect identification rate. In the second stage, the writer-specific recognition system gets significantly better recognition results, compared to the writer-independent recognizer. The final word recognition rate on the IAM-OnDB-t1 benchmark task is close to 80 %.
本文提出了一种基于隐马尔可夫模型(hmm)的手写识别系统。该系统是在智能会议室研究的背景下开发的,分两个阶段运行。首先,开发了一种基于高斯混合模型(GMM)的智能会议室写作者识别系统,用于识别在白板上写字的人。然后应用了一种适合于个体写作者的识别系统。提出了两种不同的方法来获得依赖于书写器的识别器。第一种方法使用可用的特定于编写器的数据从头开始为每个编写器训练单独的识别系统,而第二种方法采用与编写器无关的识别器,并使用来自所考虑的编写器的数据对其进行调整。实验在IAM-OnDB上进行。在第一阶段,作者识别系统产生了完美的识别率。在第二阶段,与独立于写作者的识别器相比,特定于写作者的识别系统获得了明显更好的识别结果。IAM-OnDB-t1基准任务的最终单词识别率接近80%。
{"title":"Writer-Dependent Recognition of Handwritten Whiteboard Notes in Smart Meeting Room Environments","authors":"M. Liwicki, A. Schlapbach, H. Bunke","doi":"10.1109/DAS.2008.8","DOIUrl":"https://doi.org/10.1109/DAS.2008.8","url":null,"abstract":"In this paper we present a writer-dependent handwriting recognition system based on hidden Markov models (HMMs). This system, which has been developed in the context of research on smart meeting rooms, operates in two stages. First, a Gaussian mixture model (GMM)-based writer identification system developed for smart meeting rooms identifies the person writing on the whiteboard. Then a recognition system adapted to the individual writer is applied. Two different methods for obtaining writer-dependent recognizers are proposed. The first method uses the available writer-specific data to train an individual recognition system for each writer from scratch, while the second method takes a writer-independent recognizer and adapts it with the data from the considered writer. The experiments have been performed on the IAM-OnDB. In the first stage,the writer identification system produces a perfect identification rate. In the second stage, the writer-specific recognition system gets significantly better recognition results, compared to the writer-independent recognizer. The final word recognition rate on the IAM-OnDB-t1 benchmark task is close to 80 %.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"190 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124225029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Writer Verification of Arabic Handwriting 阿拉伯笔迹的作家验证
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.81
S. Srihari, G. R. Ball
Expanding on an earlier study to objectively validate the hypothesis that handwriting is individualistic, we extend the study to include handwriting in the Arabic script. Handwriting samples from twelve native speakers of Arabic were obtained. Analyzing differences in handwriting was done by using computer algorithms for extracting features from scanned images of handwriting. Attributes characteristic of the handwriting were obtained, e.g., line separation, slant, character shapes, etc. These attributes, which are a subset of attributes used by forensic document examiners (FDEs), were used to quantitatively establish individuality by using machine learning approaches. Using global attributes of handwriting, the ability to determine the writer with a high degree of confidence was established. The work is a step towards providing scientific support for admitting handwriting evidence in court.
在早期研究的基础上,为了客观地验证笔迹是个人主义的假设,我们将研究扩展到包括阿拉伯文字的笔迹。从12个以阿拉伯语为母语的人那里获得了笔迹样本。使用计算机算法从扫描的笔迹图像中提取特征来分析笔迹的差异。获得了笔迹的属性特征,如行距、斜度、字符形状等。这些属性是法医文件审查员(fde)使用的属性的子集,用于通过使用机器学习方法定量地建立个性。利用笔迹的全局属性,建立了高度自信地确定作者的能力。这项工作是向在法庭上承认笔迹证据提供科学支持迈出的一步。
{"title":"Writer Verification of Arabic Handwriting","authors":"S. Srihari, G. R. Ball","doi":"10.1109/DAS.2008.81","DOIUrl":"https://doi.org/10.1109/DAS.2008.81","url":null,"abstract":"Expanding on an earlier study to objectively validate the hypothesis that handwriting is individualistic, we extend the study to include handwriting in the Arabic script. Handwriting samples from twelve native speakers of Arabic were obtained. Analyzing differences in handwriting was done by using computer algorithms for extracting features from scanned images of handwriting. Attributes characteristic of the handwriting were obtained, e.g., line separation, slant, character shapes, etc. These attributes, which are a subset of attributes used by forensic document examiners (FDEs), were used to quantitatively establish individuality by using machine learning approaches. Using global attributes of handwriting, the ability to determine the writer with a high degree of confidence was established. The work is a step towards providing scientific support for admitting handwriting evidence in court.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123648246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Automated OCR Ground Truth Generation 自动OCR地面真相生成
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.59
J. V. Beusekom, F. Shafait, T. Breuel
Most optical character recognition (OCR) systems need to be trained and tested on the symbols that are to be recognized. Therefore, ground truth data is needed. This data consists of character images together with their ASCII code. Among the approaches for generating ground truth of real world data, one promising technique is to use electronic version of the scanned documents. Using an alignment method, the character bounding boxes extracted from the electronic document are matched to the scanned image. Current alignment methods are not robust to different similarity transforms. They also need calibration to deal with non-linear local distortions introduced by the printing/scanning process. In this paper we present a significant improvement over existing methods, allowing to skip the calibration step and having a more accurate alignment, under all similarity transforms. Our method finds a robust and pixel accurate scanner independent alignment of the scanned image with the electronic document, allowing the extraction of accurate ground truth character information. The accuracy of the alignment is demonstrated using documents from the UW3 dataset. The results show that the mean distance between the estimated and the ground truth character bounding box position is less than one pixel.
大多数光学字符识别(OCR)系统需要对要识别的符号进行训练和测试。因此,需要地面真值数据。该数据由字符图像及其ASCII码组成。在生成真实世界数据的基础真值的方法中,一种很有前途的技术是使用扫描文档的电子版本。采用对齐方法,将从电子文档中提取的字符边界框与扫描图像进行匹配。现有的对齐方法对不同的相似性变换不具有鲁棒性。它们还需要校准以处理印刷/扫描过程中引入的非线性局部扭曲。在本文中,我们提出了对现有方法的重大改进,允许在所有相似变换下跳过校准步骤并具有更准确的对准。我们的方法找到了一种鲁棒和像素精度的扫描仪独立的扫描图像与电子文档对齐,允许提取准确的地面真实特征信息。使用来自UW3数据集的文档演示了对齐的准确性。结果表明,估计的字符边界框位置与地面真实字符边界框位置的平均距离小于1个像素。
{"title":"Automated OCR Ground Truth Generation","authors":"J. V. Beusekom, F. Shafait, T. Breuel","doi":"10.1109/DAS.2008.59","DOIUrl":"https://doi.org/10.1109/DAS.2008.59","url":null,"abstract":"Most optical character recognition (OCR) systems need to be trained and tested on the symbols that are to be recognized. Therefore, ground truth data is needed. This data consists of character images together with their ASCII code. Among the approaches for generating ground truth of real world data, one promising technique is to use electronic version of the scanned documents. Using an alignment method, the character bounding boxes extracted from the electronic document are matched to the scanned image. Current alignment methods are not robust to different similarity transforms. They also need calibration to deal with non-linear local distortions introduced by the printing/scanning process. In this paper we present a significant improvement over existing methods, allowing to skip the calibration step and having a more accurate alignment, under all similarity transforms. Our method finds a robust and pixel accurate scanner independent alignment of the scanned image with the electronic document, allowing the extraction of accurate ground truth character information. The accuracy of the alignment is demonstrated using documents from the UW3 dataset. The results show that the mean distance between the estimated and the ground truth character bounding box position is less than one pixel.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"237 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123649377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
A Graphics Image Processing System 图形图像处理系统
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.84
Linlin Li, C. Tan
Patent document images maintained by the U.S. patent database have a specific format, in which figures and text descriptions are separated into different sections. This makes it difficult for users to refer to a figure while reading the description or vice versa. The system introduced in this paper is to prepare these patent images for a friendly user browsing interface. The system is able to extract captions and labels from figures. After obtaining captions and labels, figures and the relevant descriptions are linked together. Hence, users are able to easily find the relevant figure by clicking captions or labels in the description, or vice versa.
美国专利数据库维护的专利文件图像具有特定的格式,其中图形和文本描述被分成不同的部分。这使得用户在阅读描述时很难参考数字,反之亦然。本文所介绍的系统就是将这些专利图像制作成一个友好的用户浏览界面。该系统能够从图形中提取字幕和标签。在获得标题和标签后,将图形和相关描述链接在一起。因此,用户可以通过单击描述中的标题或标签轻松找到相关图形,反之亦然。
{"title":"A Graphics Image Processing System","authors":"Linlin Li, C. Tan","doi":"10.1109/DAS.2008.84","DOIUrl":"https://doi.org/10.1109/DAS.2008.84","url":null,"abstract":"Patent document images maintained by the U.S. patent database have a specific format, in which figures and text descriptions are separated into different sections. This makes it difficult for users to refer to a figure while reading the description or vice versa. The system introduced in this paper is to prepare these patent images for a friendly user browsing interface. The system is able to extract captions and labels from figures. After obtaining captions and labels, figures and the relevant descriptions are linked together. Hence, users are able to easily find the relevant figure by clicking captions or labels in the description, or vice versa.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114583608","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2008 The Eighth IAPR International Workshop on Document Analysis Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1