首页 > 最新文献

2008 The Eighth IAPR International Workshop on Document Analysis Systems最新文献

英文 中文
Extraction of Text Objects in Video Documents: Recent Progress 视频文档中文本对象的提取:最新进展
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.49
Jing Zhang, R. Kasturi
Text extraction in video documents, as an important research field of content-based information indexing and retrieval, has been developing rapidly since 1990s. This has led to much progress in text extraction, performance evaluation, and related applications. By reviewing the approaches proposed during the past five years, this paper introduces the progress made in this area and discusses promising directions for future research.
视频文档文本提取作为基于内容的信息索引与检索的一个重要研究领域,自20世纪90年代以来得到了迅速发展。这在文本提取、性能评估和相关应用方面取得了很大进展。本文通过对近五年来提出的方法的回顾,介绍了该领域的研究进展,并对未来的研究方向进行了展望。
{"title":"Extraction of Text Objects in Video Documents: Recent Progress","authors":"Jing Zhang, R. Kasturi","doi":"10.1109/DAS.2008.49","DOIUrl":"https://doi.org/10.1109/DAS.2008.49","url":null,"abstract":"Text extraction in video documents, as an important research field of content-based information indexing and retrieval, has been developing rapidly since 1990s. This has led to much progress in text extraction, performance evaluation, and related applications. By reviewing the approaches proposed during the past five years, this paper introduces the progress made in this area and discusses promising directions for future research.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116697354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 194
Handling of Surface Modifications for Robust Image Based Mail Piece Comparison 基于鲁棒图像的邮件件比较的表面修改处理
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.39
K. Worm, B. Meffert
Image based mail piece identification is a new technology for optimizing the postal sorting process. Exploiting the mail piece surface uniqueness, characteristic features are used to identify each mail piece within a large amount of mail pieces. This process facilitates the storage of mail piece relevant data and its assignment in different sorting steps without manipulating the mail piece surface. Mail piece transportation and the mechanical mail piece sorting process may cause address and label movements, image rotations and further application-specific modifications. In contrast to other document identification systems image based mail piece identification requires high robustness towards above mentioned surface modifications for stable identification. This paper introduces four identification models for these special scenarios. Based on the proposed models text based feature extraction methods, feature representations and in particular appropriate distance metrics are presented which guarantee a robust mail piece identification. The applicability of the proposed procedure is shown in different experiments.
基于图像的邮件件识别是一种优化邮件分拣过程的新技术。利用邮件表面的唯一性,利用特征特征在大量邮件中识别每一个邮件。此过程便于在不操纵邮件表面的情况下存储邮件件相关数据并在不同的分拣步骤中进行分配。邮件运输和机械邮件分拣过程可能导致地址和标签移动、图像旋转和进一步的特定应用修改。与其他文档识别系统相比,基于图像的邮件件识别需要对上述表面修改具有较高的鲁棒性才能实现稳定的识别。本文介绍了针对这些特殊场景的四种识别模型。在此基础上,提出了基于文本的特征提取方法、特征表示和适当的距离度量,保证了邮件件识别的鲁棒性。不同的实验表明了所提方法的适用性。
{"title":"Handling of Surface Modifications for Robust Image Based Mail Piece Comparison","authors":"K. Worm, B. Meffert","doi":"10.1109/DAS.2008.39","DOIUrl":"https://doi.org/10.1109/DAS.2008.39","url":null,"abstract":"Image based mail piece identification is a new technology for optimizing the postal sorting process. Exploiting the mail piece surface uniqueness, characteristic features are used to identify each mail piece within a large amount of mail pieces. This process facilitates the storage of mail piece relevant data and its assignment in different sorting steps without manipulating the mail piece surface. Mail piece transportation and the mechanical mail piece sorting process may cause address and label movements, image rotations and further application-specific modifications. In contrast to other document identification systems image based mail piece identification requires high robustness towards above mentioned surface modifications for stable identification. This paper introduces four identification models for these special scenarios. Based on the proposed models text based feature extraction methods, feature representations and in particular appropriate distance metrics are presented which guarantee a robust mail piece identification. The applicability of the proposed procedure is shown in different experiments.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124799620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient Binarization of Historical and Degraded Document Images 历史和退化文档图像的有效二值化
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.66
B. Gatos, I. Pratikakis, S. Perantonis
This paper presents a new adaptive approach for the binarization and enhancement of historical and degraded documents. The proposed method is based on (i) efficient pre-processing; (ii) the combination of the results of several state-of-the-art binarization methodologies; (iii) the incorporation of edge information and (iv) the application of efficient image post-processing based on mathematical morphology for the enhancement of the final result. The proposed method demonstrated superior performance against six well-known techniques on numerous historical handwritten and machine-printed documents mainly from the Library of Congress of the United States archive. The performance evaluation was based on a consistent and concrete methodology.
本文提出了一种新的自适应方法对历史文献和退化文献进行二值化和增强。该方法基于(i)高效的预处理;结合几种最先进的二值化方法的结果;(iii)结合边缘信息和(iv)应用基于数学形态学的高效图像后处理来增强最终结果。在美国国会图书馆的大量历史手写和机器打印文件上,所提出的方法比六种已知的技术表现出更好的性能。绩效评价是根据一套一致的具体方法进行的。
{"title":"Efficient Binarization of Historical and Degraded Document Images","authors":"B. Gatos, I. Pratikakis, S. Perantonis","doi":"10.1109/DAS.2008.66","DOIUrl":"https://doi.org/10.1109/DAS.2008.66","url":null,"abstract":"This paper presents a new adaptive approach for the binarization and enhancement of historical and degraded documents. The proposed method is based on (i) efficient pre-processing; (ii) the combination of the results of several state-of-the-art binarization methodologies; (iii) the incorporation of edge information and (iv) the application of efficient image post-processing based on mathematical morphology for the enhancement of the final result. The proposed method demonstrated superior performance against six well-known techniques on numerous historical handwritten and machine-printed documents mainly from the Library of Congress of the United States archive. The performance evaluation was based on a consistent and concrete methodology.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123222059","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comprehensive Global Typography Extraction System for Electronic Book Documents 电子图书文档综合全球版式提取系统
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.30
Liangcai Gao, Zhi Tang, Xiaofan Lin, Ruiheng Qiu
Book documents usually have consistent typographies throughout the whole book, including headers, footers, columns, text line directions, and fonts used in the each level of headings. Such document-level typography information is of great value for downstream document processing applications. This paper presents a document analysis system that can extract a comprehensive set of typographies used in book documents. The system consists of several components: recognition of fonts used in the body text and chapter headings; detection of page body area, headers and footers; detection of columns, text line direction and line spacing of body text. Page-association is employed in the system. The preliminary experimental results demonstrate the effectiveness of the system.
图书文档通常在整本书中都有一致的排版,包括页眉、页脚、列、文本行方向和每一级标题中使用的字体。这种文档级排版信息对于下游文档处理应用程序非常有价值。本文提出了一个文档分析系统,该系统可以提取图书文档中使用的一套完整的排版。该系统由几个部分组成:正文和章节标题中使用的字体识别;检测页面主体区域、页眉和页脚;检测正文文本的列,文本行方向和行间距。系统采用页面关联。初步的实验结果证明了该系统的有效性。
{"title":"Comprehensive Global Typography Extraction System for Electronic Book Documents","authors":"Liangcai Gao, Zhi Tang, Xiaofan Lin, Ruiheng Qiu","doi":"10.1109/DAS.2008.30","DOIUrl":"https://doi.org/10.1109/DAS.2008.30","url":null,"abstract":"Book documents usually have consistent typographies throughout the whole book, including headers, footers, columns, text line directions, and fonts used in the each level of headings. Such document-level typography information is of great value for downstream document processing applications. This paper presents a document analysis system that can extract a comprehensive set of typographies used in book documents. The system consists of several components: recognition of fonts used in the body text and chapter headings; detection of page body area, headers and footers; detection of columns, text line direction and line spacing of body text. Page-association is employed in the system. The preliminary experimental results demonstrate the effectiveness of the system.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114908046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Truthing for Pixel-Accurate Segmentation 像素精确分割的真相
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.47
Michael A. Moll, H. Baird, Chang An
We discuss problems in developing policies for ground truthing document images for pixel-accurate segmentation. First, we describe ground truthing policies that apply to four different scales: (1) paragraph, (2) text line, (3) character, and (4) pixel. We then analyze difficult and/or ambiguous cases that will challenge any policy, e.g. blank space, overlapping content, etc. Experiments have shown the benefit of using "tighter'' zones that capture more detail (e.g., at the text line level, instead of paragraph). We show that tighter ground truth does significantly improve classification results, by 45% in recent experiments. It is important to face the fact that a pixel-accurate segmentation can be better than manually obtained ground truth. In practice, perfectly accurate pixel-level ground truth may not be achievable of course, but we believe it is important to explore methods to semi-automatically improve existing ground truth.
我们讨论了为实现像素精确分割而制定地面真实文档图像策略的问题。首先,我们描述了适用于四种不同尺度的地面真相策略:(1)段落,(2)文本行,(3)字符,(4)像素。然后,我们分析将挑战任何政策的困难和/或模糊情况,例如空白,重叠内容等。实验表明,使用“更紧凑”的区域可以捕获更多细节(例如,在文本行级别,而不是段落级别)。我们表明,在最近的实验中,更严格的基础真值确实显著提高了分类结果,提高了45%。重要的是要面对这样一个事实,即像素精确的分割可以比人工获得的地面真值更好。当然,在实践中,完全精确的像素级地面真值可能无法实现,但我们认为探索半自动提高现有地面真值的方法是很重要的。
{"title":"Truthing for Pixel-Accurate Segmentation","authors":"Michael A. Moll, H. Baird, Chang An","doi":"10.1109/DAS.2008.47","DOIUrl":"https://doi.org/10.1109/DAS.2008.47","url":null,"abstract":"We discuss problems in developing policies for ground truthing document images for pixel-accurate segmentation. First, we describe ground truthing policies that apply to four different scales: (1) paragraph, (2) text line, (3) character, and (4) pixel. We then analyze difficult and/or ambiguous cases that will challenge any policy, e.g. blank space, overlapping content, etc. Experiments have shown the benefit of using \"tighter'' zones that capture more detail (e.g., at the text line level, instead of paragraph). We show that tighter ground truth does significantly improve classification results, by 45% in recent experiments. It is important to face the fact that a pixel-accurate segmentation can be better than manually obtained ground truth. In practice, perfectly accurate pixel-level ground truth may not be achievable of course, but we believe it is important to explore methods to semi-automatically improve existing ground truth.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116403012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Accuracy Improvement and Objective Evaluation of Annotation Extraction from Printed Documents 打印文档注释提取的准确性改进与客观评价
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.80
T. Nakai, K. Iwata, K. Kise
There is an approach of annotation extraction from printed documents in which annotations are extracted by comparing the image of an annotated document and its original document image. In one of the previous methods, the image of an original document is actually printed and scanned in order to reproduce image degradations of the image of the annotated document. However such a method lacks convenience since users have to use the same printer and scanner to obtain images of an annotated document and its original document. In this paper, we propose an improved annotation extraction method in which the image degradations are compensated by image processing. In the proposed method, the difference between original and annotated document images due to image degradations is reduced by not only removal of the degradations in the annotated document images but also reproduction of the degradations in the original document images. The proposed method consists of three steps of processing which are for dithering, for color change, and for local displacement. We also propose an objective evaluation of extracted annotations to compare the experimental results accurately. Experimental results of the proposed method have shown that the recall of extracted annotations was 80.94% and the precision was 85.59%.
有一种从打印文档中提取注释的方法,通过比较注释文档的图像和原始文档图像来提取注释。在前面的一种方法中,实际打印和扫描原始文档的图像,以再现注释文档图像的图像退化。然而,这种方法缺乏便利性,因为用户必须使用相同的打印机和扫描仪来获取注释文档及其原始文档的图像。本文提出了一种改进的注释提取方法,该方法通过图像处理来补偿图像的退化。在该方法中,通过去除注释文档图像中的退化,并复制原始文档图像中的退化,来减小由于图像退化而导致的原始文档图像与注释文档图像之间的差异。该方法由抖动处理、颜色变化处理和局部位移处理三个步骤组成。我们还提出了对提取的注释进行客观评价,以便准确地比较实验结果。实验结果表明,该方法提取的注释查全率为80.94%,查准率为85.59%。
{"title":"Accuracy Improvement and Objective Evaluation of Annotation Extraction from Printed Documents","authors":"T. Nakai, K. Iwata, K. Kise","doi":"10.1109/DAS.2008.80","DOIUrl":"https://doi.org/10.1109/DAS.2008.80","url":null,"abstract":"There is an approach of annotation extraction from printed documents in which annotations are extracted by comparing the image of an annotated document and its original document image. In one of the previous methods, the image of an original document is actually printed and scanned in order to reproduce image degradations of the image of the annotated document. However such a method lacks convenience since users have to use the same printer and scanner to obtain images of an annotated document and its original document. In this paper, we propose an improved annotation extraction method in which the image degradations are compensated by image processing. In the proposed method, the difference between original and annotated document images due to image degradations is reduced by not only removal of the degradations in the annotated document images but also reproduction of the degradations in the original document images. The proposed method consists of three steps of processing which are for dithering, for color change, and for local displacement. We also propose an objective evaluation of extracted annotations to compare the experimental results accurately. Experimental results of the proposed method have shown that the recall of extracted annotations was 80.94% and the precision was 85.59%.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116406313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An Efficient Edge Based Technique for Text Detection in Video Frames 基于边缘的视频帧文本检测方法
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.17
P. Shivakumara, Weihua Huang, C. Tan
Both graphic text and scene text detection in video images with complex background and low resolution is still a challenging and interesting problem for researchers in the field of image processing and computer vision. In this paper, we present a novel technique for detecting both graphic text and scene text in video images by finding segments containing text in an input image and then using statistical features such as vertical and horizontal bars for edges in the segments for detecting true text blocks efficiently. To identify a segment containing text, heuristic rules are formed based on combination of filters and edge analysis. Furthermore, the same rules are extended to grow the boundaries of a candidate segment in order to include complete text in the input image. The experimental results of the proposed method show that the technique performs better than existing methods in terms of a number of metrics.
背景复杂、分辨率低的视频图像中图形文本和场景文本的检测仍然是图像处理和计算机视觉领域的研究热点和难点。在本文中,我们提出了一种检测视频图像中图形文本和场景文本的新技术,通过在输入图像中找到包含文本的片段,然后使用统计特征(如片段中的垂直和水平条)来有效地检测真正的文本块。为了识别包含文本的片段,将过滤器和边缘分析相结合,形成启发式规则。此外,将相同的规则扩展到扩展候选段的边界,以便在输入图像中包含完整的文本。实验结果表明,该方法在多个指标上都优于现有方法。
{"title":"An Efficient Edge Based Technique for Text Detection in Video Frames","authors":"P. Shivakumara, Weihua Huang, C. Tan","doi":"10.1109/DAS.2008.17","DOIUrl":"https://doi.org/10.1109/DAS.2008.17","url":null,"abstract":"Both graphic text and scene text detection in video images with complex background and low resolution is still a challenging and interesting problem for researchers in the field of image processing and computer vision. In this paper, we present a novel technique for detecting both graphic text and scene text in video images by finding segments containing text in an input image and then using statistical features such as vertical and horizontal bars for edges in the segments for detecting true text blocks efficiently. To identify a segment containing text, heuristic rules are formed based on combination of filters and edge analysis. Furthermore, the same rules are extended to grow the boundaries of a candidate segment in order to include complete text in the input image. The experimental results of the proposed method show that the technique performs better than existing methods in terms of a number of metrics.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128374012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 69
A Complete Optical Character Recognition Methodology for Historical Documents 一个完整的历史文献光学字符识别方法
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.73
G. Vamvakas, B. Gatos, N. Stamatopoulos, S. Perantonis
In this paper a complete OCR methodology for recognizing historical documents, either printed or handwritten without any knowledge of the font, is presented. This methodology consists of three steps: The first two steps refer to creating a database for training using a set of documents, while the third one refers to recognition of new document images. First, a pre-processing step that includes image binarization and enhancement takes place. At a second step a top-down segmentation approach is used in order to detect text lines, words and characters. A clustering scheme is then adopted in order to group characters of similar shape. This is a semi-automatic procedure since the user is able to interact at any time in order to correct possible errors of clustering and assign an ASCII label. After this step, a database is created in order to be used for recognition. Finally, in the third step, for every new document image the above segmentation approach takes place while the recognition is based on the character database that has been produced at the previous step.
本文提出了一种完整的OCR方法,用于识别历史文档,无论是印刷的还是手写的,都不需要任何字体知识。该方法包括三个步骤:前两个步骤是指使用一组文档创建用于训练的数据库,而第三个步骤是指识别新的文档图像。首先,预处理步骤包括图像二值化和增强。第二步,使用自顶向下的分割方法来检测文本行、词和字符。然后采用聚类方法对形状相似的字符进行分组。这是一个半自动的过程,因为用户可以在任何时候进行交互,以纠正聚类的可能错误并分配一个ASCII标签。在此步骤之后,将创建一个数据库,以便用于识别。最后,在第三步中,对每个新的文档图像进行上述分割方法,同时基于前一步产生的字符数据库进行识别。
{"title":"A Complete Optical Character Recognition Methodology for Historical Documents","authors":"G. Vamvakas, B. Gatos, N. Stamatopoulos, S. Perantonis","doi":"10.1109/DAS.2008.73","DOIUrl":"https://doi.org/10.1109/DAS.2008.73","url":null,"abstract":"In this paper a complete OCR methodology for recognizing historical documents, either printed or handwritten without any knowledge of the font, is presented. This methodology consists of three steps: The first two steps refer to creating a database for training using a set of documents, while the third one refers to recognition of new document images. First, a pre-processing step that includes image binarization and enhancement takes place. At a second step a top-down segmentation approach is used in order to detect text lines, words and characters. A clustering scheme is then adopted in order to group characters of similar shape. This is a semi-automatic procedure since the user is able to interact at any time in order to correct possible errors of clustering and assign an ASCII label. After this step, a database is created in order to be used for recognition. Finally, in the third step, for every new document image the above segmentation approach takes place while the recognition is based on the character database that has been produced at the previous step.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"151 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130401513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Detecting Gradients in Text Images Using the Hough Transform 使用霍夫变换检测文本图像中的梯度
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.55
Dimosthenis Karatzas
The use of gradients in text images is nowadays quite frequent. Existing segmentation methods encounter serious problems when it comes to modern text images where gradients might appear in the background or the foreground or both at the same time. This paper presents an approach for lightness gradient areas detection based on the Hough Transform. The issues arising are discussed, and results are presented on a dataset comprising Web images, logos and scanned documents.
在文本图像中使用渐变现在是相当频繁的。当涉及到现代文本图像时,现有的分割方法遇到了严重的问题,其中梯度可能出现在背景或前景中,或者同时出现在两者中。提出了一种基于霍夫变换的亮度梯度区域检测方法。讨论了产生的问题,并在包含Web图像、徽标和扫描文档的数据集中展示了结果。
{"title":"Detecting Gradients in Text Images Using the Hough Transform","authors":"Dimosthenis Karatzas","doi":"10.1109/DAS.2008.55","DOIUrl":"https://doi.org/10.1109/DAS.2008.55","url":null,"abstract":"The use of gradients in text images is nowadays quite frequent. Existing segmentation methods encounter serious problems when it comes to modern text images where gradients might appear in the background or the foreground or both at the same time. This paper presents an approach for lightness gradient areas detection based on the Hough Transform. The issues arising are discussed, and results are presented on a dataset comprising Web images, logos and scanned documents.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121517825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Super-Resolution of Text Images Using Edge-Directed Tangent Field 使用边缘定向切线场的文本图像超分辨率
Pub Date : 2008-09-16 DOI: 10.1109/DAS.2008.26
Jyotirmoy Banerjee, C. V. Jawahar
This paper presents an edge-directed super-resolution algorithm for document images without using any training set. This technique creates an image with smooth regions in both the foreground and the background, while allowing sharp discontinuities across and smoothness along the edges. Our method preserves sharp corners in text images by using the local edge direction, which is computed first by evaluating the gradient field and then taking its tangent. Super-resolution of document images is characterized by bimodality, smoothness along the edges as well as subsampling consistency. These characteristics are enforced in a Markov random field (MRF) framework by defining an appropriate energy function. In our method, subsampling of super-resolution image will return the original low-resolution one, proving the correctness of the method. The super-resolution image, is generated by iteratively reducing this energy function. Experimental results on a variety of input images, demonstrate the effectiveness of our method for document image super-resolution.
提出了一种不使用任何训练集的文档图像边缘定向超分辨算法。这种技术可以在前景和背景中创建一个平滑区域的图像,同时允许明显的不连续性和平滑的边缘。我们的方法通过使用局部边缘方向来保留文本图像中的尖锐角,该方向首先通过计算梯度场然后取其切线来计算。文档图像的超分辨率具有双峰性、边缘平滑性和次采样一致性等特点。通过定义适当的能量函数,在马尔可夫随机场(MRF)框架中实现这些特征。在我们的方法中,超分辨率图像的子采样将返回原始的低分辨率图像,证明了该方法的正确性。通过对该能量函数进行迭代约简,生成超分辨率图像。在多种输入图像上的实验结果证明了该方法对文档图像超分辨率的有效性。
{"title":"Super-Resolution of Text Images Using Edge-Directed Tangent Field","authors":"Jyotirmoy Banerjee, C. V. Jawahar","doi":"10.1109/DAS.2008.26","DOIUrl":"https://doi.org/10.1109/DAS.2008.26","url":null,"abstract":"This paper presents an edge-directed super-resolution algorithm for document images without using any training set. This technique creates an image with smooth regions in both the foreground and the background, while allowing sharp discontinuities across and smoothness along the edges. Our method preserves sharp corners in text images by using the local edge direction, which is computed first by evaluating the gradient field and then taking its tangent. Super-resolution of document images is characterized by bimodality, smoothness along the edges as well as subsampling consistency. These characteristics are enforced in a Markov random field (MRF) framework by defining an appropriate energy function. In our method, subsampling of super-resolution image will return the original low-resolution one, proving the correctness of the method. The super-resolution image, is generated by iteratively reducing this energy function. Experimental results on a variety of input images, demonstrate the effectiveness of our method for document image super-resolution.","PeriodicalId":423207,"journal":{"name":"2008 The Eighth IAPR International Workshop on Document Analysis Systems","volume":"155 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133909660","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
期刊
2008 The Eighth IAPR International Workshop on Document Analysis Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1