首页 > 最新文献

2016 12th IAPR Workshop on Document Analysis Systems (DAS)最新文献

英文 中文
Automatic Handwritten Character Segmentation for Paleographical Character Shape Analysis 用于古文字形状分析的自动手写字符分割
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.74
Théodore Bluche, D. Stutzmann, Christopher Kermorvant
Written texts are both physical (signs, shapes and graphical systems) and abstract objects (ideas), whose meanings and social connotations evolve through time. To study this dual nature of texts, palaeographers need to analyse large scale corpora at the finest granularity, such as character shape. This goal can only be reached through an automatic segmentation process. In this paper, we present a method, based on Handwritten Text Recognition, to automatically align images of digitized manuscripts with texts from scholarly editions, at the levels of page, column, line, word, and character. It has been successfully applied to two datasets of medieval manuscripts, which are now almost fully segmented at character level. The quality of the word and character segmentations are evaluated and further palaeographical analysis are presented.
书面文本既是物理的(符号、形状和图形系统),也是抽象的对象(思想),其意义和社会内涵随着时间的推移而演变。为了研究文本的这种双重性质,古学家需要在最细的粒度上分析大型语料库,例如字符形状。这一目标只能通过自动分割过程来实现。在本文中,我们提出了一种基于手写文本识别的方法,在页、列、行、词和字符级别上自动将数字化手稿的图像与学术版本的文本对齐。它已成功地应用于中世纪手稿的两个数据集,现在几乎完全分割在字符水平。对字词切分的质量进行了评价,并提出了进一步的古地理分析。
{"title":"Automatic Handwritten Character Segmentation for Paleographical Character Shape Analysis","authors":"Théodore Bluche, D. Stutzmann, Christopher Kermorvant","doi":"10.1109/DAS.2016.74","DOIUrl":"https://doi.org/10.1109/DAS.2016.74","url":null,"abstract":"Written texts are both physical (signs, shapes and graphical systems) and abstract objects (ideas), whose meanings and social connotations evolve through time. To study this dual nature of texts, palaeographers need to analyse large scale corpora at the finest granularity, such as character shape. This goal can only be reached through an automatic segmentation process. In this paper, we present a method, based on Handwritten Text Recognition, to automatically align images of digitized manuscripts with texts from scholarly editions, at the levels of page, column, line, word, and character. It has been successfully applied to two datasets of medieval manuscripts, which are now almost fully segmented at character level. The quality of the word and character segmentations are evaluated and further palaeographical analysis are presented.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125870543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
OCR Error Correction Using Character Correction and Feature-Based Word Classification 基于字符校正和特征词分类的OCR纠错
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.44
Ido Kissos, N. Dershowitz
This paper explores the use of a learned classifier for post-OCR text correction. Experiments with the Arabic language show that this approach, which integrates a weighted confusion matrix and a shallow language model, improves the vast majority of segmentation and recognition errors, the most frequent types of error on our dataset.
本文探讨了使用学习分类器进行后ocr文本校正。阿拉伯语的实验表明,这种方法集成了加权混淆矩阵和浅语言模型,改善了绝大多数分割和识别错误,这是我们数据集中最常见的错误类型。
{"title":"OCR Error Correction Using Character Correction and Feature-Based Word Classification","authors":"Ido Kissos, N. Dershowitz","doi":"10.1109/DAS.2016.44","DOIUrl":"https://doi.org/10.1109/DAS.2016.44","url":null,"abstract":"This paper explores the use of a learned classifier for post-OCR text correction. Experiments with the Arabic language show that this approach, which integrates a weighted confusion matrix and a shallow language model, improves the vast majority of segmentation and recognition errors, the most frequent types of error on our dataset.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131166329","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 67
Semi-automatic Text and Graphics Extraction of Manga Using Eye Tracking Information 基于眼动追踪信息的漫画半自动文本和图形提取
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.72
Christophe Rigaud, Thanh Nam Le, J. Burie, J. Ogier, Shoya Ishimaru, M. Iwata, K. Kise
The popularity of storing, distributing and reading comic books electronically has made the task of comics analysis an interesting research problem. Different work have been carried out aiming at understanding their layout structure and the graphic content. However the results are still far from universally applicable, largely due to the huge variety in expression styles and page arrangement, especially in manga (Japanese comics). In this paper, we propose a comic image analysis approach using eye-tracking data recorded during manga reading sessions. As humans are extremely capable of interpreting the structured drawing content, and show different reading behaviors based on the nature of the content, their eye movements follow distinguishable patterns over text or graphic regions. Therefore, eye gaze data can add rich information to the understanding of the manga content. Experimental results show that the fixations and saccades indeed form consistent patterns among readers, and can be used for manga textual and graphical analysis.
以电子方式存储、分发和阅读漫画书的普及使得分析漫画书的任务成为一个有趣的研究问题。为了了解它们的布局结构和图形内容,进行了不同的工作。然而,这些结果还远远不能普遍适用,这主要是由于表达风格和页面安排的巨大差异,特别是在漫画(日本漫画)中。在本文中,我们提出了一种漫画图像分析方法,使用在漫画阅读过程中记录的眼动数据。由于人类对结构化的绘画内容具有极强的解读能力,并根据内容的性质表现出不同的阅读行为,因此他们的眼球运动在文本或图形区域上遵循可区分的模式。因此,眼睛注视数据可以为理解漫画内容增加丰富的信息。实验结果表明,注视和扫视确实在读者之间形成了一致的模式,可以用于漫画的文本和图形分析。
{"title":"Semi-automatic Text and Graphics Extraction of Manga Using Eye Tracking Information","authors":"Christophe Rigaud, Thanh Nam Le, J. Burie, J. Ogier, Shoya Ishimaru, M. Iwata, K. Kise","doi":"10.1109/DAS.2016.72","DOIUrl":"https://doi.org/10.1109/DAS.2016.72","url":null,"abstract":"The popularity of storing, distributing and reading comic books electronically has made the task of comics analysis an interesting research problem. Different work have been carried out aiming at understanding their layout structure and the graphic content. However the results are still far from universally applicable, largely due to the huge variety in expression styles and page arrangement, especially in manga (Japanese comics). In this paper, we propose a comic image analysis approach using eye-tracking data recorded during manga reading sessions. As humans are extremely capable of interpreting the structured drawing content, and show different reading behaviors based on the nature of the content, their eye movements follow distinguishable patterns over text or graphic regions. Therefore, eye gaze data can add rich information to the understanding of the manga content. Experimental results show that the fixations and saccades indeed form consistent patterns among readers, and can be used for manga textual and graphical analysis.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130239364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Word Spotting in Historical Document Collections with Online-Handwritten Queries 联机手写查询在历史文献馆藏中的词识别
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.41
Christian Wieprecht, Leonard Rothacker, G. Fink
Pen-based systems are becoming more and more important due to the growing availability of touch sensitive devices in various forms and sizes. Their interfaces offer the possibility to directly interact with a system by natural handwriting. In contrast to other input modalities it is not required to switch to special modes, like software-keyboards. In this paper we propose a new method for querying digital archives of historical documents. Word images are retrieved with respect to search terms that users write on a pen-based system by hand. The captured trajectory is used as a query which we call query-by-online-trajectory word spotting. By using attribute embeddings for both online-trajectory and visual features, word images are retrieved based on their distance to the query in a common subspace. The system is therefore robust, as no explicit transcription for queries or word images is required. We evaluate our approach for writer-dependent as well as writer-independent scenarios, where we present highly accurate retrieval results in the former and compelling retrieval results in the latter case. Our performance is very competitive in comparison to related methods from the literature.
由于各种形式和尺寸的触摸敏感设备越来越多,基于笔的系统变得越来越重要。它们的接口提供了通过自然手写直接与系统交互的可能性。与其他输入模式相比,它不需要切换到特殊模式,如软件键盘。本文提出了一种新的历史文献数字档案查询方法。根据用户在基于笔的系统上手写的搜索词检索单词图像。捕获的轨迹被用作查询,我们称之为查询-按在线轨迹查找单词。通过对在线轨迹和视觉特征使用属性嵌入,基于它们在公共子空间中的距离来检索单词图像。因此,该系统是健壮的,因为不需要对查询或单词图像进行显式转录。我们评估了作者依赖和作者独立两种情况下的方法,前者提供了高度准确的检索结果,后者提供了令人信服的检索结果。与文献中的相关方法相比,我们的表现非常有竞争力。
{"title":"Word Spotting in Historical Document Collections with Online-Handwritten Queries","authors":"Christian Wieprecht, Leonard Rothacker, G. Fink","doi":"10.1109/DAS.2016.41","DOIUrl":"https://doi.org/10.1109/DAS.2016.41","url":null,"abstract":"Pen-based systems are becoming more and more important due to the growing availability of touch sensitive devices in various forms and sizes. Their interfaces offer the possibility to directly interact with a system by natural handwriting. In contrast to other input modalities it is not required to switch to special modes, like software-keyboards. In this paper we propose a new method for querying digital archives of historical documents. Word images are retrieved with respect to search terms that users write on a pen-based system by hand. The captured trajectory is used as a query which we call query-by-online-trajectory word spotting. By using attribute embeddings for both online-trajectory and visual features, word images are retrieved based on their distance to the query in a common subspace. The system is therefore robust, as no explicit transcription for queries or word images is required. We evaluate our approach for writer-dependent as well as writer-independent scenarios, where we present highly accurate retrieval results in the former and compelling retrieval results in the latter case. Our performance is very competitive in comparison to related methods from the literature.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132820100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Document Image Quality Assessment Using Discriminative Sparse Representation 基于判别稀疏表示的文档图像质量评估
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.24
Xujun Peng, Huaigu Cao, P. Natarajan
The goal of document image quality assessment (DIQA) is to build a computational model which can predict the degree of degradation for document images. Based on the estimated quality scores, the immediate feedback can be provided by document processing and analysis systems, which helps to maintain, organize, recognize and retrieve the information from document images. Recently, the bag-of-visual-words (BoV) based approaches have gained increasing attention from researchers to fulfill the task of quality assessment, but how to use BoV to represent images more accurately is still a challenging problem. In this paper, we propose to utilize a sparse representation based method to estimate document image's quality with respect to the OCR capability. Unlike the conventional sparse representation approaches, we introduce the target quality scores into the training phase of sparse representation. The proposed method improves the discriminability of the system and ensures the obtained codebook is more suitable for our assessment task. The experimental results on a public dataset show that the proposed method outperforms other hand-crafted and BoV based DIQA approaches.
文档图像质量评估(DIQA)的目标是建立一个能够预测文档图像退化程度的计算模型。根据估计的质量分数,文档处理和分析系统可以提供即时反馈,帮助维护、组织、识别和检索文档图像中的信息。近年来,基于视觉词袋(BoV)的图像质量评价方法越来越受到研究人员的关注,但如何使用视觉词袋更准确地表示图像仍然是一个具有挑战性的问题。在本文中,我们提出了一种基于稀疏表示的方法来估计文档图像的质量与OCR能力。与传统的稀疏表示方法不同,我们将目标质量分数引入到稀疏表示的训练阶段。该方法提高了系统的可分辨性,保证了得到的码本更适合我们的评估任务。在公共数据集上的实验结果表明,该方法优于其他手工制作和基于BoV的DIQA方法。
{"title":"Document Image Quality Assessment Using Discriminative Sparse Representation","authors":"Xujun Peng, Huaigu Cao, P. Natarajan","doi":"10.1109/DAS.2016.24","DOIUrl":"https://doi.org/10.1109/DAS.2016.24","url":null,"abstract":"The goal of document image quality assessment (DIQA) is to build a computational model which can predict the degree of degradation for document images. Based on the estimated quality scores, the immediate feedback can be provided by document processing and analysis systems, which helps to maintain, organize, recognize and retrieve the information from document images. Recently, the bag-of-visual-words (BoV) based approaches have gained increasing attention from researchers to fulfill the task of quality assessment, but how to use BoV to represent images more accurately is still a challenging problem. In this paper, we propose to utilize a sparse representation based method to estimate document image's quality with respect to the OCR capability. Unlike the conventional sparse representation approaches, we introduce the target quality scores into the training phase of sparse representation. The proposed method improves the discriminability of the system and ensures the obtained codebook is more suitable for our assessment task. The experimental results on a public dataset show that the proposed method outperforms other hand-crafted and BoV based DIQA approaches.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130758532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Quality Prediction System for Large-Scale Digitisation Workflows 面向大规模数字化工作流的质量预测系统
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.82
C. Clausner, S. Pletschacher, A. Antonacopoulos
The feasibility of large-scale OCR projects can so far only be assessed by running pilot studies on subsets of the target document collections and measuring the success of different workflows based on precise ground truth, which can be very costly to produce in the required volume. The premise of this paper is that, as an alternative, quality prediction may be used to approximate the success of a given OCR workflow. A new system is thus presented where a classifier is trained using metadata, image and layout features in combination with measured success rates (based on minimal ground truth). Subsequently, only document images are required as input for the numeric prediction of the quality score (no ground truth required). This way, the system can be applied to any number of similar (unseen) documents in order to assess their suitability for being processed using the particular workflow. The usefulness of the system has been validated using a realistic dataset of historical newspaper pages.
到目前为止,大规模OCR项目的可行性只能通过在目标文档集合的子集上运行试点研究来评估,并基于精确的地面事实衡量不同工作流程的成功,这在所需的数量上可能是非常昂贵的。本文的前提是,作为一种选择,质量预测可以用来近似给定OCR工作流的成功。因此,提出了一个新的系统,其中分类器是使用元数据、图像和布局特征与测量的成功率(基于最小基础真值)相结合来训练的。随后,只需要文档图像作为质量分数的数字预测的输入(不需要真实值)。通过这种方式,系统可以应用于任意数量的类似(不可见的)文档,以便评估它们是否适合使用特定工作流进行处理。使用历史报纸页面的真实数据集验证了该系统的实用性。
{"title":"Quality Prediction System for Large-Scale Digitisation Workflows","authors":"C. Clausner, S. Pletschacher, A. Antonacopoulos","doi":"10.1109/DAS.2016.82","DOIUrl":"https://doi.org/10.1109/DAS.2016.82","url":null,"abstract":"The feasibility of large-scale OCR projects can so far only be assessed by running pilot studies on subsets of the target document collections and measuring the success of different workflows based on precise ground truth, which can be very costly to produce in the required volume. The premise of this paper is that, as an alternative, quality prediction may be used to approximate the success of a given OCR workflow. A new system is thus presented where a classifier is trained using metadata, image and layout features in combination with measured success rates (based on minimal ground truth). Subsequently, only document images are required as input for the numeric prediction of the quality score (no ground truth required). This way, the system can be applied to any number of similar (unseen) documents in order to assess their suitability for being processed using the particular workflow. The usefulness of the system has been validated using a realistic dataset of historical newspaper pages.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126241189","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Text Extraction in Document Images: Highlight on Using Corner Points 文档图像中的文本提取:使用角点突出显示
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.67
Vikas Yadav, N. Ragot
During past years, text extraction in document images has been widely studied in the general context of Document Image Analysis (DIA) and especially in the framework of layout analysis. Many existing techniques rely on complex processes based on preprocessing, image transforms or component/edges extraction and their analysis. At the same time, text extraction inside videos has received an increased interest and the use of corner or key points has been proven to be very effective. Because it is noteworthy to notice that very few studies were performed on the use of corner points for text extraction in document images, we propose in this paper to evaluate the possibilities associated with this kind of approach for DIA. To do that, we designed a very simple technique based on FAST key points. A first stage divide the image into blocks and the density of points inside each one is computed. The more dense ones are kept as text blocks. Then, connectivity of blocks is checked to group them and to obtain complete text blocks. This technique has been evaluated on different kind of images: different languages (Telugu, Arabic, French), handwritten as well as typewritten, skewed documents, images at different resolution and with different kind and amount of noises (deformations, ink dot, bleed through, acquisition (blur, resolution)), etc. Even with fixed parameters for all such kind of documents images, the precision and recall are close or higher to 90% which makes this basic method already effective. Consequently, even if the proposed approach does not propose a breakthrough from theoretical aspects, it highlights that accurate text extraction could be achieved without complex approach. Moreover, this approach could also be easily improved to be more precise, robust and useful for more complex layout analysis.
近年来,在文档图像分析(DIA)的大背景下,特别是在布局分析的框架下,对文档图像中的文本提取进行了广泛的研究。许多现有的技术依赖于基于预处理、图像变换或成分/边缘提取及其分析的复杂过程。同时,视频内部的文本提取也受到越来越多的关注,使用角点或关键点被证明是非常有效的。因为值得注意的是,很少有研究在文档图像中使用角点进行文本提取,所以我们在本文中建议评估与这种DIA方法相关的可能性。为此,我们设计了一个非常简单的基于FAST关键点的技术。第一阶段将图像分成块,计算每个块内点的密度。较密集的则作为文本块保存。然后,检查文本块的连通性,对文本块进行分组,得到完整的文本块。该技术已在不同类型的图像上进行了评估:不同语言(泰卢固语,阿拉伯语,法语),手写和打字,倾斜文档,不同分辨率的图像以及不同类型和数量的噪声(变形,墨点,渗出,获取(模糊,分辨率))等。在固定参数的情况下,该方法的查全率和查全率都接近或高于90%,表明该方法是有效的。因此,即使所提出的方法没有从理论方面提出突破,它也强调了不需要复杂的方法就可以实现准确的文本提取。此外,这种方法也可以很容易地改进,以更精确,鲁棒性和有用的更复杂的布局分析。
{"title":"Text Extraction in Document Images: Highlight on Using Corner Points","authors":"Vikas Yadav, N. Ragot","doi":"10.1109/DAS.2016.67","DOIUrl":"https://doi.org/10.1109/DAS.2016.67","url":null,"abstract":"During past years, text extraction in document images has been widely studied in the general context of Document Image Analysis (DIA) and especially in the framework of layout analysis. Many existing techniques rely on complex processes based on preprocessing, image transforms or component/edges extraction and their analysis. At the same time, text extraction inside videos has received an increased interest and the use of corner or key points has been proven to be very effective. Because it is noteworthy to notice that very few studies were performed on the use of corner points for text extraction in document images, we propose in this paper to evaluate the possibilities associated with this kind of approach for DIA. To do that, we designed a very simple technique based on FAST key points. A first stage divide the image into blocks and the density of points inside each one is computed. The more dense ones are kept as text blocks. Then, connectivity of blocks is checked to group them and to obtain complete text blocks. This technique has been evaluated on different kind of images: different languages (Telugu, Arabic, French), handwritten as well as typewritten, skewed documents, images at different resolution and with different kind and amount of noises (deformations, ink dot, bleed through, acquisition (blur, resolution)), etc. Even with fixed parameters for all such kind of documents images, the precision and recall are close or higher to 90% which makes this basic method already effective. Consequently, even if the proposed approach does not propose a breakthrough from theoretical aspects, it highlights that accurate text extraction could be achieved without complex approach. Moreover, this approach could also be easily improved to be more precise, robust and useful for more complex layout analysis.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123341162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Election Tally Sheets Processing System 选举点票表处理系统
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.37
J. I. Toledo, A. Fornés, Jordi Cucurull-Juan, J. Lladós
In paper based elections, manual tallies at polling station level produce myriads of documents. These documents share a common form-like structure and a reduced vocabulary worldwide. On the other hand, each tally sheet is filled by a different writer and on different countries, different scripts are used. We present a complete document analysis system for electoral tally sheet processing combining state of the art techniques with a new handwriting recognition subprocess based on unsupervised feature discovery with Variational Autoencoders and sequence classification with BLSTM neural networks. The whole system is designed to be script independent and allows a fast and reliable results consolidation process with reduced operational cost.
在纸质选举中,投票站的人工点票产生了无数的文件。这些文档在全球范围内共享一个通用的类表单结构和简化的词汇表。另一方面,每个计数表由不同的作者填写,在不同的国家,使用不同的文字。我们提出了一个完整的文件分析系统,用于选举点票处理,结合了最先进的技术和新的手写识别子过程,该子过程基于无监督特征发现与变分自编码器和序列分类与BLSTM神经网络。整个系统被设计成独立于脚本,并允许快速可靠的结果整合过程,同时降低操作成本。
{"title":"Election Tally Sheets Processing System","authors":"J. I. Toledo, A. Fornés, Jordi Cucurull-Juan, J. Lladós","doi":"10.1109/DAS.2016.37","DOIUrl":"https://doi.org/10.1109/DAS.2016.37","url":null,"abstract":"In paper based elections, manual tallies at polling station level produce myriads of documents. These documents share a common form-like structure and a reduced vocabulary worldwide. On the other hand, each tally sheet is filled by a different writer and on different countries, different scripts are used. We present a complete document analysis system for electoral tally sheet processing combining state of the art techniques with a new handwriting recognition subprocess based on unsupervised feature discovery with Variational Autoencoders and sequence classification with BLSTM neural networks. The whole system is designed to be script independent and allows a fast and reliable results consolidation process with reduced operational cost.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114899520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Keyword Spotting in Handwritten Documents Using Projections of Oriented Gradients 利用定向梯度投影在手写文档中识别关键字
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.61
George Retsinas, G. Louloudis, N. Stamatopoulos, B. Gatos
In this paper, we present a novel approach for segmentation-based handwritten keyword spotting. The proposed approach relies upon the extraction of a simple yet efficient descriptor which is based on projections of oriented gradients. To this end, a global and a local word image descriptors, together with their combination, are proposed. Retrieval is performed using to the euclidean distance between the descriptors of a query image and the segmented word images. The proposed methods have been evaluated on the dataset of the ICFHR 2014 Competition on handwritten keyword spotting. Experimental results prove the efficiency of the proposed methods compared to several state-of-the-art techniques.
本文提出了一种基于分词的手写关键字识别方法。提出的方法依赖于提取一个简单而有效的描述符,该描述符基于定向梯度的投影。为此,提出了一种全局和局部词图像描述符及其组合。检索使用查询图像的描述符和分割的词图像之间的欧氏距离。所提出的方法已经在ICFHR 2014手写体关键词识别大赛的数据集上进行了评估。实验结果证明了该方法与几种最新技术相比的有效性。
{"title":"Keyword Spotting in Handwritten Documents Using Projections of Oriented Gradients","authors":"George Retsinas, G. Louloudis, N. Stamatopoulos, B. Gatos","doi":"10.1109/DAS.2016.61","DOIUrl":"https://doi.org/10.1109/DAS.2016.61","url":null,"abstract":"In this paper, we present a novel approach for segmentation-based handwritten keyword spotting. The proposed approach relies upon the extraction of a simple yet efficient descriptor which is based on projections of oriented gradients. To this end, a global and a local word image descriptors, together with their combination, are proposed. Retrieval is performed using to the euclidean distance between the descriptors of a query image and the segmented word images. The proposed methods have been evaluated on the dataset of the ICFHR 2014 Competition on handwritten keyword spotting. Experimental results prove the efficiency of the proposed methods compared to several state-of-the-art techniques.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124024583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Making Europe's Historical Newspapers Searchable 让欧洲历史报纸变得可搜索
Pub Date : 2016-04-11 DOI: 10.1109/DAS.2016.83
Clemens Neudecker, A. Antonacopoulos
This paper provides a rare glimpse into the overall approach for the refinement, i.e. the enrichment of scanned historical newspapers with text and layout recognition, in the Europeana Newspapers project. Within three years, the project processed more than 10 million pages of historical newspapers from 12 national and major libraries to produce the largest open access and fully searchable text collection of digital historical newspapers in Europe. In this, a wide variety of legal, logistical, technical and other challenges were encountered. After introducing the background issues in newspaper digitization in Europe, the paper discusses the technical aspects of refinement in greater detail. It explains what decisions were taken in the design of the large-scale processing workflow to address these challenges, what were the results produced and what were identified as best practices.
本文提供了一个难得的一瞥整体方法的细化,即丰富扫描历史报纸与文本和布局识别,在欧洲报纸项目。在三年内,该项目处理了来自12个国家和主要图书馆的1000多万页历史报纸,制作了欧洲最大的开放获取和完全可搜索的数字历史报纸文本集。在这方面,遇到了各种各样的法律、后勤、技术和其他挑战。在介绍了欧洲报纸数字化的背景问题后,本文更详细地讨论了细化的技术方面。它解释了在设计大规模处理工作流时采取了哪些决策来应对这些挑战,产生了哪些结果,以及哪些被确定为最佳实践。
{"title":"Making Europe's Historical Newspapers Searchable","authors":"Clemens Neudecker, A. Antonacopoulos","doi":"10.1109/DAS.2016.83","DOIUrl":"https://doi.org/10.1109/DAS.2016.83","url":null,"abstract":"This paper provides a rare glimpse into the overall approach for the refinement, i.e. the enrichment of scanned historical newspapers with text and layout recognition, in the Europeana Newspapers project. Within three years, the project processed more than 10 million pages of historical newspapers from 12 national and major libraries to produce the largest open access and fully searchable text collection of digital historical newspapers in Europe. In this, a wide variety of legal, logistical, technical and other challenges were encountered. After introducing the background issues in newspaper digitization in Europe, the paper discusses the technical aspects of refinement in greater detail. It explains what decisions were taken in the design of the large-scale processing workflow to address these challenges, what were the results produced and what were identified as best practices.","PeriodicalId":197359,"journal":{"name":"2016 12th IAPR Workshop on Document Analysis Systems (DAS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121259410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
期刊
2016 12th IAPR Workshop on Document Analysis Systems (DAS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1