首页 > 最新文献

Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)最新文献

英文 中文
Three decision levels strategy for Arabic and Latin texts differentiation in printed and handwritten natures 印刷和手写性质的阿拉伯语和拉丁语文本区分的三个决策层次策略
M. B. Jlaiel, S. Kanoun, A. Alimi, R. Mullot
Arabic and Latin script identification in printed and handwritten nature present several difficulties because the Arabic (printed or handwritten) and the handwritten Latin scripts are cursive scripts of nature. To avoid all possible confusions which can be generated, we propose in this paper a strategy which is based on three decision levels where each level will have its own features vector and will consist in identifying only one script among the scripts to identify.
由于阿拉伯语(印刷或手写)和手写拉丁字母本质上是草书,因此印刷和手写的阿拉伯和拉丁字母识别存在一些困难。为了避免所有可能产生的混淆,我们在本文中提出了一种基于三个决策层次的策略,每个决策层次都有自己的特征向量,并且只识别要识别的脚本中的一个脚本。
{"title":"Three decision levels strategy for Arabic and Latin texts differentiation in printed and handwritten natures","authors":"M. B. Jlaiel, S. Kanoun, A. Alimi, R. Mullot","doi":"10.1109/ICDAR.2007.250","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.250","url":null,"abstract":"Arabic and Latin script identification in printed and handwritten nature present several difficulties because the Arabic (printed or handwritten) and the handwritten Latin scripts are cursive scripts of nature. To avoid all possible confusions which can be generated, we propose in this paper a strategy which is based on three decision levels where each level will have its own features vector and will consist in identifying only one script among the scripts to identify.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125380476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Middle Zone Component Extraction and Recognition of Telugu Document Image 泰卢固语文档图像中间区分量提取与识别
L. Reddy, L. Satyaprasad, A. Sastry
Telugu is one of the ancient languages of South India. It has a complex orthography with a large number of distinct character shapes composed of simple and compound characters. The work reported in literature till the recent period is based on the connected component approach. Less attention is observed on the generalized character model and its application in the OCR development. Script syllable follows canonical structure where a consonant vowel core is preceded by one or two optional consonants .Formation of a syllable posses unique structural nature. In the present work, structural features of the syllable and the component model are combined to extract middle zone components. The shape of the middle zone components is closely related to a circle whereas other components are found with different topological features. Recognition rate of 99 percent is observed with the proposed method.
泰卢固语是印度南部的一种古老语言。它有一个复杂的正字法,有大量不同的汉字形状,由单字和复合字组成。直到最近,文献报道的工作都是基于关联成分方法。一般对广义字符模型及其在OCR开发中的应用关注较少。手写体音节遵循规范结构,其中辅音元音核心前面有一个或两个可选的辅音。音节的形成具有独特的结构性质。本文采用音节结构特征和成分模型相结合的方法提取中间区成分。中间区成分的形状与圆密切相关,而其他成分具有不同的拓扑特征。该方法的识别率达到99%。
{"title":"Middle Zone Component Extraction and Recognition of Telugu Document Image","authors":"L. Reddy, L. Satyaprasad, A. Sastry","doi":"10.1109/ICDAR.2007.169","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.169","url":null,"abstract":"Telugu is one of the ancient languages of South India. It has a complex orthography with a large number of distinct character shapes composed of simple and compound characters. The work reported in literature till the recent period is based on the connected component approach. Less attention is observed on the generalized character model and its application in the OCR development. Script syllable follows canonical structure where a consonant vowel core is preceded by one or two optional consonants .Formation of a syllable posses unique structural nature. In the present work, structural features of the syllable and the component model are combined to extract middle zone components. The shape of the middle zone components is closely related to a circle whereas other components are found with different topological features. Recognition rate of 99 percent is observed with the proposed method.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126901299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Knowledge-Based Recognition of Utility Map Sub-Diagrams 基于知识的实用地图子图识别
S. Hickinbotham, A. Cohn
An integrated map of all utility services in a locale would facilitate better management of the road infrastructure and the utilities themselves. To meet this goal, there exists a need to integrate raster scans of paper maps into GIS by capturing the semantic relationships between the objects in the drawings. In this context, commercially available vectorisation algorithms do not produce a sufficiently rich object representation. We present a structural object recognition system that successfully isolates sectional sub- diagrams in maps of underground utilities. This is built upon a vectorisation system based on a constrained Delau-nay triangulation of pen strokes.
一个地点所有公用事业服务的综合地图将有助于更好地管理道路基础设施和公用事业本身。为了实现这一目标,需要通过捕获绘图对象之间的语义关系,将纸质地图的栅格扫描集成到GIS中。在这种情况下,商业上可用的矢量化算法不能产生足够丰富的对象表示。提出了一种结构目标识别系统,该系统成功地分离了地下公用事业图中的截面子图。这是建立在一个矢量化系统的基础上的约束Delau-nay三角笔画。
{"title":"Knowledge-Based Recognition of Utility Map Sub-Diagrams","authors":"S. Hickinbotham, A. Cohn","doi":"10.1109/ICDAR.2007.152","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.152","url":null,"abstract":"An integrated map of all utility services in a locale would facilitate better management of the road infrastructure and the utilities themselves. To meet this goal, there exists a need to integrate raster scans of paper maps into GIS by capturing the semantic relationships between the objects in the drawings. In this context, commercially available vectorisation algorithms do not produce a sufficiently rich object representation. We present a structural object recognition system that successfully isolates sectional sub- diagrams in maps of underground utilities. This is built upon a vectorisation system based on a constrained Delau-nay triangulation of pen strokes.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"152 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127314354","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Data Mining Approach to Reading Order Detection 一种基于数据挖掘的阅读顺序检测方法
Pub Date : 2007-09-23 DOI: 10.1109/ICDAR.2007.4377050
Michelangelo Ceci, Margherita Berardi, G. Porcelli, D. Malerba
Determining the reading order for layout components extracted from a document image can be a crucial problem for several applications. It enables the reconstruction of a single textual element from texts associated to multiple layout components and makes both information extraction and content-based retrieval of documents more effective. A common aspect for all methods reported in the literature is that they strongly depend on the specific domain and are scarcely reusable when the classes of documents or the task at hand changes. In this paper, we investigate the problem of detecting the reading order of layout components by resorting to a data mining approach which acquires the domain specific knowledge from a set of training examples. The input of the learning method is the description of the "chains" of layout components defined by the user. Only spatial information is exploited to describe a chain, thus making the proposed approach also applicable to the cases in which no text can be associated to a layout component. The method induces a probabilistic classifier based on the Bayesian framework which is used for reconstructing either single or multiple chains of layout components. It has been evaluated on a set of document images.
确定从文档图像中提取的布局组件的读取顺序对于许多应用程序来说都是一个关键问题。它支持从与多个布局组件相关联的文本中重建单个文本元素,并使信息提取和基于内容的文档检索更加有效。文献中报道的所有方法的一个共同方面是,它们强烈依赖于特定的领域,当文档类或手头的任务发生变化时,它们几乎无法重用。本文采用数据挖掘的方法,从一组训练样本中获取领域特定知识,研究了布局组件阅读顺序的检测问题。学习方法的输入是用户定义的布局组件“链”的描述。仅利用空间信息来描述链,因此所提出的方法也适用于没有文本可以与布局组件相关联的情况。该方法引入了一个基于贝叶斯框架的概率分类器,用于重构单链或多链布局组件。它已经在一组文档图像上进行了评估。
{"title":"A Data Mining Approach to Reading Order Detection","authors":"Michelangelo Ceci, Margherita Berardi, G. Porcelli, D. Malerba","doi":"10.1109/ICDAR.2007.4377050","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.4377050","url":null,"abstract":"Determining the reading order for layout components extracted from a document image can be a crucial problem for several applications. It enables the reconstruction of a single textual element from texts associated to multiple layout components and makes both information extraction and content-based retrieval of documents more effective. A common aspect for all methods reported in the literature is that they strongly depend on the specific domain and are scarcely reusable when the classes of documents or the task at hand changes. In this paper, we investigate the problem of detecting the reading order of layout components by resorting to a data mining approach which acquires the domain specific knowledge from a set of training examples. The input of the learning method is the description of the \"chains\" of layout components defined by the user. Only spatial information is exploited to describe a chain, thus making the proposed approach also applicable to the cases in which no text can be associated to a layout component. The method induces a probabilistic classifier based on the Bayesian framework which is used for reconstructing either single or multiple chains of layout components. It has been evaluated on a set of document images.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125183809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Document Images Retrieval Based on Multiple Features Combination 基于多特征组合的文档图像检索
Gaofeng Meng, N. Zheng, Yonghong Song, Yuanlin Zhang
Retrieving the relevant document images from a great number of digitized pages with different kinds of artificial variations and documents quality deteriorations caused by scanning and printing is a meaningful and challenging problem. We attempt to deal with this problem by combining up multiple different kinds of document features in a hybrid way. Firstly, two new kinds of document image features based on the projection histograms and crossings number histograms of an image are proposed. Secondly, the proposed two features, together with density distribution feature and local binary pattern feature, are combined in a multistage structure to develop a novel document image retrieval system. Experimental results show that the proposed novel system is very efficient and robust for retrieving different kinds of document images, even if some of them are severely degraded.
从大量的数字化页面中检索相关的文档图像是一个有意义且具有挑战性的问题,这些页面具有各种人为变化和扫描和打印引起的文档质量下降。我们试图通过混合方式组合多种不同类型的文档特性来解决这个问题。首先,提出了基于投影直方图和交叉数直方图的两种新的文档图像特征;其次,将所提出的两种特征与密度分布特征和局部二值模式特征结合在一个多级结构中,开发了一种新的文档图像检索系统;实验结果表明,该系统在检索不同类型的文档图像时具有很高的效率和鲁棒性,即使某些文档图像严重退化。
{"title":"Document Images Retrieval Based on Multiple Features Combination","authors":"Gaofeng Meng, N. Zheng, Yonghong Song, Yuanlin Zhang","doi":"10.1109/ICDAR.2007.103","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.103","url":null,"abstract":"Retrieving the relevant document images from a great number of digitized pages with different kinds of artificial variations and documents quality deteriorations caused by scanning and printing is a meaningful and challenging problem. We attempt to deal with this problem by combining up multiple different kinds of document features in a hybrid way. Firstly, two new kinds of document image features based on the projection histograms and crossings number histograms of an image are proposed. Secondly, the proposed two features, together with density distribution feature and local binary pattern feature, are combined in a multistage structure to develop a novel document image retrieval system. Experimental results show that the proposed novel system is very efficient and robust for retrieving different kinds of document images, even if some of them are severely degraded.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121877801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Retrieval of Handwritten Lines in Historical Documents 历史文献中手写行检索
Lambert Schomaker
This study describes methods for the retrieval of handwritten lines of text in a historical administrative collection. The goal is to develop generic methods for bootstrapping the retrieval system from a tabula rasa starting condition, i.e., the virtual absence of labeled samples. By exploiting the currently available computing power and the fact that computation takes place off line, it should be possible to provide a good starting point for statistical learning methods. In this manner, a closed collection can be incrementally indexed. A cross-correlation method on line-strip images is presented and results are compared to feature-based methods.
本研究描述了在历史行政收藏中检索手写文本行的方法。目标是开发通用的方法来引导检索系统从一个表格初始条件,即,标记样本的虚拟缺席。通过利用当前可用的计算能力和离线进行计算的事实,应该可以为统计学习方法提供一个良好的起点。通过这种方式,可以对封闭集合进行增量索引。提出了一种基于线条图像的互相关方法,并与基于特征的方法进行了比较。
{"title":"Retrieval of Handwritten Lines in Historical Documents","authors":"Lambert Schomaker","doi":"10.1109/ICDAR.2007.219","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.219","url":null,"abstract":"This study describes methods for the retrieval of handwritten lines of text in a historical administrative collection. The goal is to develop generic methods for bootstrapping the retrieval system from a tabula rasa starting condition, i.e., the virtual absence of labeled samples. By exploiting the currently available computing power and the fact that computation takes place off line, it should be possible to provide a good starting point for statistical learning methods. In this manner, a closed collection can be incrementally indexed. A cross-correlation method on line-strip images is presented and results are compared to feature-based methods.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123124590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
A Case-Based Reasoning Approach for Invoice Structure Extraction 基于案例的发票结构提取推理方法
Hatem Hamza, Y. Belaïd, A. Belaïd
This paper shows the use of case-based reasoning (CBR) for invoice structure extraction and analysis. This method, called CBR-DIA (CBR for document invoice analysis), is adaptive and does not need any previous training. It analyses a document by retrieving and analysing similar documents or elements of documents (cases) stored in a database. The retrieval step is performed thanks to graph comparison techniques like graph probing and edit distance. The analysis step is done thanks to the information found in the nearest retrieved cases. Applied on 950 invoices, CBR-DIA reaches a recognition rate of 85.29% for documents of known classes and 76.33% for documents of unknown classes.
本文介绍了基于案例推理(CBR)在发票结构提取和分析中的应用。这种方法称为CBR- dia (CBR用于文档发票分析),是自适应的,不需要任何先前的培训。它通过检索和分析存储在数据库中的类似文档或文档元素(案例)来分析文档。检索步骤是通过图探测和编辑距离等图比较技术执行的。由于在最近的检索案例中找到了信息,分析步骤就完成了。应用于950张发票,CBR-DIA对已知类别单据的识别率为85.29%,对未知类别单据的识别率为76.33%。
{"title":"A Case-Based Reasoning Approach for Invoice Structure Extraction","authors":"Hatem Hamza, Y. Belaïd, A. Belaïd","doi":"10.1109/ICDAR.2007.3","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.3","url":null,"abstract":"This paper shows the use of case-based reasoning (CBR) for invoice structure extraction and analysis. This method, called CBR-DIA (CBR for document invoice analysis), is adaptive and does not need any previous training. It analyses a document by retrieving and analysing similar documents or elements of documents (cases) stored in a database. The retrieval step is performed thanks to graph comparison techniques like graph probing and edit distance. The analysis step is done thanks to the information found in the nearest retrieved cases. Applied on 950 invoices, CBR-DIA reaches a recognition rate of 85.29% for documents of known classes and 76.33% for documents of unknown classes.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129868197","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
HMM-Based Recognizer with Segmentation-free Strategy for Unconstrained Chinese Handwritten Text 基于hmm的无分割中文手写文本识别方法
Tong-Hua Su, Tian-Wen Zhang, Hu-Jie Huang, Yu Zhou
A segmentation-free strategy based on hidden Markov models (HMMs) is presented for offline recognition of unconstrained Chinese handwriting. As the first step, handwritten textlines are converted to observation sequence by sliding windows and character segmentation stage is avoided prior to recognition. Following that, embedded Baum-Welch algorithm is adopted to train character HMMs. Finally, best character string maximizing the a posteriori is located through Viterbi algorithm. Experiments are conducted on the HIT-MW database written by more than 780 writers. The results show: First, our baseline recognizer outperforms one segmentation-based OCR product with 35% relative improvement; second, more discriminative feature and compact representation, and state-tying technique to alleviate the data sparsity can enhance the recognizer with high confidence. The final recognizer has improved the performance by 10.77% than the baseline system.
提出了一种基于隐马尔可夫模型(hmm)的无分割策略,用于无约束汉字的离线识别。第一步,通过滑动窗口将手写文本行转换为观测序列,避免了识别前的字符分割阶段。然后,采用嵌入式Baum-Welch算法对字符hmm进行训练。最后,通过Viterbi算法找到后验值最大的最佳字符串。在780多位作者编写的HIT-MW数据库上进行了实验。结果表明:首先,我们的基线识别器比基于分割的OCR产品性能提高了35%;其次,采用更多的判别特征和更紧凑的表示,以及状态绑定技术来缓解数据稀疏性,可以提高识别器的高置信度。最终的识别器比基线系统的性能提高了10.77%。
{"title":"HMM-Based Recognizer with Segmentation-free Strategy for Unconstrained Chinese Handwritten Text","authors":"Tong-Hua Su, Tian-Wen Zhang, Hu-Jie Huang, Yu Zhou","doi":"10.1109/ICDAR.2007.133","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.133","url":null,"abstract":"A segmentation-free strategy based on hidden Markov models (HMMs) is presented for offline recognition of unconstrained Chinese handwriting. As the first step, handwritten textlines are converted to observation sequence by sliding windows and character segmentation stage is avoided prior to recognition. Following that, embedded Baum-Welch algorithm is adopted to train character HMMs. Finally, best character string maximizing the a posteriori is located through Viterbi algorithm. Experiments are conducted on the HIT-MW database written by more than 780 writers. The results show: First, our baseline recognizer outperforms one segmentation-based OCR product with 35% relative improvement; second, more discriminative feature and compact representation, and state-tying technique to alleviate the data sparsity can enhance the recognizer with high confidence. The final recognizer has improved the performance by 10.77% than the baseline system.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132866144","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Fusing Asynchronous Feature Streams for On-line Writer Identification 融合异步特征流的在线写入器识别
A. Schlapbach, H. Bunke
In this paper, we present a new approach to improving the performance of a writer identification system by fusing asynchronous feature streams. Different feature streams are extracted from on-line handwritten text acquired from a whiteboard. The feature streams are used to train a text and language independent writer identification system based on Gaussian mixture models (GMMs). From a stroke consisting of n points, n point-based feature vectors and one stroke-based feature vector are extracted. The resulting feature streams thus have an unequal number of feature vectors. We evaluate different methods to directly fuse the feature streams and show that, by means of feature fusion, we can improve the performance of the writer identification system on a data set produced by 200 different writers.
在本文中,我们提出了一种通过融合异步特征流来提高写入器识别系统性能的新方法。从白板上获取的在线手写文本中提取不同的特征流。利用特征流训练了基于高斯混合模型的独立于文本和语言的作家识别系统。从由n个点组成的笔画中提取n个基于点的特征向量和一个基于笔画的特征向量。由此产生的特征流具有不相等数量的特征向量。我们评估了不同的直接融合特征流的方法,并表明,通过特征融合,我们可以提高作者识别系统在由200个不同作者产生的数据集上的性能。
{"title":"Fusing Asynchronous Feature Streams for On-line Writer Identification","authors":"A. Schlapbach, H. Bunke","doi":"10.1109/ICDAR.2007.122","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.122","url":null,"abstract":"In this paper, we present a new approach to improving the performance of a writer identification system by fusing asynchronous feature streams. Different feature streams are extracted from on-line handwritten text acquired from a whiteboard. The feature streams are used to train a text and language independent writer identification system based on Gaussian mixture models (GMMs). From a stroke consisting of n points, n point-based feature vectors and one stroke-based feature vector are extracted. The resulting feature streams thus have an unequal number of feature vectors. We evaluate different methods to directly fuse the feature streams and show that, by means of feature fusion, we can improve the performance of the writer identification system on a data set produced by 200 different writers.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134380976","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Synthesis of Chinese Character Using Affine Transformation 用仿射变换合成汉字
Lianwen Jin, XiaoNa Zu
We present a novel Chinese synthesis method based on affine transform. A set of basic Chinese character element (BCCE) are designed, which can be used to generate any Chinese character in standard GB2312-80 level 1. Structure similarity measurement is used to evaluate the synthesis quality. Experiments showed that the synthesized characters look smooth and natural. Storage of synthesized characters can be greatly reduced. The proposed Chinese character synthesis method has many potential applications, such as building small-size Chinese font, building compact classifier for Chinese OCR, and etc.
提出了一种基于仿射变换的中文合成方法。设计了一套基本汉字元素(BCCE),可用于生成符合GB2312-80一级标准的任意汉字。采用结构相似性度量来评价合成质量。实验表明,合成的字符看起来平滑、自然。合成字符的存储量可以大大减少。提出的汉字合成方法在构建小字体、构建中文OCR压缩分类器等方面具有广泛的应用前景。
{"title":"Synthesis of Chinese Character Using Affine Transformation","authors":"Lianwen Jin, XiaoNa Zu","doi":"10.1109/ICDAR.2007.239","DOIUrl":"https://doi.org/10.1109/ICDAR.2007.239","url":null,"abstract":"We present a novel Chinese synthesis method based on affine transform. A set of basic Chinese character element (BCCE) are designed, which can be used to generate any Chinese character in standard GB2312-80 level 1. Structure similarity measurement is used to evaluate the synthesis quality. Experiments showed that the synthesized characters look smooth and natural. Storage of synthesized characters can be greatly reduced. The proposed Chinese character synthesis method has many potential applications, such as building small-size Chinese font, building compact classifier for Chinese OCR, and etc.","PeriodicalId":279268,"journal":{"name":"Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132048617","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
Ninth International Conference on Document Analysis and Recognition (ICDAR 2007)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1