首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
GM-transfer: Graph-based model for transfer learning 迁移:基于图的迁移学习模型
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166601
Shizhun Yang, Chenping Hou, Yi Wu
Traditional data mining and machine learning technologies may fail when the training data and the testing data are drawn from different feature spaces and different distributions. Transfer learning, which uses the data from source domain and target domain, can tackle this problem. In this paper, we propose an improved Graph-based Model for Transfer learning (GM-Transfer). We construct a tripartite graph to represent the transfer learning problem and model the relations between the source domain data and the target domain data more efficiently. By learning the informational graph, the knowledge from the source domain data can be transferred to help improve the learning efficiency on the target domain data. Experiments show the effectiveness of our algorithm.
当训练数据和测试数据来自不同的特征空间和不同的分布时,传统的数据挖掘和机器学习技术可能会失败。使用源域和目标域数据的迁移学习可以解决这一问题。本文提出了一种改进的迁移学习模型(GM-Transfer)。我们构造了一个三部分图来表示迁移学习问题,并更有效地对源域数据和目标域数据之间的关系进行建模。通过对信息图的学习,可以将源领域数据中的知识进行迁移,从而提高对目标领域数据的学习效率。实验证明了算法的有效性。
{"title":"GM-transfer: Graph-based model for transfer learning","authors":"Shizhun Yang, Chenping Hou, Yi Wu","doi":"10.1109/ACPR.2011.6166601","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166601","url":null,"abstract":"Traditional data mining and machine learning technologies may fail when the training data and the testing data are drawn from different feature spaces and different distributions. Transfer learning, which uses the data from source domain and target domain, can tackle this problem. In this paper, we propose an improved Graph-based Model for Transfer learning (GM-Transfer). We construct a tripartite graph to represent the transfer learning problem and model the relations between the source domain data and the target domain data more efficiently. By learning the informational graph, the knowledge from the source domain data can be transferred to help improve the learning efficiency on the target domain data. Experiments show the effectiveness of our algorithm.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"251 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116209243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ℓ2;1-norm based Regression for Classification 1-范数回归分类
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166615
Chuan-Xian Ren, D. Dai, Hong Yan
We present a novel classification method formulating an objective model by ℓ2;1-norm based regression. The ℓ2;1-norm based loss function is robust to outliers or the large variations within given data, and the ℓ2;1-norm regularization term selects correlated samples across the whole training set with grouped sparsity. This constrained optimization problem can be efficiently solved by an iterative procedure. Several benchmark data sets including facial images and gene expression data are used for evaluating the robustness and effectiveness of the new proposed algorithm, and the results show the competitive performance.
本文提出了一种新的分类方法,即利用1,2 -范数回归建立目标模型。基于1,2范数的损失函数对异常值或给定数据内的大变化具有鲁棒性,并且1,2范数正则化项在整个训练集中选择具有分组稀疏性的相关样本。这种约束优化问题可以通过迭代过程有效地求解。利用人脸图像和基因表达数据等基准数据集对新算法的鲁棒性和有效性进行了评价,结果表明该算法具有较强的竞争力。
{"title":"ℓ2;1-norm based Regression for Classification","authors":"Chuan-Xian Ren, D. Dai, Hong Yan","doi":"10.1109/ACPR.2011.6166615","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166615","url":null,"abstract":"We present a novel classification method formulating an objective model by ℓ2;1-norm based regression. The ℓ2;1-norm based loss function is robust to outliers or the large variations within given data, and the ℓ2;1-norm regularization term selects correlated samples across the whole training set with grouped sparsity. This constrained optimization problem can be efficiently solved by an iterative procedure. Several benchmark data sets including facial images and gene expression data are used for evaluating the robustness and effectiveness of the new proposed algorithm, and the results show the competitive performance.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121563807","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Palmprint verification using binary contrast context vector 使用二元对比上下文向量的掌纹验证
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166566
Yi C. Feng, Lei Huang, Chang-ping Liu
Palmprint recognition has attracted much attention in recent years. Many algorithms based texture coding achieve high accuracy. However they are still sensitive to local unsteady region introduced by variations of hand pose and other conditions. In this paper we proposed a novel feature extraction algorithm, namely binary contrast context vector (BCCV), to represent multiple contrast distribution for a local region. Due to forming the local contrast value into a binary vector, contrast context could be used to match more effectively. Furthermore, by using BCCV we apply an adaptive threshold to mask the stable local region before matching. Our experiment results on public palmprint database shows that the proposed BCCV achieves lower equal error rate (EER) than other two state-of-the-art approaches.
近年来,掌纹识别引起了人们的广泛关注。许多基于纹理编码的算法都达到了较高的精度。然而,由于手姿等条件的变化,它们对局部不稳定区域仍然很敏感。本文提出了一种新的特征提取算法,即二元对比上下文向量(binary contrast context vector, BCCV)来表示局部区域的多个对比分布。由于将局部对比度值形成二值向量,可以更有效地利用对比度上下文进行匹配。此外,利用BCCV方法在匹配前对稳定的局部区域进行自适应阈值屏蔽。我们在公共掌纹数据库上的实验结果表明,与其他两种最先进的方法相比,所提出的BCCV方法的等错误率(EER)更低。
{"title":"Palmprint verification using binary contrast context vector","authors":"Yi C. Feng, Lei Huang, Chang-ping Liu","doi":"10.1109/ACPR.2011.6166566","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166566","url":null,"abstract":"Palmprint recognition has attracted much attention in recent years. Many algorithms based texture coding achieve high accuracy. However they are still sensitive to local unsteady region introduced by variations of hand pose and other conditions. In this paper we proposed a novel feature extraction algorithm, namely binary contrast context vector (BCCV), to represent multiple contrast distribution for a local region. Due to forming the local contrast value into a binary vector, contrast context could be used to match more effectively. Furthermore, by using BCCV we apply an adaptive threshold to mask the stable local region before matching. Our experiment results on public palmprint database shows that the proposed BCCV achieves lower equal error rate (EER) than other two state-of-the-art approaches.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132306158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Translation-invariant scene grouping 平移不变场景分组
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166542
Pin-Ching Su, Hwann-Tzong Chen, Koichi Ito, T. Aoki
We present a new approach to the problem of grouping similar scene images. The proposed method characterizes both the global feature layout and the local oriented edge responses of an image, and provides a translation-invariant similarity measure to compare scene images. Our method is effective in generating initial clustering results for applications that require extensive local-feature matching on unorganized image collections, such as large-scale 3D reconstruction and scene completion. The advantage of our method is that it can estimate image similarity via integrating global and local information. The experimental evaluations on various image datasets show that our method is able to approximate well the similarities derived from local-feature matching with a lower computational cost.
我们提出了一种新的方法来解决相似场景图像的分组问题。该方法同时表征了图像的全局特征布局和局部定向边缘响应,并提供了一种平移不变的相似度量来比较场景图像。对于需要在无组织图像集合上进行大量局部特征匹配的应用,例如大规模3D重建和场景补全,我们的方法在生成初始聚类结果方面是有效的。该方法的优点是可以通过综合全局和局部信息来估计图像的相似度。在不同图像数据集上的实验评估表明,我们的方法能够很好地逼近由局部特征匹配得到的相似度,并且计算成本较低。
{"title":"Translation-invariant scene grouping","authors":"Pin-Ching Su, Hwann-Tzong Chen, Koichi Ito, T. Aoki","doi":"10.1109/ACPR.2011.6166542","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166542","url":null,"abstract":"We present a new approach to the problem of grouping similar scene images. The proposed method characterizes both the global feature layout and the local oriented edge responses of an image, and provides a translation-invariant similarity measure to compare scene images. Our method is effective in generating initial clustering results for applications that require extensive local-feature matching on unorganized image collections, such as large-scale 3D reconstruction and scene completion. The advantage of our method is that it can estimate image similarity via integrating global and local information. The experimental evaluations on various image datasets show that our method is able to approximate well the similarities derived from local-feature matching with a lower computational cost.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"139 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133912814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiple view semi-supervised discriminant analysis 多视图半监督判别分析
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166562
Xuesong Yin, Xiaodong Chen, Xiaofang Ruan, Yarong Huang
Beyond conventional semi-supervised dimensionality reduction methods which data are represented in a single vector or graph space, multiple view semi-supervised ones are to learn a hidden consensus pattern from multiple representations of multiple view data together with some domain knowledge. Under multiple view settings, we propose a new Multiple view Semi-supervised Discriminant Analysis (MSDA). Specifically, the labeled data are used to infer the discriminant structure in each view. Simultaneously, all the data, including the labeled and the unlabeled instances, are used to discover the intrinsic geometrical structure in each view. Thus, we can learn an optimal pattern from the multiple patterns of multiple representations with serial combination after getting the projection of each view. Experiments carried out on real-world data sets by MSDA show a clear improvement over the results of representative dimensionality reduction algorithms.
传统的半监督降维方法是将数据表示在单个向量或图空间中,而多视图半监督降维方法则是从多视图数据的多个表示中学习隐藏的一致模式,并结合一些领域知识。在多视图设置下,我们提出了一种新的多视图半监督判别分析(MSDA)。具体来说,标记的数据用于推断每个视图中的判别结构。同时,利用所有的数据,包括标记和未标记的实例,来发现每个视图的内在几何结构。因此,在得到每个视图的投影后,我们可以从多个表示的多个模式中学习到一个最优模式。在实际数据集上进行的实验表明,MSDA比代表性降维算法的结果有明显改善。
{"title":"Multiple view semi-supervised discriminant analysis","authors":"Xuesong Yin, Xiaodong Chen, Xiaofang Ruan, Yarong Huang","doi":"10.1109/ACPR.2011.6166562","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166562","url":null,"abstract":"Beyond conventional semi-supervised dimensionality reduction methods which data are represented in a single vector or graph space, multiple view semi-supervised ones are to learn a hidden consensus pattern from multiple representations of multiple view data together with some domain knowledge. Under multiple view settings, we propose a new Multiple view Semi-supervised Discriminant Analysis (MSDA). Specifically, the labeled data are used to infer the discriminant structure in each view. Simultaneously, all the data, including the labeled and the unlabeled instances, are used to discover the intrinsic geometrical structure in each view. Thus, we can learn an optimal pattern from the multiple patterns of multiple representations with serial combination after getting the projection of each view. Experiments carried out on real-world data sets by MSDA show a clear improvement over the results of representative dimensionality reduction algorithms.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133068153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Face attractiveness improvement using beauty prototypes and decision 利用美貌原型和决策提高面部吸引力
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166544
Mingming Sun, D. Zhang, Jing-yu Yang
To improve the attractiveness of a face, intuitively, one can drive the face approaching some beautiful faces. There are two major problem to solve for implementing the intuitive solution. One problem is that how to define and discover suitable beauty prototypes. Another is that how to determine the balance between the original face and the beauty prototype to produce the desired face. In this paper, we proposed a quantitive method to solve these two problems. First, a set of beautiful face prototypes are identified as cluster centers of beautiful faces, which avoid involving specific personal facial characteristic. Second, a beauty decision function is learned as a classifier that can tell whether a face is beautiful or not. Then, the facial attractiveness improvement procedure finds the nearest beauty prototype for the original face, and then approaches the prototype from the original face until the beauty decision function tells the approaching face is beautiful. With this method, the face is beautified and the difference between the beautified face and the original face is minimized. The experimental results verify the validity of the proposed methods.
为了提高一张脸的吸引力,直觉上,一个人可以驱使脸接近一些美丽的脸。实现直观的解决方案需要解决两个主要问题。一个问题是如何定义和发现合适的美原型。另一个问题是如何确定原始面孔和美的原型之间的平衡,以产生理想的面孔。在本文中,我们提出了一种定量的方法来解决这两个问题。首先,将一组美丽面孔原型识别为美丽面孔的聚类中心,避免涉及特定的个人面部特征;其次,学习一个美丽决策函数作为一个分类器,可以判断一张脸是否漂亮。然后,面部吸引力改进程序为原始面孔找到最接近的美丽原型,然后从原始面孔接近原型,直到美丽决策函数告诉接近的面孔是美丽的。该方法对人脸进行美化,使美化后的人脸与原始人脸之间的差异最小化。实验结果验证了所提方法的有效性。
{"title":"Face attractiveness improvement using beauty prototypes and decision","authors":"Mingming Sun, D. Zhang, Jing-yu Yang","doi":"10.1109/ACPR.2011.6166544","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166544","url":null,"abstract":"To improve the attractiveness of a face, intuitively, one can drive the face approaching some beautiful faces. There are two major problem to solve for implementing the intuitive solution. One problem is that how to define and discover suitable beauty prototypes. Another is that how to determine the balance between the original face and the beauty prototype to produce the desired face. In this paper, we proposed a quantitive method to solve these two problems. First, a set of beautiful face prototypes are identified as cluster centers of beautiful faces, which avoid involving specific personal facial characteristic. Second, a beauty decision function is learned as a classifier that can tell whether a face is beautiful or not. Then, the facial attractiveness improvement procedure finds the nearest beauty prototype for the original face, and then approaches the prototype from the original face until the beauty decision function tells the approaching face is beautiful. With this method, the face is beautified and the difference between the beautified face and the original face is minimized. The experimental results verify the validity of the proposed methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115576518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Spatially-adaptive regularized super-resolution image reconstruction using a gradient-based saliency measure 基于梯度显著性测度的空间自适应正则化超分辨率图像重建
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166567
Zhenyu Liu, Jing Tian, Li Chen, Yongtao Wang
This paper addresses the super-resolution image reconstruction problem with the aim to produce a higher-resolution image based on its low-resolution counterparts. The proposed approach adaptively adjusts the degree of regularization using the saliency measure of the local content of the image. This is in contrast to that a spatially-invariant regularization is used for the whole image in conventional approaches. Furthermore, a gradient-based assessment criterion is proposed to measure the saliency of the image. Experiments are conducted to demonstrate the superior performance of the proposed approach.
本文解决了超分辨率图像重建问题,目的是在低分辨率图像的基础上产生更高分辨率的图像。该方法利用图像局部内容的显著性度量自适应调整正则化程度。这与传统方法中对整个图像使用空间不变正则化形成对比。在此基础上,提出了一种基于梯度的图像显著性评价准则。实验证明了该方法的优越性能。
{"title":"Spatially-adaptive regularized super-resolution image reconstruction using a gradient-based saliency measure","authors":"Zhenyu Liu, Jing Tian, Li Chen, Yongtao Wang","doi":"10.1109/ACPR.2011.6166567","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166567","url":null,"abstract":"This paper addresses the super-resolution image reconstruction problem with the aim to produce a higher-resolution image based on its low-resolution counterparts. The proposed approach adaptively adjusts the degree of regularization using the saliency measure of the local content of the image. This is in contrast to that a spatially-invariant regularization is used for the whole image in conventional approaches. Furthermore, a gradient-based assessment criterion is proposed to measure the saliency of the image. Experiments are conducted to demonstrate the superior performance of the proposed approach.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124391036","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel method of eliminating the background in Fourier transform profilometry based on Bi-dimensional Empirical Mode Decomposition 基于二维经验模态分解的傅里叶变换轮廓术背景消除新方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166670
Chenxing Wang, F. Da
To address the issue of spectrum overlapping in Fourier transform profilometry, a new method based on Bi-dimensional Empirical Mode Decomposition (BEMD) is proposed. BEMD is an adaptive data decomposition method, so it does not need filters or basic functions which are important for Fourier transform or wavelet transform. In this paper, the complicated original signal of distorted fringe pattern is decomposed into several Bi-dimensional Intrinsic Mode Functions (BIMFs) as well as the residual component, with which the background component and some other frequency noises of fringe pattern can be eliminated effectively. It is beneficial to extract the first frequency component exactly for the subsequent wrapped phase retrieval in Fourier transform. Simulation and experiments illustrate the feasibility and the exactness of the proposed method.
针对傅里叶变换轮廓术中存在的频谱重叠问题,提出了一种基于二维经验模态分解(BEMD)的新方法。BEMD是一种自适应数据分解方法,它不需要傅里叶变换和小波变换所需要的滤波器和基本函数。本文将复杂的畸变条纹图原始信号分解为若干个二维本征模态函数(bimf)和残差分量,利用残差分量可以有效地消除条纹图的背景分量和其他一些频率噪声。在傅里叶变换中准确提取第一频率分量有利于后续的包裹相位恢复。仿真和实验验证了该方法的可行性和准确性。
{"title":"A novel method of eliminating the background in Fourier transform profilometry based on Bi-dimensional Empirical Mode Decomposition","authors":"Chenxing Wang, F. Da","doi":"10.1109/ACPR.2011.6166670","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166670","url":null,"abstract":"To address the issue of spectrum overlapping in Fourier transform profilometry, a new method based on Bi-dimensional Empirical Mode Decomposition (BEMD) is proposed. BEMD is an adaptive data decomposition method, so it does not need filters or basic functions which are important for Fourier transform or wavelet transform. In this paper, the complicated original signal of distorted fringe pattern is decomposed into several Bi-dimensional Intrinsic Mode Functions (BIMFs) as well as the residual component, with which the background component and some other frequency noises of fringe pattern can be eliminated effectively. It is beneficial to extract the first frequency component exactly for the subsequent wrapped phase retrieval in Fourier transform. Simulation and experiments illustrate the feasibility and the exactness of the proposed method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124409902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Drug-taking instruments recognition 吸毒工具识别
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166575
Ruiguang Hu, Nianhua Xie, Weiming Hu
In this paper we propose an algorithm for the recognition of three kinds of drug-taking instruments, including bongs, hookahs and spoons. A global feature - Pyramid of Histograms of Orientation Gradients (PHOG) - is used to represent images. PHOG is calculated by partitioning an image into increasingly fine sub-regions and concatenating the appropriately weighted histograms of orientation gradients of each sub-region at each level. Then, different classifiers can be employed to handle this recognition problem. In our experiments, Support Vector Machines (SVM) with five different kernels and Random Forest are evaluated for our application and SVM with χ2 kernel shows the best performance. We also compare our method with the standard Bag-of-Words (BOW) model using SIFT features. Experimental results demonstrate that in our application, directly using appropriate global feature (PHOG) is better than using local feature (SIFT) and BOW model in both performance and complexity.
本文提出了一种识别三种吸毒工具的算法,包括烟斗、水烟和勺子。一个全局特征-方向梯度直方图金字塔(PHOG) -被用来表示图像。PHOG的计算方法是将图像划分为越来越精细的子区域,并将每个子区域在每个级别上的方向梯度的适当加权直方图连接起来。然后,可以使用不同的分类器来处理这个识别问题。在我们的实验中,我们评估了五种不同核的支持向量机(SVM)和随机森林的应用,其中带有χ2核的支持向量机(SVM)表现出最好的性能。我们还将我们的方法与使用SIFT特征的标准词袋(BOW)模型进行了比较。实验结果表明,在我们的应用中,直接使用适当的全局特征(PHOG)在性能和复杂度上都优于使用局部特征(SIFT)和BOW模型。
{"title":"Drug-taking instruments recognition","authors":"Ruiguang Hu, Nianhua Xie, Weiming Hu","doi":"10.1109/ACPR.2011.6166575","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166575","url":null,"abstract":"In this paper we propose an algorithm for the recognition of three kinds of drug-taking instruments, including bongs, hookahs and spoons. A global feature - Pyramid of Histograms of Orientation Gradients (PHOG) - is used to represent images. PHOG is calculated by partitioning an image into increasingly fine sub-regions and concatenating the appropriately weighted histograms of orientation gradients of each sub-region at each level. Then, different classifiers can be employed to handle this recognition problem. In our experiments, Support Vector Machines (SVM) with five different kernels and Random Forest are evaluated for our application and SVM with χ2 kernel shows the best performance. We also compare our method with the standard Bag-of-Words (BOW) model using SIFT features. Experimental results demonstrate that in our application, directly using appropriate global feature (PHOG) is better than using local feature (SIFT) and BOW model in both performance and complexity.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124450076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image-based building reconstruction with Manhattan-world assumption 基于曼哈顿世界假设的图像建筑重建
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6892193
Ruiling Deng, Gang Zeng, Rui Gan, H. Zha
The 3D reconstruction of buildings is a challenging research problem especially for image-based methods due to the absence of textured surfaces and difficulty in detecting high-level architectural structures. In this paper, we present an image-based reconstruction algorithm for efficiently modeling of buildings with the Manhattan-world assumption. The first key component of the algorithm is a clustering of geometric primitives (e.g. stereo points and lines) into sparse planes in Manhattan-world coordinates. The combination of such clustered planes greatly limits the possibility of building models to be reconstructed. In the second stage, we employ the graph-cut minimization to obtain an optimal model based on an energy functional that embeds image consistency, surface smoothness and Manhattanworld constraints. Real world building reconstruction results demonstrate the efficiency of the proposed algorithm in handling large scale data and its robustness against the variety of architectural structures.
建筑物的三维重建是一个具有挑战性的研究问题,特别是基于图像的方法,由于缺乏纹理表面和难以检测高层建筑结构。在本文中,我们提出了一种基于图像的重建算法,以有效地模拟曼哈顿世界假设下的建筑物。该算法的第一个关键组件是将几何原语(例如立体点和线)聚类到曼哈顿世界坐标中的稀疏平面中。这种聚集平面的组合极大地限制了重建建筑模型的可能性。在第二阶段,我们采用图切最小化来获得基于能量函数的最优模型,该函数嵌入了图像一致性、表面光滑性和曼哈顿世界约束。真实世界的建筑重建结果证明了该算法在处理大规模数据方面的有效性以及对各种建筑结构的鲁棒性。
{"title":"Image-based building reconstruction with Manhattan-world assumption","authors":"Ruiling Deng, Gang Zeng, Rui Gan, H. Zha","doi":"10.1109/ACPR.2011.6892193","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6892193","url":null,"abstract":"The 3D reconstruction of buildings is a challenging research problem especially for image-based methods due to the absence of textured surfaces and difficulty in detecting high-level architectural structures. In this paper, we present an image-based reconstruction algorithm for efficiently modeling of buildings with the Manhattan-world assumption. The first key component of the algorithm is a clustering of geometric primitives (e.g. stereo points and lines) into sparse planes in Manhattan-world coordinates. The combination of such clustered planes greatly limits the possibility of building models to be reconstructed. In the second stage, we employ the graph-cut minimization to obtain an optimal model based on an energy functional that embeds image consistency, surface smoothness and Manhattanworld constraints. Real world building reconstruction results demonstrate the efficiency of the proposed algorithm in handling large scale data and its robustness against the variety of architectural structures.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122026537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1