首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
Geolocation based image annotation 基于地理位置的图像标注
Pub Date : 2011-12-01 DOI: 10.1109/ACPR.2011.6166619
Atsushi Shimada, H. Nagahara, R. Taniguchi, V. Charvillat
The growth of photo-sharing website such as Flickr and Picasa enables us to access the billions of images easily. Recent years, many researchers leverage such photo-sharing site to tackle the image annotation problem. The aim of the image annotation is to give a proper label to an unknown image. Generally, image features and label features are used to acquire the relationship between them. Meanwhile, we use not only such image and label features but also geolocation which indicate the information where the image was taken. We formulate the image annotation problem as two important issues; image-based labeling and label-based localization. The former issue is to estimate a proper label from a given image. The latter is the issue to estimate the location from the label. Our approach combine these two estimation strategies. We conducted some experiments and found that our approach outperformed the traditional approach.
Flickr和Picasa等照片分享网站的发展使我们能够轻松访问数十亿张图片。近年来,许多研究者利用这样的照片分享网站来解决图像标注问题。图像标注的目的是给未知图像一个合适的标签。通常使用图像特征和标签特征来获取它们之间的关系。同时,我们不仅利用图像和标签的特征,还利用地理位置来表示图像的拍摄位置信息。我们将图像标注问题表述为两个重要问题;基于图像的标注和基于标签的定位。前一个问题是从给定的图像中估计一个合适的标签。后者是根据标签估计位置的问题。我们的方法结合了这两种评估策略。我们进行了一些实验,发现我们的方法优于传统方法。
{"title":"Geolocation based image annotation","authors":"Atsushi Shimada, H. Nagahara, R. Taniguchi, V. Charvillat","doi":"10.1109/ACPR.2011.6166619","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166619","url":null,"abstract":"The growth of photo-sharing website such as Flickr and Picasa enables us to access the billions of images easily. Recent years, many researchers leverage such photo-sharing site to tackle the image annotation problem. The aim of the image annotation is to give a proper label to an unknown image. Generally, image features and label features are used to acquire the relationship between them. Meanwhile, we use not only such image and label features but also geolocation which indicate the information where the image was taken. We formulate the image annotation problem as two important issues; image-based labeling and label-based localization. The former issue is to estimate a proper label from a given image. The latter is the issue to estimate the location from the label. Our approach combine these two estimation strategies. We conducted some experiments and found that our approach outperformed the traditional approach.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115808084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Discriminant appearance weighting for action recognition 动作识别的判别外观加权
Pub Date : 2011-12-01 DOI: 10.1109/ACPR.2011.6166599
Tetsu Matsukawa, Takio Kurita
Extending popular histogram representations of local motion patterns, we present a novel weighted integration method based on an assumption that a motion importance should be changed by its appearance to obtain better recognition accuracies. The proposed integration method of motion and appearance patterns can weight information involving “what is moving” by discriminant way. The discriminant weights can be learned efficiently and naturally using two-dimensional fisher discriminant analysis (or, fisher weight maps) of co-occurrence matrices. Original fisher weight maps lose shift invariance of histogram features, while the proposed method preserves it. Experimental results on KTH human action dataset and UT-interaction dataset revealed the effectiveness of the proposed integration compared to naive integration methods of independent motion and appearance features and also other state-of-the-art methods.
我们扩展了流行的局部运动模式直方图表示,提出了一种新的加权积分方法,该方法基于一个假设,即运动的重要性应该随着其外观而改变,以获得更好的识别精度。所提出的运动和外观模式的整合方法可以通过判别的方式对涉及“什么在运动”的信息进行加权。利用共现矩阵的二维fisher判别分析(或fisher权值图)可以高效、自然地学习判别权值。原始的fisher权值图失去了直方图特征的平移不变性,而该方法保留了直方图特征的平移不变性。在KTH人类动作数据集和ut交互数据集上的实验结果表明,与独立运动和外观特征的朴素集成方法以及其他最先进的方法相比,所提出的集成方法是有效的。
{"title":"Discriminant appearance weighting for action recognition","authors":"Tetsu Matsukawa, Takio Kurita","doi":"10.1109/ACPR.2011.6166599","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166599","url":null,"abstract":"Extending popular histogram representations of local motion patterns, we present a novel weighted integration method based on an assumption that a motion importance should be changed by its appearance to obtain better recognition accuracies. The proposed integration method of motion and appearance patterns can weight information involving “what is moving” by discriminant way. The discriminant weights can be learned efficiently and naturally using two-dimensional fisher discriminant analysis (or, fisher weight maps) of co-occurrence matrices. Original fisher weight maps lose shift invariance of histogram features, while the proposed method preserves it. Experimental results on KTH human action dataset and UT-interaction dataset revealed the effectiveness of the proposed integration compared to naive integration methods of independent motion and appearance features and also other state-of-the-art methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"429 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123416905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Tree crown detection in high resolution optical images during the early growth stages of Eucalyptus plantations in Brazil 巴西桉树人工林生长早期高分辨率光学图像的树冠检测
Pub Date : 2011-11-29 DOI: 10.1109/ACPR.2011.6166666
Jia Zhou, C. Proisy, P. Couteron, X. Descombes, J. Zerubia, G. Maire, Y. Nouvellon
Individual tree detection methods are more and more present, and improve, in forestry and silviculture domains with the increasing availability of satellite metric imagery [2–7]. Automatic detection on these very high spatial resolution images aims to determine the tree positions and crown sizes. In this paper, we use a mathematical model based on marked point processes, which showed advantages w.r.t. several individual tree detection algorithms for plantations [2], to analyze an Eucalyptus plantation in Brazil, with 2 optical images acquired by the WorldView-2 satellite. A tentative detection simultaneously with 2 images of different dates (multi-date) has been tested for the first time, which estimates individual tree crown variation during these dates. While, for most current detection methods, only the static state of tree crowns at the moment of one image's acquisition is estimated. The relevance of detection is discussed considering the detection performance in tree localizations and crown sizes. Then, tree crown growth are deduced from detection results and compared with the expected dynamics of corresponding populations.
随着卫星公制图像可用性的提高,在林业和林业领域,单株树检测方法越来越多,并且得到了改进[2-7]。自动检测这些非常高的空间分辨率图像的目的是确定树的位置和树冠大小。在本文中,我们使用基于标记点过程的数学模型,利用WorldView-2卫星获取的2张光学图像,对巴西的桉树人工林进行了分析,该模型在几种单独的树木检测算法中显示出优势[2]。首次对2张不同日期(多日期)的图像同时进行了试探性检测,估计了这些日期内单个树冠的变化。然而,目前大多数检测方法只估计了一张图像采集时刻树冠的静态状态。考虑树定位和树冠大小的检测性能,讨论了检测的相关性。然后,根据检测结果推导出树冠生长,并与相应种群的预期动态进行比较。
{"title":"Tree crown detection in high resolution optical images during the early growth stages of Eucalyptus plantations in Brazil","authors":"Jia Zhou, C. Proisy, P. Couteron, X. Descombes, J. Zerubia, G. Maire, Y. Nouvellon","doi":"10.1109/ACPR.2011.6166666","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166666","url":null,"abstract":"Individual tree detection methods are more and more present, and improve, in forestry and silviculture domains with the increasing availability of satellite metric imagery [2–7]. Automatic detection on these very high spatial resolution images aims to determine the tree positions and crown sizes. In this paper, we use a mathematical model based on marked point processes, which showed advantages w.r.t. several individual tree detection algorithms for plantations [2], to analyze an Eucalyptus plantation in Brazil, with 2 optical images acquired by the WorldView-2 satellite. A tentative detection simultaneously with 2 images of different dates (multi-date) has been tested for the first time, which estimates individual tree crown variation during these dates. While, for most current detection methods, only the static state of tree crowns at the moment of one image's acquisition is estimated. The relevance of detection is discussed considering the detection performance in tree localizations and crown sizes. Then, tree crown growth are deduced from detection results and compared with the expected dynamics of corresponding populations.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126619552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Video Object Segmentation by Hierarchical Localized Classification of Regions 基于层次局部区域分类的视频目标分割
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166545
Chenguang Zhang, H. Ai
Video Object Segmentation (VOS) is to cut out a selected object from video sequences, where the main difficulties are shape deformation, appearance variations and background clutter. To cope with these difficulties, we propose a novel method, named as Hierarchical Localized Classification of Regions (HLCR). We suggest that appearance models as well as the spatial and temporal coherence between frames are the keys to break through bottleneck. Locally, in order to identify foreground regions, we propose to use Hierarchial Localized Classifiers, which organize regional features as decision trees. In global, we adopt Gaussian Mixture Color Models (GMMs). After integrating the local and global results into a probability mask, we can achieve the final segmentation result by graph cut. Experiments on various challenging video sequences demonstrate the efficiency and adaptability of the proposed method.
视频目标分割(VOS)是从视频序列中分割出选定的目标,其难点主要是形状变形、外观变化和背景杂波。为了解决这些困难,我们提出了一种新的方法,称为层次局部区域分类(HLCR)。我们认为外观模型以及帧间的时空一致性是突破瓶颈的关键。在局部,为了识别前景区域,我们建议使用层次局部分类器,将区域特征组织为决策树。在全局上,我们采用高斯混合颜色模型(GMMs)。将局部和全局结果整合成一个概率掩模后,通过图切得到最终的分割结果。在各种具有挑战性的视频序列上的实验证明了该方法的有效性和适应性。
{"title":"Video Object Segmentation by Hierarchical Localized Classification of Regions","authors":"Chenguang Zhang, H. Ai","doi":"10.1109/ACPR.2011.6166545","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166545","url":null,"abstract":"Video Object Segmentation (VOS) is to cut out a selected object from video sequences, where the main difficulties are shape deformation, appearance variations and background clutter. To cope with these difficulties, we propose a novel method, named as Hierarchical Localized Classification of Regions (HLCR). We suggest that appearance models as well as the spatial and temporal coherence between frames are the keys to break through bottleneck. Locally, in order to identify foreground regions, we propose to use Hierarchial Localized Classifiers, which organize regional features as decision trees. In global, we adopt Gaussian Mixture Color Models (GMMs). After integrating the local and global results into a probability mask, we can achieve the final segmentation result by graph cut. Experiments on various challenging video sequences demonstrate the efficiency and adaptability of the proposed method.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115212503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure-constrained distribution matching using quadratic programming and its application to pronunciation evaluation 二次规划结构约束分布匹配及其在发音评价中的应用
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166673
Y. Qiao, Masayuki Suzuki, N. Minematsu, K. Hirose
We proposed a structural representation of speech that is robust to speaker difference due to its transformation-invariant property in previous works, where we compared two speech structures by calculating the distance between two structural vectors, each composed of the lengths of a structure's edges. However, this distance cannot yield matching scores directly related to individual events (nodes) of the two structures. In spite of comparing structural vectors directly, this paper takes structures as constraints for optimal pattern matching. We derive the formulas of objective functions and constraint functions for optimization. Under assumptions of Gaussian and shared covariance matrices, we show that this optimal problem can be reduced to a quadratically constrained quadratic programming problem. To relieve the too strong invariance problem, we use a subspace decomposition method and perform the optimization in each subspace. We evaluate the proposed method on a task to assess the goodness of students' English pronunciation. Experimental results show that the proposed method achieves higher correlations with teachers' manual scores than compared methods.
在之前的工作中,我们提出了一种对说话者差异具有鲁棒性的语音结构表示,由于其变换不变性,我们通过计算两个结构向量之间的距离来比较两个语音结构,每个结构向量由结构边缘的长度组成。然而,这个距离不能产生与两个结构的单个事件(节点)直接相关的匹配分数。除了直接比较结构向量外,本文还将结构作为最优模式匹配的约束条件。导出了优化的目标函数和约束函数的表达式。在高斯矩阵和共享协方差矩阵的假设下,我们证明了这个最优问题可以简化为一个二次约束的二次规划问题。为了解决不变性太强的问题,我们采用子空间分解的方法,在每个子空间上进行优化。我们在一个评估学生英语发音好坏的任务中对所提出的方法进行了评估。实验结果表明,该方法与教师手工成绩的相关性高于其他方法。
{"title":"Structure-constrained distribution matching using quadratic programming and its application to pronunciation evaluation","authors":"Y. Qiao, Masayuki Suzuki, N. Minematsu, K. Hirose","doi":"10.1109/ACPR.2011.6166673","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166673","url":null,"abstract":"We proposed a structural representation of speech that is robust to speaker difference due to its transformation-invariant property in previous works, where we compared two speech structures by calculating the distance between two structural vectors, each composed of the lengths of a structure's edges. However, this distance cannot yield matching scores directly related to individual events (nodes) of the two structures. In spite of comparing structural vectors directly, this paper takes structures as constraints for optimal pattern matching. We derive the formulas of objective functions and constraint functions for optimization. Under assumptions of Gaussian and shared covariance matrices, we show that this optimal problem can be reduced to a quadratically constrained quadratic programming problem. To relieve the too strong invariance problem, we use a subspace decomposition method and perform the optimization in each subspace. We evaluate the proposed method on a task to assess the goodness of students' English pronunciation. Experimental results show that the proposed method achieves higher correlations with teachers' manual scores than compared methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117191479","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effective image representation based on bi-layer visual codebook 基于双层视觉码本的有效图像表示
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166534
Yan Song, Jinhui Tang, Xia Li, Q. Tian, Lirong Dai
Recently, the Bag-of-visual Words (BoW) based image representation has drawn much attention in image categorization and retrieval applications. It is known that the visual codebook construction and the related quantization methods play the important roles in BoW model. Traditionally, visual codebook is generated by clustering local features into groups, and the original feature is hard quantized to its nearest centers. It is known that the quantization error may degrade the effectiveness of the BoW representation. To address this problem, several soft quantization based methods have been proposed in literature. However, the effectiveness of these methods is still unsatisfactory. In this paper, we propose a novel and effective image representation method based on a bi-layer codebook. In this method, we first construct the bi-layer codebook to explicitly reduce the quantization error. And then, inspired by the locality-constrained linear coding method[18], we propose a ridge regression based quantization to assign multiple visual words to the local feature. Furthermore, the k nearest neighbor strategy is integrated to improve the efficiency of quantization. To evaluate the proposed image representation, we compare it with the existing image representations on two benchmark datasets in the image classification experiments. The experimental results demonstrate the superiority over the state-of-the-art techniques.
近年来,基于视觉词袋(Bag-of-visual Words, BoW)的图像表示方法在图像分类和检索中得到了广泛的应用。可视化码本的构建及其量化方法在BoW模型中起着重要的作用。传统的视觉码本是通过将局部特征聚类成组来生成的,原始特征很难量化到最近的中心。众所周知,量化误差会降低BoW表示的有效性。为了解决这个问题,文献中提出了几种基于软量化的方法。然而,这些方法的有效性仍然令人不满意。本文提出了一种新颖有效的基于双层码本的图像表示方法。在该方法中,我们首先构造双层码本来显式地减小量化误差。然后,受位置约束线性编码方法[18]的启发,我们提出了一种基于脊回归的量化方法,将多个视觉词分配给局部特征。在此基础上,结合k近邻策略,提高量化效率。为了评估所提出的图像表示,我们将其与现有的图像表示在两个基准数据集上进行了图像分类实验。实验结果表明,该方法优于目前最先进的技术。
{"title":"Effective image representation based on bi-layer visual codebook","authors":"Yan Song, Jinhui Tang, Xia Li, Q. Tian, Lirong Dai","doi":"10.1109/ACPR.2011.6166534","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166534","url":null,"abstract":"Recently, the Bag-of-visual Words (BoW) based image representation has drawn much attention in image categorization and retrieval applications. It is known that the visual codebook construction and the related quantization methods play the important roles in BoW model. Traditionally, visual codebook is generated by clustering local features into groups, and the original feature is hard quantized to its nearest centers. It is known that the quantization error may degrade the effectiveness of the BoW representation. To address this problem, several soft quantization based methods have been proposed in literature. However, the effectiveness of these methods is still unsatisfactory. In this paper, we propose a novel and effective image representation method based on a bi-layer codebook. In this method, we first construct the bi-layer codebook to explicitly reduce the quantization error. And then, inspired by the locality-constrained linear coding method[18], we propose a ridge regression based quantization to assign multiple visual words to the local feature. Furthermore, the k nearest neighbor strategy is integrated to improve the efficiency of quantization. To evaluate the proposed image representation, we compare it with the existing image representations on two benchmark datasets in the image classification experiments. The experimental results demonstrate the superiority over the state-of-the-art techniques.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124803799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Low resolution facial image recognition via multiple kernel criterion 基于多核准则的低分辨率人脸图像识别
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166709
Chuan-Xian Ren, D. Dai, Hong Yan
Practical face recognition systems are sometimes confronted with low-resolution (LR) images. Most existing feature extraction algorithms aim to preserve relational structure among objects of the input space in a linear embedding space. However, it has been a consensus that such complex visual learning tasks will be well be solved by adopting multiple descriptors to more precisely characterize the data for improving performance. In this paper, we addresses the problem of matching LR and high-resolution images that are difficult for conventional methods in practice due to the lack of an efficient similarity measure, and a multiple kernel criterion (MKC) is proposed for LR face recognition without any super-resolution (SR) preprocessing. Different image descriptors including RsL2, LBP, Gradientface and IMED are considered as the multiple kernel generators and the Gaussian function is exploited as the distance induced kernel. MKC solves this problem by minimizing the inconsistency between the similarities captured by the multiple kernels, and the nonlinear objective function can be alternatively minimized by a constrained eigenvalue decomposition. Experiments on benchmark databases show that our MKC method indeed improves the recognition performance.
实际的人脸识别系统有时会遇到低分辨率(LR)图像。现有的特征提取算法大多是为了在线性嵌入空间中保持输入空间对象之间的关系结构。然而,人们一致认为,通过采用多个描述符来更精确地描述数据以提高性能,可以很好地解决这种复杂的视觉学习任务。本文针对传统方法由于缺乏有效的相似性度量而难以实现LR和高分辨率图像匹配的问题,提出了一种无需超分辨率预处理的LR人脸识别多核准则(MKC)。将RsL2、LBP、Gradientface和IMED等不同的图像描述符作为多核生成器,利用高斯函数作为距离诱导核。MKC通过最小化多个核捕获的相似性之间的不一致性来解决这个问题,并且可以通过约束特征值分解来交替最小化非线性目标函数。在基准数据库上的实验表明,MKC方法确实提高了识别性能。
{"title":"Low resolution facial image recognition via multiple kernel criterion","authors":"Chuan-Xian Ren, D. Dai, Hong Yan","doi":"10.1109/ACPR.2011.6166709","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166709","url":null,"abstract":"Practical face recognition systems are sometimes confronted with low-resolution (LR) images. Most existing feature extraction algorithms aim to preserve relational structure among objects of the input space in a linear embedding space. However, it has been a consensus that such complex visual learning tasks will be well be solved by adopting multiple descriptors to more precisely characterize the data for improving performance. In this paper, we addresses the problem of matching LR and high-resolution images that are difficult for conventional methods in practice due to the lack of an efficient similarity measure, and a multiple kernel criterion (MKC) is proposed for LR face recognition without any super-resolution (SR) preprocessing. Different image descriptors including RsL2, LBP, Gradientface and IMED are considered as the multiple kernel generators and the Gaussian function is exploited as the distance induced kernel. MKC solves this problem by minimizing the inconsistency between the similarities captured by the multiple kernels, and the nonlinear objective function can be alternatively minimized by a constrained eigenvalue decomposition. Experiments on benchmark databases show that our MKC method indeed improves the recognition performance.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"398 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123714913","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Counting pedestrians in crowded scenes with efficient sparse learning 基于高效稀疏学习的拥挤场景行人计数
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166650
M. Shimosaka, S. Masuda, R. Fukui, Taketoshi Mori, Tomomasa Sato
Counting pedestrians in crowded scenes provides powerful cues for several applications such as traffic, safety, and advertising analysis in urban areas. Recent research progress has shown that direct mapping from image statistics (e.g. area or texture histograms of people regions) to the number of pedestrians, also known as counting by regression, is a promise way of robust pedestrian counting. While leveraging arbitrary image features is encouraged in the counting by regression to improve the accuracy, this leads to risk of over-fitting issue. Furthermore, the most image statistics are sensitive to the way of foreground region segmentation. Hence, careful selection process on both segmentation and feature levels is needed. This paper presents an efficient sparse training method via LARS (Least Angle Regression) to achieve the selection process on both levels, which provides the both sparsity of Lasso and Group Lasso. The experimental results using synthetic and pedestrian counting dataset show that our method provides robust performance with reasonable training cost among the state of the art pedestrian counting methods.
在拥挤的场景中计算行人数量为城市地区的交通、安全和广告分析等多种应用提供了强大的线索。最近的研究进展表明,将图像统计(如人区域的面积或纹理直方图)直接映射到行人数量,也称为回归计数,是一种很有前途的稳健行人计数方法。虽然在回归计数中鼓励利用任意图像特征来提高准确性,但这会导致过度拟合问题的风险。此外,大多数图像统计量对前景区域分割的方式比较敏感。因此,需要在分割和特征级别上进行仔细的选择过程。本文提出了一种基于最小角度回归(LARS)的高效稀疏训练方法来实现两个层次的选择过程,同时提供了Lasso和Group Lasso的稀疏性。实验结果表明,该方法在现有行人计数方法中具有较好的鲁棒性和较好的训练成本。
{"title":"Counting pedestrians in crowded scenes with efficient sparse learning","authors":"M. Shimosaka, S. Masuda, R. Fukui, Taketoshi Mori, Tomomasa Sato","doi":"10.1109/ACPR.2011.6166650","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166650","url":null,"abstract":"Counting pedestrians in crowded scenes provides powerful cues for several applications such as traffic, safety, and advertising analysis in urban areas. Recent research progress has shown that direct mapping from image statistics (e.g. area or texture histograms of people regions) to the number of pedestrians, also known as counting by regression, is a promise way of robust pedestrian counting. While leveraging arbitrary image features is encouraged in the counting by regression to improve the accuracy, this leads to risk of over-fitting issue. Furthermore, the most image statistics are sensitive to the way of foreground region segmentation. Hence, careful selection process on both segmentation and feature levels is needed. This paper presents an efficient sparse training method via LARS (Least Angle Regression) to achieve the selection process on both levels, which provides the both sparsity of Lasso and Group Lasso. The experimental results using synthetic and pedestrian counting dataset show that our method provides robust performance with reasonable training cost among the state of the art pedestrian counting methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114598871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Probability of a unique crypto key generation based on finger's different images with two scanners 用两个扫描仪根据手指的不同图像生成唯一加密密钥的概率
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166527
Guoqiang Ma, Juan Liu, Bin Ni
In practical cases, one finger is always scanned with different scanners at different time, which generates different fingerprint images. The key issue of the fingerprint-based cryptosystem is how to realize a unique crypto key from a fingerprint feature data which may differ from scanner to scanner. In this paper, the main difference between the fingerprint key-generation cryptosystems and the fingerprint-based recognition system was studied and a crypto key generation scheme based on one fingerprint image was proposed. The experimental results showed that about forty percent fingers can generate a same crypto key even if the images are from different scanners. That is to say, a unique crypto key could be generated based on different fingerprint images of one finger even these images were scanned with different scanners. This means the efforts of future work should be the technology of image quality checking and key generation scheme.
在实际情况中,一个手指在不同的时间被不同的扫描仪扫描,从而产生不同的指纹图像。基于指纹的密码系统的关键问题是如何从不同扫描器的指纹特征数据中实现唯一的加密密钥。本文研究了指纹密钥生成密码系统与基于指纹的识别系统的主要区别,提出了一种基于单张指纹图像的密码密钥生成方案。实验结果表明,即使图像来自不同的扫描仪,也有大约40%的手指可以生成相同的加密密钥。也就是说,即使用不同的扫描仪扫描一个手指的不同指纹图像,也可以生成一个唯一的加密密钥。这意味着今后的工作重点应该是图像质量检测技术和密钥生成方案。
{"title":"Probability of a unique crypto key generation based on finger's different images with two scanners","authors":"Guoqiang Ma, Juan Liu, Bin Ni","doi":"10.1109/ACPR.2011.6166527","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166527","url":null,"abstract":"In practical cases, one finger is always scanned with different scanners at different time, which generates different fingerprint images. The key issue of the fingerprint-based cryptosystem is how to realize a unique crypto key from a fingerprint feature data which may differ from scanner to scanner. In this paper, the main difference between the fingerprint key-generation cryptosystems and the fingerprint-based recognition system was studied and a crypto key generation scheme based on one fingerprint image was proposed. The experimental results showed that about forty percent fingers can generate a same crypto key even if the images are from different scanners. That is to say, a unique crypto key could be generated based on different fingerprint images of one finger even these images were scanned with different scanners. This means the efforts of future work should be the technology of image quality checking and key generation scheme.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117300875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Global and local training for moving object classification in surveillance-oriented scene 面向监视场景中运动目标分类的全局和局部训练
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166561
Xin Zhao, Jianwei Ding, Kaiqi Huang, T. Tan
This paper presents a new training framework for multi-class moving object classification in surveillance-oriented scene. In many practical multi-class classification tasks, the instances are close to each other in the input feature space when they have similar features. These instances may have different class labels. Since the moving objects may have various view and shape, the above phenomenon is common in multi-class moving object classification. In our framework, firstly the input feature space is divided into several local clusters. Then, global training and local training are carried out sequential with an efficient online learning based algorithm. The induced global classifier is used to assign candidate instances to the most reliable clusters. Meanwhile, the trained local classifiers within those clusters can determine which classes the candidate instances belong to. Our experimental results illustrate the effectiveness of our method for moving object classification in surveillance-oriented scene.
本文提出了一种新的训练框架,用于面向监视场景的多类运动目标分类。在许多实际的多类分类任务中,当实例具有相似的特征时,它们在输入特征空间中彼此接近。这些实例可能有不同的类标签。由于运动对象可能具有多种视图和形状,因此上述现象在多类运动对象分类中很常见。在我们的框架中,首先将输入特征空间划分为几个局部聚类。然后,采用高效的在线学习算法,依次进行全局训练和局部训练。使用诱导全局分类器将候选实例分配到最可靠的聚类中。同时,在这些聚类中经过训练的局部分类器可以确定候选实例属于哪些类。实验结果表明,该方法在面向监视的场景中对运动目标进行分类是有效的。
{"title":"Global and local training for moving object classification in surveillance-oriented scene","authors":"Xin Zhao, Jianwei Ding, Kaiqi Huang, T. Tan","doi":"10.1109/ACPR.2011.6166561","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166561","url":null,"abstract":"This paper presents a new training framework for multi-class moving object classification in surveillance-oriented scene. In many practical multi-class classification tasks, the instances are close to each other in the input feature space when they have similar features. These instances may have different class labels. Since the moving objects may have various view and shape, the above phenomenon is common in multi-class moving object classification. In our framework, firstly the input feature space is divided into several local clusters. Then, global training and local training are carried out sequential with an efficient online learning based algorithm. The induced global classifier is used to assign candidate instances to the most reliable clusters. Meanwhile, the trained local classifiers within those clusters can determine which classes the candidate instances belong to. Our experimental results illustrate the effectiveness of our method for moving object classification in surveillance-oriented scene.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"13 Suppl 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128541488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1