首页 > 最新文献

The First Asian Conference on Pattern Recognition最新文献

英文 中文
Towards exaggerated image stereotypes 走向夸张的形象刻板印象
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166569
Cheng Chen, F. Lauze, C. Igel, Aasa Feragen, M. Loog, M. Nielsen
Given a training set of images and a binary classifier, we introduce the notion of an exaggerated image stereotype for some image class of interest, which emphasizes/exaggerates the characteristic patterns in an image and visualizes which visual information the classification relies on. This is useful for gaining insight into the classification mechanism. The exaggerated image stereotypes results in a proper trade-off between classification accuracy and likelihood of being generated from the class of interest. This is done by optimizing an objective function which consists of a discriminative term based on the classification result, and a generative term based on the assumption of the class distribution. We use this idea with Fisher's Linear Discriminant rule, and assume a multivariate normal distribution for samples within a class. The proposed framework has been applied on handwritten digit data, illustrating specific features differentiating digits. Then it is applied to a face dataset using Active Appearance Model (AAM), where male faces stereotypes are evolved from initial female faces.
给定一个图像训练集和一个二值分类器,我们为一些感兴趣的图像类别引入了夸张图像刻板印象的概念,它强调/夸大图像中的特征模式,并可视化分类所依赖的视觉信息。这对于深入了解分类机制非常有用。夸张的图像刻板印象导致分类准确性和从感兴趣的类生成的可能性之间的适当权衡。这是通过优化目标函数来实现的,该目标函数由基于分类结果的判别项和基于类分布假设的生成项组成。我们将这个想法与费雪的线性判别规则一起使用,并假设一个类内样本的多元正态分布。该框架已应用于手写数字数据,说明了区分数字的具体特征。然后使用主动外观模型(AAM)将其应用于人脸数据集,其中男性面孔的刻板印象是从最初的女性面孔演变而来的。
{"title":"Towards exaggerated image stereotypes","authors":"Cheng Chen, F. Lauze, C. Igel, Aasa Feragen, M. Loog, M. Nielsen","doi":"10.1109/ACPR.2011.6166569","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166569","url":null,"abstract":"Given a training set of images and a binary classifier, we introduce the notion of an exaggerated image stereotype for some image class of interest, which emphasizes/exaggerates the characteristic patterns in an image and visualizes which visual information the classification relies on. This is useful for gaining insight into the classification mechanism. The exaggerated image stereotypes results in a proper trade-off between classification accuracy and likelihood of being generated from the class of interest. This is done by optimizing an objective function which consists of a discriminative term based on the classification result, and a generative term based on the assumption of the class distribution. We use this idea with Fisher's Linear Discriminant rule, and assume a multivariate normal distribution for samples within a class. The proposed framework has been applied on handwritten digit data, illustrating specific features differentiating digits. Then it is applied to a face dataset using Active Appearance Model (AAM), where male faces stereotypes are evolved from initial female faces.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"272 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122837763","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Human centric object detection in highly crowded scenes 高度拥挤场景中以人为中心的目标检测
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166674
Genquan Duan, H. Ai, Takayoshi Yamashita, S. Lao
In this paper, we propose to detect human centric objects, including face, head shoulder, upper body, left body, right body and whole body, which can provide essential information to locate humans in highly crowed scenes. In the literature, the approaches to detect multi-class objects are either taking each class independently to learn and apply its classifier successively or taking all classes as a whole to learn individual classifier based on sharing features and to detect by step-by-step dividing. Different from these works, we consider two issues, one is the similarities and discriminations of different classes and the other is the semantic relations among them. Our main idea is to predict class labels quickly using a Salient Patch Model (SPM) first, and then do detection accurately using detectors of predicted classes in which a Semantic Relation Model (SRM) is proposed to capture relations among classes for efficient inferences. SPM and SRM are designed for these two issues respectively. Experiments on challenging real-world datasets demonstrate that our proposed approach can achieve significant performance improvements.
本文提出对以人为中心的物体进行检测,包括人脸、头肩、上半身、左身体、右身体和全身,可以为在高度拥挤的场景中定位人类提供必要的信息。在文献中,多类物体检测的方法是:将每一类单独学习并依次应用其分类器,或将所有类作为一个整体,根据共享特征学习单个分类器,分步进行检测。与这些作品不同的是,我们考虑了两个问题,一个是不同类别之间的相似和区别,另一个是它们之间的语义关系。我们的主要思想是首先使用显著补丁模型(SPM)快速预测类标签,然后使用预测类的检测器进行准确检测,其中提出了语义关系模型(SRM)来捕获类之间的关系以进行有效推断。SPM和SRM分别针对这两个问题设计。在具有挑战性的真实数据集上的实验表明,我们提出的方法可以实现显着的性能改进。
{"title":"Human centric object detection in highly crowded scenes","authors":"Genquan Duan, H. Ai, Takayoshi Yamashita, S. Lao","doi":"10.1109/ACPR.2011.6166674","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166674","url":null,"abstract":"In this paper, we propose to detect human centric objects, including face, head shoulder, upper body, left body, right body and whole body, which can provide essential information to locate humans in highly crowed scenes. In the literature, the approaches to detect multi-class objects are either taking each class independently to learn and apply its classifier successively or taking all classes as a whole to learn individual classifier based on sharing features and to detect by step-by-step dividing. Different from these works, we consider two issues, one is the similarities and discriminations of different classes and the other is the semantic relations among them. Our main idea is to predict class labels quickly using a Salient Patch Model (SPM) first, and then do detection accurately using detectors of predicted classes in which a Semantic Relation Model (SRM) is proposed to capture relations among classes for efficient inferences. SPM and SRM are designed for these two issues respectively. Experiments on challenging real-world datasets demonstrate that our proposed approach can achieve significant performance improvements.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"113 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124747145","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quadratic-chi similarity metric learning for histogram feature 直方图特征的二次chi相似度度量学习
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166698
Xinyuan Cai, Baihua Xiao, Chunheng Wang, Rongguo Zhang
Histogram features, such as SIFT, HOG, LBP et al, are widely used in modern computer vision algorithms. According to [18], chi-square distance is an effective measure for comparing histogram features. In this paper, we propose a new method, named the Quadric-chi similarity metric learning (QCSML) for histogram features. The main contribution of this paper is that we propose a new metric learning method based on chi-square distance, in contrast with traditional Mahalanobis distance metric learning methods. The use of quadric-chi similarity in our method leads to an effective learning algorithm. Our method is tested on SIFT features for face identification, and compared with the state-of-art metric learning method (LDML) on the benchmark dataset, the Labeled Faces in the Wild (LFW). Experimental results show that our method can achieve clear performance gains over LDML.
直方图特征在现代计算机视觉算法中得到了广泛的应用,如SIFT、HOG、LBP等。根据[18],卡方距离是比较直方图特征的有效度量。在本文中,我们提出了一种新的直方图特征的相似度度量学习(QCSML)方法。本文的主要贡献在于,与传统的马氏距离度量学习方法相比,我们提出了一种新的基于卡方距离的度量学习方法。在我们的方法中使用二次chi相似度导致了一个有效的学习算法。我们的方法在SIFT特征上进行了人脸识别测试,并在基准数据集Labeled Faces in the Wild (LFW)上与最先进的度量学习方法(LDML)进行了比较。实验结果表明,与LDML相比,我们的方法可以获得明显的性能提升。
{"title":"Quadratic-chi similarity metric learning for histogram feature","authors":"Xinyuan Cai, Baihua Xiao, Chunheng Wang, Rongguo Zhang","doi":"10.1109/ACPR.2011.6166698","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166698","url":null,"abstract":"Histogram features, such as SIFT, HOG, LBP et al, are widely used in modern computer vision algorithms. According to [18], chi-square distance is an effective measure for comparing histogram features. In this paper, we propose a new method, named the Quadric-chi similarity metric learning (QCSML) for histogram features. The main contribution of this paper is that we propose a new metric learning method based on chi-square distance, in contrast with traditional Mahalanobis distance metric learning methods. The use of quadric-chi similarity in our method leads to an effective learning algorithm. Our method is tested on SIFT features for face identification, and compared with the state-of-art metric learning method (LDML) on the benchmark dataset, the Labeled Faces in the Wild (LFW). Experimental results show that our method can achieve clear performance gains over LDML.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124818472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A local learning based Image-To-Class distance for image classification 基于局部学习的图像到类距离分类方法
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166577
Xinyuan Cai, Baihua Xiao, Chunheng Wang, Rongguo Zhang
Image-To-Class distance is first proposed in Naive-Bayes Nearest-Neighbor. NBNN is a feature-based image classifier, and can achieve impressive classification accuracy. However, the performance of NBNN relies heavily on the large number of training samples. If using small number of training samples, the performance will degrade. The goal of this paper is to address this issue. The main contribution of this paper is that we propose a robust Image-to-Class distance by local learning. We define the patch-to-class distance as the distance between the input patch to its nearest neighbor in one class, which is reconstructed in the local manifold space; and then our image-to-class distance is the sum of patch-to-class distance. Furthermore, we take advantage of large-margin metric learning framework to obtain a proper Mahalanobis metric for each class. We evaluate the proposed method on four benchmark datasets: Caltech, Corel, Scene13, and Graz. The results show that our defined Image-To-Class Distance is more robust than NBNN and Optimal-NBNN, and by combining with the learned metric for each class, our method can achieve significant improvement over previous reported results on these datasets.
Image-To-Class距离最早是在Naive-Bayes最近邻算法中提出的。NBNN是一种基于特征的图像分类器,可以达到令人印象深刻的分类精度。然而,NBNN的性能很大程度上依赖于大量的训练样本。如果使用少量的训练样本,性能会下降。本文的目标就是解决这个问题。本文的主要贡献在于我们通过局部学习提出了一个鲁棒的图像到班级的距离。我们将patch到类的距离定义为输入patch到类中最近邻居的距离,该距离在局部流形空间中重构;然后图像到类的距离就是斑块到类的距离之和。此外,我们利用大余量度量学习框架为每个类别获得合适的马氏度规。我们在四个基准数据集上评估了所提出的方法:Caltech, Corel, Scene13和Graz。结果表明,我们定义的图像到类距离比NBNN和Optimal-NBNN具有更强的鲁棒性,并且通过结合每个类的学习度量,我们的方法可以在这些数据集上取得显著的改进。
{"title":"A local learning based Image-To-Class distance for image classification","authors":"Xinyuan Cai, Baihua Xiao, Chunheng Wang, Rongguo Zhang","doi":"10.1109/ACPR.2011.6166577","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166577","url":null,"abstract":"Image-To-Class distance is first proposed in Naive-Bayes Nearest-Neighbor. NBNN is a feature-based image classifier, and can achieve impressive classification accuracy. However, the performance of NBNN relies heavily on the large number of training samples. If using small number of training samples, the performance will degrade. The goal of this paper is to address this issue. The main contribution of this paper is that we propose a robust Image-to-Class distance by local learning. We define the patch-to-class distance as the distance between the input patch to its nearest neighbor in one class, which is reconstructed in the local manifold space; and then our image-to-class distance is the sum of patch-to-class distance. Furthermore, we take advantage of large-margin metric learning framework to obtain a proper Mahalanobis metric for each class. We evaluate the proposed method on four benchmark datasets: Caltech, Corel, Scene13, and Graz. The results show that our defined Image-To-Class Distance is more robust than NBNN and Optimal-NBNN, and by combining with the learned metric for each class, our method can achieve significant improvement over previous reported results on these datasets.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124944082","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Face recognition with continuous occlusion using partially iteratively reweighted sparse coding 基于部分迭代重加权稀疏编码的连续遮挡人脸识别
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166617
Xiao-Xin Li, D. Dai, Xiao-Fei Zhang, Chuan-Xian Ren
Partially occluded faces are common in automatic face recognition in the real world. Existing methods, such as sparse error correction with Markov random fields, correntropy-based sparse representation and robust sparse coding, are all based on error correction, which relies on the perfect reconstruction of the occluded facial image and limits their recognition rates especially when the occluded regions are large. It helps to enhance recognition rates if we can detect the occluded portions and exclude them from further classification. Based on a magnitude order measure, we propose an innovative effective occlusion detection algorithm, called Partially Iteratively Reweighted Sparse Coding (PIRSC). Compared to the state-of-the-art methods, our PIRSC based classifier greatly improve the face recognition rate especially when the occlusion percentage is large.
部分遮挡人脸是现实世界中人脸自动识别中常见的现象。现有的基于马尔可夫随机场的稀疏纠错、基于相关熵的稀疏表示和鲁棒稀疏编码等方法都是基于纠错的,这依赖于对被遮挡的面部图像的完美重建,并且限制了它们的识别率,特别是当被遮挡区域较大时。如果我们能够检测出遮挡部分并将其排除在进一步分类之外,将有助于提高识别率。基于数量级度量,我们提出了一种创新的有效遮挡检测算法,称为部分迭代重加权稀疏编码(PIRSC)。与现有的分类器相比,我们的基于PIRSC的分类器大大提高了人脸识别率,特别是在遮挡百分比较大的情况下。
{"title":"Face recognition with continuous occlusion using partially iteratively reweighted sparse coding","authors":"Xiao-Xin Li, D. Dai, Xiao-Fei Zhang, Chuan-Xian Ren","doi":"10.1109/ACPR.2011.6166617","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166617","url":null,"abstract":"Partially occluded faces are common in automatic face recognition in the real world. Existing methods, such as sparse error correction with Markov random fields, correntropy-based sparse representation and robust sparse coding, are all based on error correction, which relies on the perfect reconstruction of the occluded facial image and limits their recognition rates especially when the occluded regions are large. It helps to enhance recognition rates if we can detect the occluded portions and exclude them from further classification. Based on a magnitude order measure, we propose an innovative effective occlusion detection algorithm, called Partially Iteratively Reweighted Sparse Coding (PIRSC). Compared to the state-of-the-art methods, our PIRSC based classifier greatly improve the face recognition rate especially when the occlusion percentage is large.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122093256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Fusion of features and classifiers for off-line handwritten signature verification 特征与分类器融合的离线手写签名验证
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166701
Juan Hu, Youbin Chen
A method for writer-independent off-line handwritten signature verification based on grey level feature extraction and Real Adaboost algorithm is proposed. Firstly, both global and local features are used simultaneously. The texture information such as co-occurrence matrix and local binary pattern are analyzed and used as features. Secondly, Support Vector Machines (SVMs) and the squared Mahalanobis distance classifier are introduced. Finally, Real Adaboost algorithm is applied. Experiments on the public signature database GPDS Corpus show that our proposed method has achieved the FRR 5.64% and the FAR 5.37% which are the best so far compared with other published results.
提出了一种基于灰度特征提取和Real Adaboost算法的离线手写签名验证方法。首先,全局特征和局部特征同时使用。对共现矩阵和局部二值模式等纹理信息进行分析并作为特征。其次,介绍了支持向量机(svm)和马氏距离平方分类器。最后,应用Real Adaboost算法。在公共特征库GPDS语料库上进行的实验表明,与已有的研究结果相比,本文提出的方法的FRR为5.64%,FAR为5.37%,是目前为止最好的。
{"title":"Fusion of features and classifiers for off-line handwritten signature verification","authors":"Juan Hu, Youbin Chen","doi":"10.1109/ACPR.2011.6166701","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166701","url":null,"abstract":"A method for writer-independent off-line handwritten signature verification based on grey level feature extraction and Real Adaboost algorithm is proposed. Firstly, both global and local features are used simultaneously. The texture information such as co-occurrence matrix and local binary pattern are analyzed and used as features. Secondly, Support Vector Machines (SVMs) and the squared Mahalanobis distance classifier are introduced. Finally, Real Adaboost algorithm is applied. Experiments on the public signature database GPDS Corpus show that our proposed method has achieved the FRR 5.64% and the FAR 5.37% which are the best so far compared with other published results.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126536137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An improvment of weight scheme on adaBoost in the presence of noisy data adaBoost中存在噪声数据时权值方案的改进
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166557
Shihai Wang, Geng Li
The first strand of this research is concerned with the classification noise issue. Classification noise, (worry labeling), is a further consequence of the difficulties in accurately labeling the real training data. For efficient reduction of the negative influence produced by noisy samples, we propose a new weight scheme with a nonlinear model with the local proximity assumption for the Boosting algorithm. The effectiveness of our method has been evaluated by using a set of University of California Irvine Machine Learning Repository (UCI) [1] benchmarks. We report promising results.
本研究的第一部分涉及分类噪声问题。分类噪声(忧虑标注)是难以准确标注真实训练数据的进一步后果。为了有效地降低噪声样本产生的负面影响,我们提出了一种新的加权算法,该算法采用非线性模型和局部接近假设。我们的方法的有效性已经通过使用一组加州大学欧文分校机器学习存储库(UCI)[1]基准进行了评估。我们报告了有希望的结果。
{"title":"An improvment of weight scheme on adaBoost in the presence of noisy data","authors":"Shihai Wang, Geng Li","doi":"10.1109/ACPR.2011.6166557","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166557","url":null,"abstract":"The first strand of this research is concerned with the classification noise issue. Classification noise, (worry labeling), is a further consequence of the difficulties in accurately labeling the real training data. For efficient reduction of the negative influence produced by noisy samples, we propose a new weight scheme with a nonlinear model with the local proximity assumption for the Boosting algorithm. The effectiveness of our method has been evaluated by using a set of University of California Irvine Machine Learning Repository (UCI) [1] benchmarks. We report promising results.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126556984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
What is happening in a still picture? 静止画面中发生了什么?
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166555
Piji Li, Jun Ma
We consider the problem of generating concise sentences to describe still pictures automatically. We treat objects in images (nouns in sentences) as hidden information of actions (verbs). Therefore, the sentence generation problem can be transformed into action detection and scene classification problems. We employ Latent Multiple Kernel Learning (L-MKL) to learn the action detectors from “Exemplarlets”, and utilize MKL to learn the scene classifiers. The image features employed include distribution of edges, dense visual words and feature descriptors at different levels of spatial pyramid. For a new image we can detect the action using a sliding-window detector learnt via L-MKL, predict the scene the action happened in and build haction, scenei tuples. Finally, these tuples will be translated into concise sentences according to previously defined grammar template. We show both the classification and sentence generating results on our newly collected dataset of six actions as well as demonstrate improved performance over existing methods.
我们考虑了自动生成简明句子来描述静态图片的问题。我们将图像中的物体(句子中的名词)视为动作(动词)的隐藏信息。因此,句子生成问题可以转化为动作检测和场景分类问题。我们利用潜多核学习(L-MKL)从“Exemplarlets”中学习动作检测器,并利用潜多核学习学习场景分类器。采用的图像特征包括边缘分布、密集的视觉词和空间金字塔不同层次上的特征描述符。对于新图像,我们可以使用通过L-MKL学习的滑动窗口检测器来检测动作,预测动作发生的场景并构建动作、场景元组。最后,根据之前定义的语法模板将这些元组翻译成简明的句子。我们在新收集的六个动作数据集上展示了分类和句子生成结果,并展示了比现有方法更好的性能。
{"title":"What is happening in a still picture?","authors":"Piji Li, Jun Ma","doi":"10.1109/ACPR.2011.6166555","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166555","url":null,"abstract":"We consider the problem of generating concise sentences to describe still pictures automatically. We treat objects in images (nouns in sentences) as hidden information of actions (verbs). Therefore, the sentence generation problem can be transformed into action detection and scene classification problems. We employ Latent Multiple Kernel Learning (L-MKL) to learn the action detectors from “Exemplarlets”, and utilize MKL to learn the scene classifiers. The image features employed include distribution of edges, dense visual words and feature descriptors at different levels of spatial pyramid. For a new image we can detect the action using a sliding-window detector learnt via L-MKL, predict the scene the action happened in and build haction, scenei tuples. Finally, these tuples will be translated into concise sentences according to previously defined grammar template. We show both the classification and sentence generating results on our newly collected dataset of six actions as well as demonstrate improved performance over existing methods.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127903171","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Identical object segmentation through level sets with similarity constraint 基于相似度约束的水平集分割同一目标
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166609
Hongbin Xie, Gang Zeng, Rui Gan, H. Zha
Unsupervised identical object segmentation remains a challenging problem in vision research due to the difficulties in obtaining high-level structural knowledge about the scene. In this paper, we present an algorithm based on level set with a novel similarity constraint term for identical objects segmentation. The key component of the proposed algorithm is to embed the similarity constraint into curve evolution, where the evolving speed is high in regions of similar appearance and becomes low in areas with distinct contents. The algorithm starts with a pair of seed matches (e.g. SIFT) and evolve the small initial circle to form large similar regions under the similarity constraint. The similarity constraint is related to local alignment with assumption that the warp between identical objects is affine transformation. The right warp aligns the identical objects and promotes the similar regions growth. The alignment and expansion alternate until the curve reaches the boundaries of similar objects. Real experiments validates the efficiency and effectiveness of the proposed algorithm.
由于难以获得关于场景的高层次结构知识,无监督的同物分割一直是视觉研究中的一个难题。本文提出了一种基于水平集的相同目标分割算法,并提出了一种新的相似约束项。该算法的关键部分是将相似性约束嵌入到曲线演化中,即在外观相似的区域进化速度快,而在内容不同的区域进化速度慢。该算法从一对种子匹配(如SIFT)开始,在相似度约束下,将小的初始圆进化成大的相似区域。相似性约束与局部对齐有关,假设相同物体之间的翘曲是仿射变换。右曲使相同的物体对齐,并促进相似区域的生长。对齐和扩展交替进行,直到曲线到达相似物体的边界。实际实验验证了该算法的有效性和有效性。
{"title":"Identical object segmentation through level sets with similarity constraint","authors":"Hongbin Xie, Gang Zeng, Rui Gan, H. Zha","doi":"10.1109/ACPR.2011.6166609","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166609","url":null,"abstract":"Unsupervised identical object segmentation remains a challenging problem in vision research due to the difficulties in obtaining high-level structural knowledge about the scene. In this paper, we present an algorithm based on level set with a novel similarity constraint term for identical objects segmentation. The key component of the proposed algorithm is to embed the similarity constraint into curve evolution, where the evolving speed is high in regions of similar appearance and becomes low in areas with distinct contents. The algorithm starts with a pair of seed matches (e.g. SIFT) and evolve the small initial circle to form large similar regions under the similarity constraint. The similarity constraint is related to local alignment with assumption that the warp between identical objects is affine transformation. The right warp aligns the identical objects and promotes the similar regions growth. The alignment and expansion alternate until the curve reaches the boundaries of similar objects. Real experiments validates the efficiency and effectiveness of the proposed algorithm.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"88 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123523327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust hemorrhage detection in diabetic retinopathy image 糖尿病视网膜病变图像出血检测的鲁棒性
Pub Date : 2011-11-01 DOI: 10.1109/ACPR.2011.6166529
Dongbo Zhang, Xiong Li, Xingyu Shang, Yao Yi, Yaonan Wang
To improve the robust performance to detect hemorrhage lesions in diabetic retinopathy image, a background estimation and vessel exclusion based algorithm is proposed in this paper. Candidate hemorrhages are located by background estimation and Mahalanobis distance, and then on the basis of shape analysis, vessel exclusion is conducted to remove non hemorrhage pixels. Experiments results show that the performance of our method is effective to reduce the false negative results arise from inaccurate vessel structure.
为了提高糖尿病视网膜病变图像出血病灶检测的鲁棒性,提出了一种基于背景估计和血管排除的糖尿病视网膜病变图像检测算法。通过背景估计和马氏距离对候选出血点进行定位,然后在形状分析的基础上进行血管排除,去除非出血点。实验结果表明,该方法可以有效地降低因血管结构不准确而产生的假阴性结果。
{"title":"Robust hemorrhage detection in diabetic retinopathy image","authors":"Dongbo Zhang, Xiong Li, Xingyu Shang, Yao Yi, Yaonan Wang","doi":"10.1109/ACPR.2011.6166529","DOIUrl":"https://doi.org/10.1109/ACPR.2011.6166529","url":null,"abstract":"To improve the robust performance to detect hemorrhage lesions in diabetic retinopathy image, a background estimation and vessel exclusion based algorithm is proposed in this paper. Candidate hemorrhages are located by background estimation and Mahalanobis distance, and then on the basis of shape analysis, vessel exclusion is conducted to remove non hemorrhage pixels. Experiments results show that the performance of our method is effective to reduce the false negative results arise from inaccurate vessel structure.","PeriodicalId":287232,"journal":{"name":"The First Asian Conference on Pattern Recognition","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123835987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
The First Asian Conference on Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1