首页 > 最新文献

2013 IEEE International Conference on Computer Vision最新文献

英文 中文
Markov Network-Based Unified Classifier for Face Identification 基于马尔可夫网络的人脸识别统一分类器
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.245
Wonjun Hwang, Kyungshik Noh, Junmo Kim
We propose a novel unifying framework using a Markov network to learn the relationship between multiple classifiers in face recognition. We assume that we have several complementary classifiers and assign observation nodes to the features of a query image and hidden nodes to the features of gallery images. We connect each hidden node to its corresponding observation node and to the hidden nodes of other neighboring classifiers. For each observation-hidden node pair, we collect a set of gallery candidates that are most similar to the observation instance, and the relationship between the hidden nodes is captured in terms of the similarity matrix between the collected gallery images. Posterior probabilities in the hidden nodes are computed by the belief-propagation algorithm. The novelty of the proposed framework is the method that takes into account the classifier dependency using the results of each neighboring classifier. We present extensive results on two different evaluation protocols, known and unknown image variation tests, using three different databases, which shows that the proposed framework always leads to good accuracy in face recognition.
我们提出了一个新的统一框架,使用马尔可夫网络来学习人脸识别中多个分类器之间的关系。我们假设我们有几个互补的分类器,并将观察节点分配给查询图像的特征,将隐藏节点分配给图库图像的特征。我们将每个隐藏节点与其对应的观测节点和其他邻近分类器的隐藏节点连接起来。对于每个观测隐藏节点对,我们收集一组与观测实例最相似的候选图库,并根据收集的图库图像之间的相似性矩阵捕获隐藏节点之间的关系。隐藏节点的后验概率由信念传播算法计算。该框架的新颖之处在于,它使用每个相邻分类器的结果来考虑分类器的依赖性。我们使用三种不同的数据库对两种不同的评估方案,已知和未知图像变异测试进行了广泛的结果,表明所提出的框架始终具有良好的人脸识别准确性。
{"title":"Markov Network-Based Unified Classifier for Face Identification","authors":"Wonjun Hwang, Kyungshik Noh, Junmo Kim","doi":"10.1109/ICCV.2013.245","DOIUrl":"https://doi.org/10.1109/ICCV.2013.245","url":null,"abstract":"We propose a novel unifying framework using a Markov network to learn the relationship between multiple classifiers in face recognition. We assume that we have several complementary classifiers and assign observation nodes to the features of a query image and hidden nodes to the features of gallery images. We connect each hidden node to its corresponding observation node and to the hidden nodes of other neighboring classifiers. For each observation-hidden node pair, we collect a set of gallery candidates that are most similar to the observation instance, and the relationship between the hidden nodes is captured in terms of the similarity matrix between the collected gallery images. Posterior probabilities in the hidden nodes are computed by the belief-propagation algorithm. The novelty of the proposed framework is the method that takes into account the classifier dependency using the results of each neighboring classifier. We present extensive results on two different evaluation protocols, known and unknown image variation tests, using three different databases, which shows that the proposed framework always leads to good accuracy in face recognition.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"2008 1","pages":"1952-1959"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82534252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Probabilistic Elastic Part Model for Unsupervised Face Detector Adaptation 无监督人脸检测器自适应的概率弹性部分模型
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.103
Haoxiang Li, G. Hua, Zhe L. Lin, Jonathan Brandt, Jianchao Yang
We propose an unsupervised detector adaptation algorithm to adapt any offline trained face detector to a specific collection of images, and hence achieve better accuracy. The core of our detector adaptation algorithm is a probabilistic elastic part (PEP) model, which is offline trained with a set of face examples. It produces a statistically aligned part based face representation, namely the PEP representation. To adapt a general face detector to a collection of images, we compute the PEP representations of the candidate detections from the general face detector, and then train a discriminative classifier with the top positives and negatives. Then we re-rank all the candidate detections with this classifier. This way, a face detector tailored to the statistics of the specific image collection is adapted from the original detector. We present extensive results on three datasets with two state-of-the-art face detectors. The significant improvement of detection accuracy over these state of-the-art face detectors strongly demonstrates the efficacy of the proposed face detector adaptation algorithm.
我们提出了一种无监督检测器自适应算法,使任何离线训练的人脸检测器适应特定的图像集合,从而达到更好的准确性。我们的检测器自适应算法的核心是一个概率弹性部分(PEP)模型,该模型使用一组人脸样本进行离线训练。它产生一个基于统计对齐部分的人脸表示,即PEP表示。为了使通用人脸检测器适应一组图像,我们计算了通用人脸检测器中候选检测的PEP表示,然后用最上面的阳性和阴性训练一个判别分类器。然后我们用这个分类器对所有的候选检测重新排序。通过这种方式,根据特定图像收集的统计数据定制的人脸检测器是由原始检测器改编的。我们在三个数据集上展示了两个最先进的面部检测器的广泛结果。与现有的人脸检测器相比,检测精度的显著提高充分证明了所提出的人脸检测器自适应算法的有效性。
{"title":"Probabilistic Elastic Part Model for Unsupervised Face Detector Adaptation","authors":"Haoxiang Li, G. Hua, Zhe L. Lin, Jonathan Brandt, Jianchao Yang","doi":"10.1109/ICCV.2013.103","DOIUrl":"https://doi.org/10.1109/ICCV.2013.103","url":null,"abstract":"We propose an unsupervised detector adaptation algorithm to adapt any offline trained face detector to a specific collection of images, and hence achieve better accuracy. The core of our detector adaptation algorithm is a probabilistic elastic part (PEP) model, which is offline trained with a set of face examples. It produces a statistically aligned part based face representation, namely the PEP representation. To adapt a general face detector to a collection of images, we compute the PEP representations of the candidate detections from the general face detector, and then train a discriminative classifier with the top positives and negatives. Then we re-rank all the candidate detections with this classifier. This way, a face detector tailored to the statistics of the specific image collection is adapted from the original detector. We present extensive results on three datasets with two state-of-the-art face detectors. The significant improvement of detection accuracy over these state of-the-art face detectors strongly demonstrates the efficacy of the proposed face detector adaptation algorithm.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"22 1","pages":"793-800"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81683129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Modeling Occlusion by Discriminative AND-OR Structures 基于鉴别与或结构的遮挡建模
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.318
Bo Li, Wenze Hu, Tianfu Wu, Song-Chun Zhu
Occlusion presents a challenge for detecting objects in real world applications. To address this issue, this paper models object occlusion with an AND-OR structure which (i) represents occlusion at semantic part level, and (ii) captures the regularities of different occlusion configurations (i.e., the different combinations of object part visibilities). This paper focuses on car detection on street. Since annotating part occlusion on real images is time-consuming and error-prone, we propose to learn the the AND-OR structure automatically using synthetic images of CAD models placed at different relative positions. The model parameters are learned from real images under the latent structural SVM (LSSVM) framework. In inference, an efficient dynamic programming (DP) algorithm is utilized. In experiments, we test our method on both car detection and car view estimation. Experimental results show that (i) Our CAD simulation strategy is capable of generating occlusion patterns for real scenarios, (ii) The proposed AND-OR structure model is effective for modeling occlusions, which outperforms the deformable part-based model (DPM) DPM, voc5 in car detection on both our self-collected street parking dataset and the Pascal VOC 2007 car dataset pascal-voc-2007}, (iii) The learned model is on-par with the state-of-the-art methods on car view estimation tested on two public datasets.
遮挡对现实世界中的物体检测提出了挑战。为了解决这一问题,本文使用and - or结构对物体遮挡进行建模,该结构(i)表示语义部分级别的遮挡,(ii)捕获不同遮挡配置(即物体部分可见性的不同组合)的规律。本文主要研究道路上的车辆检测。由于在真实图像上标注局部遮挡非常耗时且容易出错,我们提出使用放置在不同相对位置的CAD模型合成图像自动学习and或结构。在潜在结构支持向量机(LSSVM)框架下,从真实图像中学习模型参数。在推理中,采用了一种高效的动态规划算法。在实验中,我们测试了我们的方法在汽车检测和汽车视图估计。实验结果表明:(1)我们的CAD仿真策略能够生成真实场景的遮挡模式;(2)我们提出的and - or结构模型对于遮挡建模是有效的,在我们的自收集的街道停车数据集和Pascal VOC 2007汽车数据集上,都优于基于变形零件的模型(DPM) DPM, voc5。(iii)学习的模型与在两个公共数据集上测试的最先进的汽车视图估计方法相当。
{"title":"Modeling Occlusion by Discriminative AND-OR Structures","authors":"Bo Li, Wenze Hu, Tianfu Wu, Song-Chun Zhu","doi":"10.1109/ICCV.2013.318","DOIUrl":"https://doi.org/10.1109/ICCV.2013.318","url":null,"abstract":"Occlusion presents a challenge for detecting objects in real world applications. To address this issue, this paper models object occlusion with an AND-OR structure which (i) represents occlusion at semantic part level, and (ii) captures the regularities of different occlusion configurations (i.e., the different combinations of object part visibilities). This paper focuses on car detection on street. Since annotating part occlusion on real images is time-consuming and error-prone, we propose to learn the the AND-OR structure automatically using synthetic images of CAD models placed at different relative positions. The model parameters are learned from real images under the latent structural SVM (LSSVM) framework. In inference, an efficient dynamic programming (DP) algorithm is utilized. In experiments, we test our method on both car detection and car view estimation. Experimental results show that (i) Our CAD simulation strategy is capable of generating occlusion patterns for real scenarios, (ii) The proposed AND-OR structure model is effective for modeling occlusions, which outperforms the deformable part-based model (DPM) DPM, voc5 in car detection on both our self-collected street parking dataset and the Pascal VOC 2007 car dataset pascal-voc-2007}, (iii) The learned model is on-par with the state-of-the-art methods on car view estimation tested on two public datasets.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"17 1","pages":"2560-2567"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81751385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Complementary Projection Hashing 互补投影哈希
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.39
Zhongming Jin, Yao Hu, Yuetan Lin, Debing Zhang, Shiding Lin, Deng Cai, Xuelong Li
Recently, hashing techniques have been widely applied to solve the approximate nearest neighbors search problem in many vision applications. Generally, these hashing approaches generate 2^c buckets, where c is the length of the hash code. A good hashing method should satisfy the following two requirements: 1) mapping the nearby data points into the same bucket or nearby (measured by the Hamming distance) buckets. 2) all the data points are evenly distributed among all the buckets. In this paper, we propose a novel algorithm named Complementary Projection Hashing (CPH) to find the optimal hashing functions which explicitly considers the above two requirements. Specifically, CPH aims at sequentially finding a series of hyper planes (hashing functions) which cross the sparse region of the data. At the same time, the data points are evenly distributed in the hyper cubes generated by these hyper planes. The experiments comparing with the state-of-the-art hashing methods demonstrate the effectiveness of the proposed method.
近年来,哈希技术在许多视觉应用中被广泛应用于解决近似近邻搜索问题。通常,这些哈希方法生成2^c个桶,其中c是哈希码的长度。一个好的哈希方法应该满足以下两个要求:1)将附近的数据点映射到相同的桶或附近的桶(通过汉明距离测量)。2)所有数据点均匀分布在所有桶中。在本文中,我们提出了一种新的算法,称为互补投影哈希(CPH),以寻找明确考虑上述两个要求的最优哈希函数。具体来说,CPH的目标是依次找到一系列跨越数据稀疏区域的超平面(哈希函数)。同时,数据点均匀分布在这些超平面生成的超立方体中。通过与现有哈希算法的对比实验,验证了该方法的有效性。
{"title":"Complementary Projection Hashing","authors":"Zhongming Jin, Yao Hu, Yuetan Lin, Debing Zhang, Shiding Lin, Deng Cai, Xuelong Li","doi":"10.1109/ICCV.2013.39","DOIUrl":"https://doi.org/10.1109/ICCV.2013.39","url":null,"abstract":"Recently, hashing techniques have been widely applied to solve the approximate nearest neighbors search problem in many vision applications. Generally, these hashing approaches generate 2^c buckets, where c is the length of the hash code. A good hashing method should satisfy the following two requirements: 1) mapping the nearby data points into the same bucket or nearby (measured by the Hamming distance) buckets. 2) all the data points are evenly distributed among all the buckets. In this paper, we propose a novel algorithm named Complementary Projection Hashing (CPH) to find the optimal hashing functions which explicitly considers the above two requirements. Specifically, CPH aims at sequentially finding a series of hyper planes (hashing functions) which cross the sparse region of the data. At the same time, the data points are evenly distributed in the hyper cubes generated by these hyper planes. The experiments comparing with the state-of-the-art hashing methods demonstrate the effectiveness of the proposed method.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"4 1","pages":"257-264"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78863047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 58
Pose-Free Facial Landmark Fitting via Optimized Part Mixtures and Cascaded Deformable Shape Model 基于优化零件混合和级联可变形形状模型的无姿态面部地标拟合
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.244
Xiang Yu, Junzhou Huang, Shaoting Zhang, Wang Yan, Dimitris N. Metaxas
This paper addresses the problem of facial landmark localization and tracking from a single camera. We present a two-stage cascaded deformable shape model to effectively and efficiently localize facial landmarks with large head pose variations. For face detection, we propose a group sparse learning method to automatically select the most salient facial landmarks. By introducing 3D face shape model, we use procrustes analysis to achieve pose-free facial landmark initialization. For deformation, the first step uses mean-shift local search with constrained local model to rapidly approach the global optimum. The second step uses component-wise active contours to discriminatively refine the subtle shape variation. Our framework can simultaneously handle face detection, pose-free landmark localization and tracking in real time. Extensive experiments are conducted on both laboratory environmental face databases and face-in-the-wild databases. All results demonstrate that our approach has certain advantages over state-of-the-art methods in handling pose variations.
本文研究了单摄像头下的人脸标记定位与跟踪问题。我们提出了一种两阶段级联的可变形形状模型,以有效地定位头部姿态变化较大的面部标志。在人脸检测方面,我们提出了一种组稀疏学习方法来自动选择最显著的人脸标志。通过引入三维脸型模型,利用procrustes分析实现无姿态面部地标初始化。对于变形,第一步采用约束局部模型的均值偏移局部搜索,快速逼近全局最优解。第二步使用组件智能活动轮廓来区分细化细微的形状变化。我们的框架可以同时处理人脸检测、无姿态地标定位和实时跟踪。在实验室环境人脸数据库和野外人脸数据库上进行了大量的实验。所有结果都表明,我们的方法在处理姿势变化方面比最先进的方法具有一定的优势。
{"title":"Pose-Free Facial Landmark Fitting via Optimized Part Mixtures and Cascaded Deformable Shape Model","authors":"Xiang Yu, Junzhou Huang, Shaoting Zhang, Wang Yan, Dimitris N. Metaxas","doi":"10.1109/ICCV.2013.244","DOIUrl":"https://doi.org/10.1109/ICCV.2013.244","url":null,"abstract":"This paper addresses the problem of facial landmark localization and tracking from a single camera. We present a two-stage cascaded deformable shape model to effectively and efficiently localize facial landmarks with large head pose variations. For face detection, we propose a group sparse learning method to automatically select the most salient facial landmarks. By introducing 3D face shape model, we use procrustes analysis to achieve pose-free facial landmark initialization. For deformation, the first step uses mean-shift local search with constrained local model to rapidly approach the global optimum. The second step uses component-wise active contours to discriminatively refine the subtle shape variation. Our framework can simultaneously handle face detection, pose-free landmark localization and tracking in real time. Extensive experiments are conducted on both laboratory environmental face databases and face-in-the-wild databases. All results demonstrate that our approach has certain advantages over state-of-the-art methods in handling pose variations.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"22 1","pages":"1944-1951"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87605666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 251
Semi-supervised Learning for Large Scale Image Cosegmentation 大规模图像共分割的半监督学习
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.56
Zhengxiang Wang, Rujie Liu
This paper introduces to use semi-supervised learning for large scale image co segmentation. Different from traditional unsupervised cosegmentation that does not use any segmentation ground truth, semi-supervised cosegmentation exploits the similarity from both the very limited training image foregrounds, as well as the common object shared between the large number of unsegmented images. This would be a much practical way to effectively co segment a large number of related images simultaneously, where previous unsupervised co segmentation work poorly due to the large variances in appearance between different images and the lack of segmentation ground truth for guidance in co segmentation. For semi-supervised co segmentation in large scale, we propose an effective method by minimizing an energy function, which consists of the inter-image distance, the intra-image distance and the balance term. We also propose an iterative updating algorithm to efficiently solve this energy function, which decomposes the original energy minimization problem into sub-problems, and updates each image alternatively to reduce the number of variables in each sub-problem for computation efficiency. Experiment results on iCoseg and Pascal VOC datasets show that the proposed co segmentation method can effectively co segment hundreds of images in less than one minute. And our semi-supervised co segmentation is able to outperform both unsupervised co segmentation as well as fully supervised single image segmentation, especially when the training data is limited.
介绍了将半监督学习用于大规模图像分割的方法。与不使用任何分割基础真理的传统无监督共分割不同,半监督共分割既利用了非常有限的训练图像前景的相似性,又利用了大量未分割图像之间共享的共同对象的相似性。这将是一种非常实用的方法,可以有效地同时对大量相关图像进行co分割,而以前的无监督co分割由于不同图像之间的外观差异很大而效果不佳,并且在co分割中缺乏分割基础真理作为指导。对于大规模的半监督协同分割,我们提出了一种有效的方法,即最小化由图像间距离、图像内距离和平衡项组成的能量函数。我们还提出了一种迭代更新算法来有效地求解该能量函数,该算法将原始能量最小化问题分解为子问题,并交替更新每个图像以减少每个子问题中的变量数量以提高计算效率。在iCoseg和Pascal VOC数据集上的实验结果表明,所提出的协同分割方法可以在不到1分钟的时间内有效地对数百张图像进行协同分割。我们的半监督协同分割能够优于无监督协同分割和完全监督单幅图像分割,特别是在训练数据有限的情况下。
{"title":"Semi-supervised Learning for Large Scale Image Cosegmentation","authors":"Zhengxiang Wang, Rujie Liu","doi":"10.1109/ICCV.2013.56","DOIUrl":"https://doi.org/10.1109/ICCV.2013.56","url":null,"abstract":"This paper introduces to use semi-supervised learning for large scale image co segmentation. Different from traditional unsupervised cosegmentation that does not use any segmentation ground truth, semi-supervised cosegmentation exploits the similarity from both the very limited training image foregrounds, as well as the common object shared between the large number of unsegmented images. This would be a much practical way to effectively co segment a large number of related images simultaneously, where previous unsupervised co segmentation work poorly due to the large variances in appearance between different images and the lack of segmentation ground truth for guidance in co segmentation. For semi-supervised co segmentation in large scale, we propose an effective method by minimizing an energy function, which consists of the inter-image distance, the intra-image distance and the balance term. We also propose an iterative updating algorithm to efficiently solve this energy function, which decomposes the original energy minimization problem into sub-problems, and updates each image alternatively to reduce the number of variables in each sub-problem for computation efficiency. Experiment results on iCoseg and Pascal VOC datasets show that the proposed co segmentation method can effectively co segment hundreds of images in less than one minute. And our semi-supervised co segmentation is able to outperform both unsupervised co segmentation as well as fully supervised single image segmentation, especially when the training data is limited.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"2020 1","pages":"393-400"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87830091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Exemplar-Based Graph Matching for Robust Facial Landmark Localization 基于样例的图像匹配鲁棒人脸地标定位
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.131
Feng Zhou, Jonathan Brandt, Zhe L. Lin
Localizing facial landmarks is a fundamental step in facial image analysis. However, the problem is still challenging due to the large variability in pose and appearance, and the existence of occlusions in real-world face images. In this paper, we present exemplar-based graph matching (EGM), a robust framework for facial landmark localization. Compared to conventional algorithms, EGM has three advantages: (1) an affine-invariant shape constraint is learned online from similar exemplars to better adapt to the test face, (2) the optimal landmark configuration can be directly obtained by solving a graph matching problem with the learned shape constraint, (3) the graph matching problem can be optimized efficiently by linear programming. To our best knowledge, this is the first attempt to apply a graph matching technique for facial landmark localization. Experiments on several challenging datasets demonstrate the advantages of EGM over state-of-the-art methods.
人脸特征点定位是人脸图像分析的基本步骤。然而,由于姿态和外观的巨大可变性以及现实世界人脸图像中存在的遮挡,该问题仍然具有挑战性。在本文中,我们提出了基于示例的图匹配(EGM),这是一种鲁棒的面部地标定位框架。与传统算法相比,EGM具有三个优点:(1)从相似样例中在线学习仿射不变形状约束,以更好地适应测试面;(2)利用学习到的形状约束求解图匹配问题,可直接获得最优地标配置;(3)通过线性规划对图匹配问题进行高效优化。据我们所知,这是第一次尝试将图匹配技术应用于面部地标定位。在几个具有挑战性的数据集上的实验证明了EGM优于最先进的方法。
{"title":"Exemplar-Based Graph Matching for Robust Facial Landmark Localization","authors":"Feng Zhou, Jonathan Brandt, Zhe L. Lin","doi":"10.1109/ICCV.2013.131","DOIUrl":"https://doi.org/10.1109/ICCV.2013.131","url":null,"abstract":"Localizing facial landmarks is a fundamental step in facial image analysis. However, the problem is still challenging due to the large variability in pose and appearance, and the existence of occlusions in real-world face images. In this paper, we present exemplar-based graph matching (EGM), a robust framework for facial landmark localization. Compared to conventional algorithms, EGM has three advantages: (1) an affine-invariant shape constraint is learned online from similar exemplars to better adapt to the test face, (2) the optimal landmark configuration can be directly obtained by solving a graph matching problem with the learned shape constraint, (3) the graph matching problem can be optimized efficiently by linear programming. To our best knowledge, this is the first attempt to apply a graph matching technique for facial landmark localization. Experiments on several challenging datasets demonstrate the advantages of EGM over state-of-the-art methods.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"33 1","pages":"1025-1032"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87919229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 112
A Simple Model for Intrinsic Image Decomposition with Depth Cues 一种基于深度线索的简单图像分解模型
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.37
Qifeng Chen, V. Koltun
We present a model for intrinsic decomposition of RGB-D images. Our approach analyzes a single RGB-D image and estimates albedo and shading fields that explain the input. To disambiguate the problem, our model estimates a number of components that jointly account for the reconstructed shading. By decomposing the shading field, we can build in assumptions about image formation that help distinguish reflectance variation from shading. These assumptions are expressed as simple nonlocal regularizers. We evaluate the model on real-world images and on a challenging synthetic dataset. The experimental results demonstrate that the presented approach outperforms prior models for intrinsic decomposition of RGB-D images.
提出了一种RGB-D图像的内禀分解模型。我们的方法分析单个RGB-D图像,并估计解释输入的反照率和阴影场。为了消除这个问题的歧义,我们的模型估计了一些共同构成重建阴影的成分。通过分解阴影场,我们可以建立关于图像形成的假设,帮助区分反射率变化和阴影。这些假设被表示为简单的非局部正则化。我们在真实世界的图像和具有挑战性的合成数据集上评估模型。实验结果表明,该方法在RGB-D图像的内禀分解方面优于现有模型。
{"title":"A Simple Model for Intrinsic Image Decomposition with Depth Cues","authors":"Qifeng Chen, V. Koltun","doi":"10.1109/ICCV.2013.37","DOIUrl":"https://doi.org/10.1109/ICCV.2013.37","url":null,"abstract":"We present a model for intrinsic decomposition of RGB-D images. Our approach analyzes a single RGB-D image and estimates albedo and shading fields that explain the input. To disambiguate the problem, our model estimates a number of components that jointly account for the reconstructed shading. By decomposing the shading field, we can build in assumptions about image formation that help distinguish reflectance variation from shading. These assumptions are expressed as simple nonlocal regularizers. We evaluate the model on real-world images and on a challenging synthetic dataset. The experimental results demonstrate that the presented approach outperforms prior models for intrinsic decomposition of RGB-D images.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"73 1","pages":"241-248"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86155707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 190
Semantic Segmentation without Annotating Segments 没有注释段的语义分割
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.271
W. Xia, Csaba Domokos, Jian Dong, L. Cheong, Shuicheng Yan
Numerous existing object segmentation frameworks commonly utilize the object bounding box as a prior. In this paper, we address semantic segmentation assuming that object bounding boxes are provided by object detectors, but no training data with annotated segments are available. Based on a set of segment hypotheses, we introduce a simple voting scheme to estimate shape guidance for each bounding box. The derived shape guidance is used in the subsequent graph-cut-based figure-ground segmentation. The final segmentation result is obtained by merging the segmentation results in the bounding boxes. We conduct an extensive analysis of the effect of object bounding box accuracy. Comprehensive experiments on both the challenging PASCAL VOC object segmentation dataset and GrabCut-50 image segmentation dataset show that the proposed approach achieves competitive results compared to previous detection or bounding box prior based methods, as well as other state-of-the-art semantic segmentation methods.
许多现有的对象分割框架通常使用对象边界框作为先验。在本文中,我们假设对象检测器提供了对象边界框来解决语义分割问题,但没有带注释段的训练数据可用。基于一组分段假设,我们引入了一种简单的投票方案来估计每个边界框的形状引导。导出的形状制导用于随后的基于图形切割的图形-地面分割。将边界框内的分割结果合并得到最终的分割结果。我们对物体边界盒精度的影响进行了广泛的分析。在具有挑战性的PASCAL VOC对象分割数据集和GrabCut-50图像分割数据集上进行的综合实验表明,与之前基于检测或边界盒先验的方法以及其他最先进的语义分割方法相比,所提出的方法取得了具有竞争力的结果。
{"title":"Semantic Segmentation without Annotating Segments","authors":"W. Xia, Csaba Domokos, Jian Dong, L. Cheong, Shuicheng Yan","doi":"10.1109/ICCV.2013.271","DOIUrl":"https://doi.org/10.1109/ICCV.2013.271","url":null,"abstract":"Numerous existing object segmentation frameworks commonly utilize the object bounding box as a prior. In this paper, we address semantic segmentation assuming that object bounding boxes are provided by object detectors, but no training data with annotated segments are available. Based on a set of segment hypotheses, we introduce a simple voting scheme to estimate shape guidance for each bounding box. The derived shape guidance is used in the subsequent graph-cut-based figure-ground segmentation. The final segmentation result is obtained by merging the segmentation results in the bounding boxes. We conduct an extensive analysis of the effect of object bounding box accuracy. Comprehensive experiments on both the challenging PASCAL VOC object segmentation dataset and GrabCut-50 image segmentation dataset show that the proposed approach achieves competitive results compared to previous detection or bounding box prior based methods, as well as other state-of-the-art semantic segmentation methods.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"49 1","pages":"2176-2183"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86455666","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Category-Independent Object-Level Saliency Detection 类别无关的对象级显著性检测
Pub Date : 2013-12-01 DOI: 10.1109/ICCV.2013.221
Yangqing Jia, Mei Han
It is known that purely low-level saliency cues such as frequency does not lead to a good salient object detection result, requiring high-level knowledge to be adopted for successful discovery of task-independent salient objects. In this paper, we propose an efficient way to combine such high-level saliency priors and low-level appearance models. We obtain the high-level saliency prior with the objectness algorithm to find potential object candidates without the need of category information, and then enforce the consistency among the salient regions using a Gaussian MRF with the weights scaled by diverse density that emphasizes the influence of potential foreground pixels. Our model obtains saliency maps that assign high scores for the whole salient object, and achieves state-of-the-art performance on benchmark datasets covering various foreground statistics.
众所周知,纯粹的低水平显著性线索(如频率)并不能带来良好的显著性对象检测结果,需要采用高水平的知识才能成功发现与任务无关的显著性对象。在本文中,我们提出了一种有效的方法来结合这些高级显着先验和低级外观模型。在不需要类别信息的情况下,通过对象性算法获得高水平的显著性先验来寻找潜在的候选对象,然后使用高斯MRF增强显著区域之间的一致性,该MRF的权重按不同密度缩放,强调潜在前景像素的影响。我们的模型获得了为整个显著性对象分配高分的显著性图,并在涵盖各种前景统计的基准数据集上实现了最先进的性能。
{"title":"Category-Independent Object-Level Saliency Detection","authors":"Yangqing Jia, Mei Han","doi":"10.1109/ICCV.2013.221","DOIUrl":"https://doi.org/10.1109/ICCV.2013.221","url":null,"abstract":"It is known that purely low-level saliency cues such as frequency does not lead to a good salient object detection result, requiring high-level knowledge to be adopted for successful discovery of task-independent salient objects. In this paper, we propose an efficient way to combine such high-level saliency priors and low-level appearance models. We obtain the high-level saliency prior with the objectness algorithm to find potential object candidates without the need of category information, and then enforce the consistency among the salient regions using a Gaussian MRF with the weights scaled by diverse density that emphasizes the influence of potential foreground pixels. Our model obtains saliency maps that assign high scores for the whole salient object, and achieves state-of-the-art performance on benchmark datasets covering various foreground statistics.","PeriodicalId":6351,"journal":{"name":"2013 IEEE International Conference on Computer Vision","volume":"83 1","pages":"1761-1768"},"PeriodicalIF":0.0,"publicationDate":"2013-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88812485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 133
期刊
2013 IEEE International Conference on Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1