首页 > 最新文献

2014 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Decomposable Nonlocal Tensor Dictionary Learning for Multispectral Image Denoising 多光谱图像去噪的可分解非局部张量字典学习
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.377
Yi Peng, Deyu Meng, Zongben Xu, Chenqiang Gao, Yi Yang, Biao Zhang
As compared to the conventional RGB or gray-scale images, multispectral images (MSI) can deliver more faithful representation for real scenes, and enhance the performance of many computer vision tasks. In practice, however, an MSI is always corrupted by various noises. In this paper we propose an effective MSI denoising approach by combinatorially considering two intrinsic characteristics underlying an MSI: the nonlocal similarity over space and the global correlation across spectrum. In specific, by explicitly considering spatial self-similarity of an MSI we construct a nonlocal tensor dictionary learning model with a group-block-sparsity constraint, which makes similar full-band patches (FBP) share the same atoms from the spatial and spectral dictionaries. Furthermore, through exploiting spectral correlation of an MSI and assuming over-redundancy of dictionaries, the constrained nonlocal MSI dictionary learning model can be decomposed into a series of unconstrained low-rank tensor approximation problems, which can be readily solved by off-the-shelf higher order statistics. Experimental results show that our method outperforms all state-of-the-art MSI denoising methods under comprehensive quantitative performance measures.
与传统的RGB或灰度图像相比,多光谱图像(MSI)可以更忠实地呈现真实场景,并提高许多计算机视觉任务的性能。然而,在实际应用中,微信号总是受到各种噪声的破坏。在本文中,我们提出了一种有效的MSI去噪方法,通过组合考虑MSI的两个内在特征:空间上的非局部相似性和频谱上的全局相关性。具体而言,通过明确考虑MSI的空间自相似性,我们构建了一个具有群块稀疏性约束的非局部张量字典学习模型,使相似的全带斑块(FBP)共享来自空间和光谱字典的相同原子。此外,通过利用MSI的谱相关性和假设字典的过冗余,约束非局部MSI字典学习模型可以分解为一系列无约束的低秩张量逼近问题,这些问题可以很容易地用现成的高阶统计量来解决。实验结果表明,在综合定量性能指标下,我们的方法优于所有最先进的MSI去噪方法。
{"title":"Decomposable Nonlocal Tensor Dictionary Learning for Multispectral Image Denoising","authors":"Yi Peng, Deyu Meng, Zongben Xu, Chenqiang Gao, Yi Yang, Biao Zhang","doi":"10.1109/CVPR.2014.377","DOIUrl":"https://doi.org/10.1109/CVPR.2014.377","url":null,"abstract":"As compared to the conventional RGB or gray-scale images, multispectral images (MSI) can deliver more faithful representation for real scenes, and enhance the performance of many computer vision tasks. In practice, however, an MSI is always corrupted by various noises. In this paper we propose an effective MSI denoising approach by combinatorially considering two intrinsic characteristics underlying an MSI: the nonlocal similarity over space and the global correlation across spectrum. In specific, by explicitly considering spatial self-similarity of an MSI we construct a nonlocal tensor dictionary learning model with a group-block-sparsity constraint, which makes similar full-band patches (FBP) share the same atoms from the spatial and spectral dictionaries. Furthermore, through exploiting spectral correlation of an MSI and assuming over-redundancy of dictionaries, the constrained nonlocal MSI dictionary learning model can be decomposed into a series of unconstrained low-rank tensor approximation problems, which can be readily solved by off-the-shelf higher order statistics. Experimental results show that our method outperforms all state-of-the-art MSI denoising methods under comprehensive quantitative performance measures.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"279 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115900173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 298
3D-Aided Face Recognition Robust to Expression and Pose Variations 三维辅助人脸识别对表情和姿态变化的鲁棒性
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.245
Baptiste Chu, S. Romdhani, Liming Chen
Expression and pose variations are major challenges for reliable face recognition (FR) in 2D. In this paper, we aim to endow state of the art face recognition SDKs with robustness to facial expression variations and pose changes by using an extended 3D Morphable Model (3DMM) which isolates identity variations from those due to facial expressions. Specifically, given a probe with expression, a novel view of the face is generated where the pose is rectified and the expression neutralized. We present two methods of expression neutralization. The first one uses prior knowledge to infer the neutral expression image from an input image. The second method, specifically designed for verification, is based on the transfer of the gallery face expression to the probe. Experiments using rectified and neutralized view with a standard commercial FR SDK on two 2D face databases, namely Multi-PIE and AR, show significant performance improvement of the commercial SDK to deal with expression and pose variations and demonstrates the effectiveness of the proposed approach.
面部表情和姿态变化是实现二维人脸识别的主要挑战。在本文中,我们的目标是通过使用扩展的3D变形模型(3DMM)来赋予最先进的面部识别sdk对面部表情变化和姿势变化的鲁棒性,该模型将面部表情引起的身份变化与身份变化隔离开来。具体来说,给定一个带有表情的探针,就会生成一个新的面部视图,其中姿势被纠正,表情被中和。我们提出了两种表达中和的方法。第一种方法是利用先验知识从输入图像中推断出中性表情图像。第二种方法是专门为验证而设计的,是基于将画廊面部表情传递到探针的方法。在Multi-PIE和AR两个2D人脸数据库上使用标准商用FR SDK进行校正和中和视图的实验表明,商用SDK在处理表情和姿态变化方面的性能有显著提高,证明了所提方法的有效性。
{"title":"3D-Aided Face Recognition Robust to Expression and Pose Variations","authors":"Baptiste Chu, S. Romdhani, Liming Chen","doi":"10.1109/CVPR.2014.245","DOIUrl":"https://doi.org/10.1109/CVPR.2014.245","url":null,"abstract":"Expression and pose variations are major challenges for reliable face recognition (FR) in 2D. In this paper, we aim to endow state of the art face recognition SDKs with robustness to facial expression variations and pose changes by using an extended 3D Morphable Model (3DMM) which isolates identity variations from those due to facial expressions. Specifically, given a probe with expression, a novel view of the face is generated where the pose is rectified and the expression neutralized. We present two methods of expression neutralization. The first one uses prior knowledge to infer the neutral expression image from an input image. The second method, specifically designed for verification, is based on the transfer of the gallery face expression to the probe. Experiments using rectified and neutralized view with a standard commercial FR SDK on two 2D face databases, namely Multi-PIE and AR, show significant performance improvement of the commercial SDK to deal with expression and pose variations and demonstrates the effectiveness of the proposed approach.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124144669","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 88
Fine-Grained Visual Comparisons with Local Learning 与局部学习的细粒度视觉比较
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.32
Aron Yu, K. Grauman
Given two images, we want to predict which exhibits a particular visual attribute more than the other-even when the two images are quite similar. Existing relative attribute methods rely on global ranking functions; yet rarely will the visual cues relevant to a comparison be constant for all data, nor will humans' perception of the attribute necessarily permit a global ordering. To address these issues, we propose a local learning approach for fine-grained visual comparisons. Given a novel pair of images, we learn a local ranking model on the fly, using only analogous training comparisons. We show how to identify these analogous pairs using learned metrics. With results on three challenging datasets-including a large newly curated dataset for fine-grained comparisons-our method outperforms stateof-the-art methods for relative attribute prediction.
给定两幅图像,我们想要预测哪一幅比另一幅更具有特定的视觉属性——即使这两幅图像非常相似。现有的相对属性方法依赖于全局排序函数;然而,与比较相关的视觉线索很少对所有数据都是恒定的,人类对属性的感知也不一定允许全局排序。为了解决这些问题,我们提出了一种用于细粒度视觉比较的局部学习方法。给定一对新的图像,我们只使用类似的训练比较,在飞行中学习局部排名模型。我们将展示如何使用学习的度量来识别这些类似的对。通过对三个具有挑战性的数据集(包括用于细粒度比较的大型新整理数据集)的结果,我们的方法在相对属性预测方面优于最先进的方法。
{"title":"Fine-Grained Visual Comparisons with Local Learning","authors":"Aron Yu, K. Grauman","doi":"10.1109/CVPR.2014.32","DOIUrl":"https://doi.org/10.1109/CVPR.2014.32","url":null,"abstract":"Given two images, we want to predict which exhibits a particular visual attribute more than the other-even when the two images are quite similar. Existing relative attribute methods rely on global ranking functions; yet rarely will the visual cues relevant to a comparison be constant for all data, nor will humans' perception of the attribute necessarily permit a global ordering. To address these issues, we propose a local learning approach for fine-grained visual comparisons. Given a novel pair of images, we learn a local ranking model on the fly, using only analogous training comparisons. We show how to identify these analogous pairs using learned metrics. With results on three challenging datasets-including a large newly curated dataset for fine-grained comparisons-our method outperforms stateof-the-art methods for relative attribute prediction.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124167362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 458
A Learning-to-Rank Approach for Image Color Enhancement 一种图像颜色增强的学习排序方法
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.382
Jianzhou Yan, Stephen Lin, S. B. Kang, Xiaoou Tang
We present a machine-learned ranking approach for automatically enhancing the color of a photograph. Unlike previous techniques that train on pairs of images before and after adjustment by a human user, our method takes into account the intermediate steps taken in the enhancement process, which provide detailed information on the person's color preferences. To make use of this data, we formulate the color enhancement task as a learning-to-rank problem in which ordered pairs of images are used for training, and then various color enhancements of a novel input image can be evaluated from their corresponding rank values. From the parallels between the decision tree structures we use for ranking and the decisions made by a human during the editing process, we posit that breaking a full enhancement sequence into individual steps can facilitate training. Our experiments show that this approach compares well to existing methods for automatic color enhancement.
我们提出了一种机器学习排序方法来自动增强照片的颜色。与以前的技术不同,我们的方法是在人类用户调整前后对图像进行训练,我们的方法考虑了增强过程中采取的中间步骤,这些步骤提供了关于人的颜色偏好的详细信息。为了利用这些数据,我们将颜色增强任务制定为一个学习排序问题,其中使用有序的图像对进行训练,然后可以从其相应的秩值评估新输入图像的各种颜色增强。从我们用于排序的决策树结构与人类在编辑过程中做出的决策之间的相似之处来看,我们假设将完整的增强序列分解为单个步骤可以促进训练。实验表明,该方法与现有的自动色彩增强方法相比效果良好。
{"title":"A Learning-to-Rank Approach for Image Color Enhancement","authors":"Jianzhou Yan, Stephen Lin, S. B. Kang, Xiaoou Tang","doi":"10.1109/CVPR.2014.382","DOIUrl":"https://doi.org/10.1109/CVPR.2014.382","url":null,"abstract":"We present a machine-learned ranking approach for automatically enhancing the color of a photograph. Unlike previous techniques that train on pairs of images before and after adjustment by a human user, our method takes into account the intermediate steps taken in the enhancement process, which provide detailed information on the person's color preferences. To make use of this data, we formulate the color enhancement task as a learning-to-rank problem in which ordered pairs of images are used for training, and then various color enhancements of a novel input image can be evaluated from their corresponding rank values. From the parallels between the decision tree structures we use for ranking and the decisions made by a human during the editing process, we posit that breaking a full enhancement sequence into individual steps can facilitate training. Our experiments show that this approach compares well to existing methods for automatic color enhancement.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124401967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 70
Deblurring Text Images via L0-Regularized Intensity and Gradient Prior 通过l0正则化强度和梯度先验去模糊文本图像
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.371
Jin-shan Pan, Zhe Hu, Zhixun Su, Ming-Hsuan Yang
We propose a simple yet effective L0-regularized prior based on intensity and gradient for text image deblurring. The proposed image prior is motivated by observing distinct properties of text images. Based on this prior, we develop an efficient optimization method to generate reliable intermediate results for kernel estimation. The proposed method does not require any complex filtering strategies to select salient edges which are critical to the state-of-the-art deblurring algorithms. We discuss the relationship with other deblurring algorithms based on edge selection and provide insight on how to select salient edges in a more principled way. In the final latent image restoration step, we develop a simple method to remove artifacts and render better deblurred images. Experimental results demonstrate that the proposed algorithm performs favorably against the state-of-the-art text image deblurring methods. In addition, we show that the proposed method can be effectively applied to deblur low-illumination images.
我们提出了一个简单而有效的基于强度和梯度的l0正则化先验文本图像去模糊。所提出的图像先验是通过观察文本图像的不同属性来激发的。在此基础上,我们开发了一种高效的优化方法来生成可靠的核估计中间结果。该方法不需要任何复杂的滤波策略来选择对最先进的去模糊算法至关重要的显著边缘。我们讨论了与其他基于边缘选择的去模糊算法的关系,并提供了如何以更有原则的方式选择显著边缘的见解。在最后的潜在图像恢复步骤中,我们开发了一种简单的方法来去除伪影并呈现更好的去模糊图像。实验结果表明,该算法与现有的文本图像去模糊方法相比具有较好的效果。此外,我们还证明了该方法可以有效地应用于低照度图像的去模糊。
{"title":"Deblurring Text Images via L0-Regularized Intensity and Gradient Prior","authors":"Jin-shan Pan, Zhe Hu, Zhixun Su, Ming-Hsuan Yang","doi":"10.1109/CVPR.2014.371","DOIUrl":"https://doi.org/10.1109/CVPR.2014.371","url":null,"abstract":"We propose a simple yet effective L0-regularized prior based on intensity and gradient for text image deblurring. The proposed image prior is motivated by observing distinct properties of text images. Based on this prior, we develop an efficient optimization method to generate reliable intermediate results for kernel estimation. The proposed method does not require any complex filtering strategies to select salient edges which are critical to the state-of-the-art deblurring algorithms. We discuss the relationship with other deblurring algorithms based on edge selection and provide insight on how to select salient edges in a more principled way. In the final latent image restoration step, we develop a simple method to remove artifacts and render better deblurred images. Experimental results demonstrate that the proposed algorithm performs favorably against the state-of-the-art text image deblurring methods. In addition, we show that the proposed method can be effectively applied to deblur low-illumination images.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"276 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114484400","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 404
Fast Edge-Preserving PatchMatch for Large Displacement Optical Flow 大位移光流快速保边补丁匹配
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.452
Linchao Bao, Qingxiong Yang, Hailin Jin
We present a fast optical flow algorithm that can handle large displacement motions. Our algorithm is inspired by recent successes of local methods in visual correspondence searching as well as approximate nearest neighbor field algorithms. The main novelty is a fast randomized edge-preserving approximate nearest neighbor field algorithm which propagates self-similarity patterns in addition to offsets. Experimental results on public optical flow benchmarks show that our method is significantly faster than state-of-the-art methods without compromising on quality, especially when scenes contain large motions.
提出了一种处理大位移运动的快速光流算法。我们的算法的灵感来自于最近在视觉对应搜索中的局部方法和近似最近邻域算法的成功。主要的新颖之处在于一种快速的随机化边缘保持近似最近邻域算法,该算法除了传播偏移量外,还传播自相似模式。公共光流基准的实验结果表明,我们的方法在不影响质量的情况下明显快于最先进的方法,特别是当场景包含大运动时。
{"title":"Fast Edge-Preserving PatchMatch for Large Displacement Optical Flow","authors":"Linchao Bao, Qingxiong Yang, Hailin Jin","doi":"10.1109/CVPR.2014.452","DOIUrl":"https://doi.org/10.1109/CVPR.2014.452","url":null,"abstract":"We present a fast optical flow algorithm that can handle large displacement motions. Our algorithm is inspired by recent successes of local methods in visual correspondence searching as well as approximate nearest neighbor field algorithms. The main novelty is a fast randomized edge-preserving approximate nearest neighbor field algorithm which propagates self-similarity patterns in addition to offsets. Experimental results on public optical flow benchmarks show that our method is significantly faster than state-of-the-art methods without compromising on quality, especially when scenes contain large motions.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"87 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114620084","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Image Fusion with Local Spectral Consistency and Dynamic Gradient Sparsity 基于局部光谱一致性和动态梯度稀疏性的图像融合
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.347
Cheng Chen, Yeqing Li, W. Liu, Junzhou Huang
In this paper, we propose a novel method for image fusion from a high resolution panchromatic image and a low resolution multispectral image at the same geographical location. Different from previous methods, we do not make any assumption about the upsampled multispectral image, but only assume that the fused image after downsampling should be close to the original multispectral image. This is a severely ill-posed problem and a dynamic gradient sparsity penalty is thus proposed for regularization. Incorporating the intra- correlations of different bands, this penalty can effectively exploit the prior information (e.g. sharp boundaries) from the panchromatic image. A new convex optimization algorithm is proposed to efficiently solve this problem. Extensive experiments on four multispectral datasets demonstrate that the proposed method significantly outperforms the state-of-the-arts in terms of both spatial and spectral qualities.
本文提出了一种基于同一地理位置的高分辨率全色图像和低分辨率多光谱图像融合的新方法。与以往的方法不同,我们对上采样后的多光谱图像不做任何假设,只假设下采样后的融合图像与原始多光谱图像接近。这是一个严重不适定的问题,因此提出了一个动态梯度稀疏性惩罚来正则化。该惩罚结合了不同波段的内相关性,可以有效地利用全色图像的先验信息(如清晰的边界)。为了有效地解决这一问题,提出了一种新的凸优化算法。在四个多光谱数据集上进行的大量实验表明,该方法在空间和光谱质量方面都明显优于目前的方法。
{"title":"Image Fusion with Local Spectral Consistency and Dynamic Gradient Sparsity","authors":"Cheng Chen, Yeqing Li, W. Liu, Junzhou Huang","doi":"10.1109/CVPR.2014.347","DOIUrl":"https://doi.org/10.1109/CVPR.2014.347","url":null,"abstract":"In this paper, we propose a novel method for image fusion from a high resolution panchromatic image and a low resolution multispectral image at the same geographical location. Different from previous methods, we do not make any assumption about the upsampled multispectral image, but only assume that the fused image after downsampling should be close to the original multispectral image. This is a severely ill-posed problem and a dynamic gradient sparsity penalty is thus proposed for regularization. Incorporating the intra- correlations of different bands, this penalty can effectively exploit the prior information (e.g. sharp boundaries) from the panchromatic image. A new convex optimization algorithm is proposed to efficiently solve this problem. Extensive experiments on four multispectral datasets demonstrate that the proposed method significantly outperforms the state-of-the-arts in terms of both spatial and spectral qualities.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"180 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114848441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 127
RAPS: Robust and Efficient Automatic Construction of Person-Specific Deformable Models RAPS:鲁棒且高效的个人化变形模型自动建构
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.231
Christos Sagonas, Yannis Panagakis, S. Zafeiriou, M. Pantic
The construction of Facial Deformable Models (FDMs) is a very challenging computer vision problem, since the face is a highly deformable object and its appearance drastically changes under different poses, expressions, and illuminations. Although several methods for generic FDMs construction, have been proposed for facial landmark localization in still images, they are insufficient for tasks such as facial behaviour analysis and facial motion capture where perfect landmark localization is required. In this case, person-specific FDMs (PSMs) are mainly employed, requiring manual facial landmark annotation for each person and person-specific training. In this paper, a novel method for the automatic construction of PSMs is proposed. To this end, an orthonormal subspace which is suitable for facial image reconstruction is learnt. Next, to correct the fittings of a generic model, image congealing (i.e., batch image aliment) is performed by employing only the learnt orthonormal subspace. Finally, the corrected fittings are used to construct the PSM. The image congealing problem is solved by formulating a suitable sparsity regularized rank minimization problem. The proposed method outperforms the state-of-the art methods that is compared to, in terms of both landmark localization accuracy and computational time.
人脸变形模型(fdm)的构建是一个非常具有挑战性的计算机视觉问题,因为人脸是一个高度可变形的物体,它的外观在不同的姿势、表情和光照下会发生巨大的变化。虽然已经提出了几种通用fdm构建方法,用于静止图像中的面部地标定位,但它们不足以用于需要完美地标定位的面部行为分析和面部动作捕捉等任务。在这种情况下,主要使用针对个人的fdm (psm),需要对每个人进行手动面部地标标注和针对个人的培训。本文提出了一种新的psm自动构建方法。为此,学习了一种适合于人脸图像重建的正交子空间。接下来,为了纠正通用模型的拟合,通过仅使用学习到的正交子空间来执行图像凝结(即批图像填充)。最后,校正后的配件用于构建PSM。通过提出一个合适的稀疏正则化秩最小化问题来解决图像凝结问题。该方法在地标定位精度和计算时间方面都优于目前比较的方法。
{"title":"RAPS: Robust and Efficient Automatic Construction of Person-Specific Deformable Models","authors":"Christos Sagonas, Yannis Panagakis, S. Zafeiriou, M. Pantic","doi":"10.1109/CVPR.2014.231","DOIUrl":"https://doi.org/10.1109/CVPR.2014.231","url":null,"abstract":"The construction of Facial Deformable Models (FDMs) is a very challenging computer vision problem, since the face is a highly deformable object and its appearance drastically changes under different poses, expressions, and illuminations. Although several methods for generic FDMs construction, have been proposed for facial landmark localization in still images, they are insufficient for tasks such as facial behaviour analysis and facial motion capture where perfect landmark localization is required. In this case, person-specific FDMs (PSMs) are mainly employed, requiring manual facial landmark annotation for each person and person-specific training. In this paper, a novel method for the automatic construction of PSMs is proposed. To this end, an orthonormal subspace which is suitable for facial image reconstruction is learnt. Next, to correct the fittings of a generic model, image congealing (i.e., batch image aliment) is performed by employing only the learnt orthonormal subspace. Finally, the corrected fittings are used to construct the PSM. The image congealing problem is solved by formulating a suitable sparsity regularized rank minimization problem. The proposed method outperforms the state-of-the art methods that is compared to, in terms of both landmark localization accuracy and computational time.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116793940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 40
Dense Non-rigid Shape Correspondence Using Random Forests 使用随机森林的密集非刚性形状对应
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.532
E. Rodolà, S. R. Bulò, Thomas Windheuser, Matthias Vestner, D. Cremers
We propose a shape matching method that produces dense correspondences tuned to a specific class of shapes and deformations. In a scenario where this class is represented by a small set of example shapes, the proposed method learns a shape descriptor capturing the variability of the deformations in the given class. The approach enables the wave kernel signature to extend the class of recognized deformations from near isometries to the deformations appearing in the example set by means of a random forest classifier. With the help of the introduced spatial regularization, the proposed method achieves significant improvements over the baseline approach and obtains state-of-the-art results while keeping short computation times.
我们提出了一种形状匹配方法,该方法可以产生针对特定类型形状和变形的密集对应。在该类由一小组示例形状表示的场景中,所提出的方法学习捕获给定类中变形的可变性的形状描述符。该方法使波核签名能够将识别变形的类别从近等距扩展到通过随机森林分类器出现在示例集中的变形。在引入空间正则化的帮助下,该方法在保持较短的计算时间的同时取得了较好的结果。
{"title":"Dense Non-rigid Shape Correspondence Using Random Forests","authors":"E. Rodolà, S. R. Bulò, Thomas Windheuser, Matthias Vestner, D. Cremers","doi":"10.1109/CVPR.2014.532","DOIUrl":"https://doi.org/10.1109/CVPR.2014.532","url":null,"abstract":"We propose a shape matching method that produces dense correspondences tuned to a specific class of shapes and deformations. In a scenario where this class is represented by a small set of example shapes, the proposed method learns a shape descriptor capturing the variability of the deformations in the given class. The approach enables the wave kernel signature to extend the class of recognized deformations from near isometries to the deformations appearing in the example set by means of a random forest classifier. With the help of the introduced spatial regularization, the proposed method achieves significant improvements over the baseline approach and obtains state-of-the-art results while keeping short computation times.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124061973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 167
Socially-Aware Large-Scale Crowd Forecasting 具有社会意识的大规模人群预测
Pub Date : 2014-06-23 DOI: 10.1109/CVPR.2014.283
Alexandre Alahi, Vignesh Ramanathan, Li Fei-Fei
In crowded spaces such as city centers or train stations, human mobility looks complex, but is often influenced only by a few causes. We propose to quantitatively study crowded environments by introducing a dataset of 42 million trajectories collected in train stations. Given this dataset, we address the problem of forecasting pedestrians' destinations, a central problem in understanding large-scale crowd mobility. We need to overcome the challenges posed by a limited number of observations (e.g. sparse cameras), and change in pedestrian appearance cues across different cameras. In addition, we often have restrictions in the way pedestrians can move in a scene, encoded as priors over origin and destination (OD) preferences. We propose a new descriptor coined as Social Affinity Maps (SAM) to link broken or unobserved trajectories of individuals in the crowd, while using the OD-prior in our framework. Our experiments show improvement in performance through the use of SAM features and OD prior. To the best of our knowledge, our work is one of the first studies that provides encouraging results towards a better understanding of crowd behavior at the scale of million pedestrians.
在城市中心或火车站等拥挤的空间,人类的流动性看起来很复杂,但往往只受少数几个原因的影响。我们建议通过引入在火车站收集的4200万个轨迹数据集来定量研究拥挤环境。基于这个数据集,我们解决了预测行人目的地的问题,这是理解大规模人群流动的核心问题。我们需要克服有限数量的观测(例如稀疏的摄像机)所带来的挑战,以及不同摄像机之间行人外观线索的变化。此外,我们通常会限制行人在场景中的移动方式,将其编码为优先于原点和目的地(OD)偏好的先验。我们提出了一个新的描述符,称为社会亲和力图(Social Affinity Maps, SAM),以连接人群中破碎或未观察到的个体轨迹,同时在我们的框架中使用OD-prior。我们的实验表明,通过使用SAM特征和OD先验,性能得到了改善。据我们所知,我们的工作是第一批为更好地理解百万行人规模的人群行为提供令人鼓舞的结果的研究之一。
{"title":"Socially-Aware Large-Scale Crowd Forecasting","authors":"Alexandre Alahi, Vignesh Ramanathan, Li Fei-Fei","doi":"10.1109/CVPR.2014.283","DOIUrl":"https://doi.org/10.1109/CVPR.2014.283","url":null,"abstract":"In crowded spaces such as city centers or train stations, human mobility looks complex, but is often influenced only by a few causes. We propose to quantitatively study crowded environments by introducing a dataset of 42 million trajectories collected in train stations. Given this dataset, we address the problem of forecasting pedestrians' destinations, a central problem in understanding large-scale crowd mobility. We need to overcome the challenges posed by a limited number of observations (e.g. sparse cameras), and change in pedestrian appearance cues across different cameras. In addition, we often have restrictions in the way pedestrians can move in a scene, encoded as priors over origin and destination (OD) preferences. We propose a new descriptor coined as Social Affinity Maps (SAM) to link broken or unobserved trajectories of individuals in the crowd, while using the OD-prior in our framework. Our experiments show improvement in performance through the use of SAM features and OD prior. To the best of our knowledge, our work is one of the first studies that provides encouraging results towards a better understanding of crowd behavior at the scale of million pedestrians.","PeriodicalId":319578,"journal":{"name":"2014 IEEE Conference on Computer Vision and Pattern Recognition","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124062216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 183
期刊
2014 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1