首页 > 最新文献

2013 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Learning SURF Cascade for Fast and Accurate Object Detection 学习SURF级联快速准确的目标检测
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.445
Jianguo Li, Yimin Zhang
This paper presents a novel learning framework for training boosting cascade based object detector from large scale dataset. The framework is derived from the well-known Viola-Jones (VJ) framework but distinguished by three key differences. First, the proposed framework adopts multi-dimensional SURF features instead of single dimensional Haar features to describe local patches. In this way, the number of used local patches can be reduced from hundreds of thousands to several hundreds. Second, it adopts logistic regression as weak classifier for each local patch instead of decision trees in the VJ framework. Third, we adopt AUC as a single criterion for the convergence test during cascade training rather than the two trade-off criteria (false-positive-rate and hit-rate) in the VJ framework. The benefit is that the false-positive-rate can be adaptive among different cascade stages, and thus yields much faster convergence speed of SURF cascade. Combining these points together, the proposed approach has three good properties. First, the boosting cascade can be trained very efficiently. Experiments show that the proposed approach can train object detectors from billions of negative samples within one hour even on personal computers. Second, the built detector is comparable to the state-of-the-art algorithm not only on the accuracy but also on the processing speed. Third, the built detector is small in model-size due to short cascade stages.
本文提出了一种新的学习框架,用于训练基于大规模数据集的增强级联目标检测器。该框架源自著名的Viola-Jones (VJ)框架,但有三个关键区别。首先,采用多维SURF特征代替单维Haar特征来描述局部斑块;这样,使用的局部补丁数量可以从数十万个减少到数百个。其次,采用逻辑回归作为每个局部patch的弱分类器,而不是VJ框架中的决策树。第三,在级联训练过程中,我们采用AUC作为收敛检验的单一标准,而不是VJ框架中的两个权衡标准(假阳性率和命中率)。其优点是假阳性率可以在不同级联阶段之间自适应,从而使SURF级联的收敛速度更快。将这些点结合在一起,所提出的方法具有三个良好的性质。首先,增强级联可以非常有效地训练。实验表明,即使在个人电脑上,该方法也可以在一小时内从数十亿个负样本中训练出目标检测器。其次,构建的探测器不仅在精度上,而且在处理速度上与最先进的算法相当。第三,由于级联阶段短,所构建的探测器模型尺寸小。
{"title":"Learning SURF Cascade for Fast and Accurate Object Detection","authors":"Jianguo Li, Yimin Zhang","doi":"10.1109/CVPR.2013.445","DOIUrl":"https://doi.org/10.1109/CVPR.2013.445","url":null,"abstract":"This paper presents a novel learning framework for training boosting cascade based object detector from large scale dataset. The framework is derived from the well-known Viola-Jones (VJ) framework but distinguished by three key differences. First, the proposed framework adopts multi-dimensional SURF features instead of single dimensional Haar features to describe local patches. In this way, the number of used local patches can be reduced from hundreds of thousands to several hundreds. Second, it adopts logistic regression as weak classifier for each local patch instead of decision trees in the VJ framework. Third, we adopt AUC as a single criterion for the convergence test during cascade training rather than the two trade-off criteria (false-positive-rate and hit-rate) in the VJ framework. The benefit is that the false-positive-rate can be adaptive among different cascade stages, and thus yields much faster convergence speed of SURF cascade. Combining these points together, the proposed approach has three good properties. First, the boosting cascade can be trained very efficiently. Experiments show that the proposed approach can train object detectors from billions of negative samples within one hour even on personal computers. Second, the built detector is comparable to the state-of-the-art algorithm not only on the accuracy but also on the processing speed. Third, the built detector is small in model-size due to short cascade stages.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"15 1","pages":"3468-3475"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73272165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 205
Articulated and Restricted Motion Subspaces and Their Signatures 铰接式和受限运动子空间及其特征
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.198
Bastien Jacquet, Roland Angst, M. Pollefeys
Articulated objects represent an important class of objects in our everyday environment. Automatic detection of the type of articulated or otherwise restricted motion and extraction of the corresponding motion parameters are therefore of high value, eg in order to augment an otherwise static 3D reconstruction with dynamic semantics, such as rotation axes and allowable translation directions for certain rigid parts or objects. Hence, in this paper, a novel theory to analyse relative transformations between two motion-restricted parts will be presented. The analysis is based on linear subspaces spanned by relative transformations. Moreover, a signature for relative transformations will be introduced which uniquely specifies the type of restricted motion encoded in these relative transformations. This theoretic framework enables the derivation of novel algebraic constraints, such as low-rank constraints for subsequent rotations around two fixed axes for example. Lastly, given the type of restricted motion as predicted by the signature, the paper shows how to extract all the motion parameters with matrix manipulations from linear algebra. Our theory is verified on several real data sets, such as a rotating blackboard or a wheel rolling on the floor amongst others.
铰接对象代表了我们日常环境中一类重要的对象。因此,自动检测关节或其他受限运动的类型并提取相应的运动参数具有很高的价值,例如,为了增强具有动态语义的静态3D重建,例如某些刚性部件或对象的旋转轴和允许的平移方向。因此,本文将提出一种新的理论来分析两个运动受限部分之间的相对变换。该分析基于由相对变换张成的线性子空间。此外,将引入一个相对变换的签名,它唯一地指定在这些相对变换中编码的受限运动的类型。这一理论框架使得新的代数约束的推导成为可能,例如围绕两个固定轴的后续旋转的低秩约束。最后,根据特征预测的受限运动类型,给出了如何从线性代数中利用矩阵处理提取所有运动参数的方法。我们的理论在几个真实的数据集上得到了验证,例如旋转的黑板或在地板上滚动的轮子等等。
{"title":"Articulated and Restricted Motion Subspaces and Their Signatures","authors":"Bastien Jacquet, Roland Angst, M. Pollefeys","doi":"10.1109/CVPR.2013.198","DOIUrl":"https://doi.org/10.1109/CVPR.2013.198","url":null,"abstract":"Articulated objects represent an important class of objects in our everyday environment. Automatic detection of the type of articulated or otherwise restricted motion and extraction of the corresponding motion parameters are therefore of high value, eg in order to augment an otherwise static 3D reconstruction with dynamic semantics, such as rotation axes and allowable translation directions for certain rigid parts or objects. Hence, in this paper, a novel theory to analyse relative transformations between two motion-restricted parts will be presented. The analysis is based on linear subspaces spanned by relative transformations. Moreover, a signature for relative transformations will be introduced which uniquely specifies the type of restricted motion encoded in these relative transformations. This theoretic framework enables the derivation of novel algebraic constraints, such as low-rank constraints for subsequent rotations around two fixed axes for example. Lastly, given the type of restricted motion as predicted by the signature, the paper shows how to extract all the motion parameters with matrix manipulations from linear algebra. Our theory is verified on several real data sets, such as a rotating blackboard or a wheel rolling on the floor amongst others.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"1506-1513"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72607878","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Boundary Cues for 3D Object Shape Recovery 3D物体形状恢复的边界线索
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.281
Kevin Karsch, Zicheng Liao, Jason Rock, Jonathan T. Barron, Derek Hoiem
Early work in computer vision considered a host of geometric cues for both shape reconstruction and recognition. However, since then, the vision community has focused heavily on shading cues for reconstruction, and moved towards data-driven approaches for recognition. In this paper, we reconsider these perhaps overlooked "boundary" cues (such as self occlusions and folds in a surface), as well as many other established constraints for shape reconstruction. In a variety of user studies and quantitative tasks, we evaluate how well these cues inform shape reconstruction (relative to each other) in terms of both shape quality and shape recognition. Our findings suggest many new directions for future research in shape reconstruction, such as automatic boundary cue detection and relaxing assumptions in shape from shading (e.g. orthographic projection, Lambertian surfaces).
计算机视觉的早期工作考虑了许多用于形状重建和识别的几何线索。然而,从那时起,视觉界就把重点放在了重建的阴影线索上,并转向了数据驱动的识别方法。在本文中,我们重新考虑了这些可能被忽视的“边界”线索(如表面的自遮挡和折叠),以及许多其他已建立的形状重建约束。在各种用户研究和定量任务中,我们评估了这些线索在形状质量和形状识别方面对形状重建(相对于彼此)的影响程度。我们的研究结果为形状重建的未来研究提供了许多新的方向,如自动边界线索检测和放松阴影对形状的假设(如正交投影,兰伯曲面)。
{"title":"Boundary Cues for 3D Object Shape Recovery","authors":"Kevin Karsch, Zicheng Liao, Jason Rock, Jonathan T. Barron, Derek Hoiem","doi":"10.1109/CVPR.2013.281","DOIUrl":"https://doi.org/10.1109/CVPR.2013.281","url":null,"abstract":"Early work in computer vision considered a host of geometric cues for both shape reconstruction and recognition. However, since then, the vision community has focused heavily on shading cues for reconstruction, and moved towards data-driven approaches for recognition. In this paper, we reconsider these perhaps overlooked \"boundary\" cues (such as self occlusions and folds in a surface), as well as many other established constraints for shape reconstruction. In a variety of user studies and quantitative tasks, we evaluate how well these cues inform shape reconstruction (relative to each other) in terms of both shape quality and shape recognition. Our findings suggest many new directions for future research in shape reconstruction, such as automatic boundary cue detection and relaxing assumptions in shape from shading (e.g. orthographic projection, Lambertian surfaces).","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"49 1","pages":"2163-2170"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73395677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Learning the Change for Automatic Image Cropping 学习自动图像裁剪的变化
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.130
Jianzhou Yan, Stephen Lin, S. B. Kang, Xiaoou Tang
Image cropping is a common operation used to improve the visual quality of photographs. In this paper, we present an automatic cropping technique that accounts for the two primary considerations of people when they crop: removal of distracting content, and enhancement of overall composition. Our approach utilizes a large training set consisting of photos before and after cropping by expert photographers to learn how to evaluate these two factors in a crop. In contrast to the many methods that exist for general assessment of image quality, ours specifically examines differences between the original and cropped photo in solving for the crop parameters. To this end, several novel image features are proposed to model the changes in image content and composition when a crop is applied. Our experiments demonstrate improvements of our method over recent cropping algorithms on a broad range of images.
图像裁剪是一种常用的操作,用于提高照片的视觉质量。在本文中,我们提出了一种自动裁剪技术,该技术考虑了人们在裁剪时的两个主要考虑因素:去除分散注意力的内容,增强整体构图。我们的方法利用由专业摄影师裁剪前后的照片组成的大型训练集来学习如何评估裁剪中的这两个因素。与现有的许多用于图像质量一般评估的方法相比,我们的方法在求解裁剪参数时专门检查原始照片和裁剪照片之间的差异。为此,提出了几个新的图像特征来模拟图像内容和组成的变化,当一个裁剪应用。我们的实验证明了我们的方法在广泛的图像上比最近的裁剪算法有所改进。
{"title":"Learning the Change for Automatic Image Cropping","authors":"Jianzhou Yan, Stephen Lin, S. B. Kang, Xiaoou Tang","doi":"10.1109/CVPR.2013.130","DOIUrl":"https://doi.org/10.1109/CVPR.2013.130","url":null,"abstract":"Image cropping is a common operation used to improve the visual quality of photographs. In this paper, we present an automatic cropping technique that accounts for the two primary considerations of people when they crop: removal of distracting content, and enhancement of overall composition. Our approach utilizes a large training set consisting of photos before and after cropping by expert photographers to learn how to evaluate these two factors in a crop. In contrast to the many methods that exist for general assessment of image quality, ours specifically examines differences between the original and cropped photo in solving for the crop parameters. To this end, several novel image features are proposed to model the changes in image content and composition when a crop is applied. Our experiments demonstrate improvements of our method over recent cropping algorithms on a broad range of images.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"26 1","pages":"971-978"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79227569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 102
Beta Process Joint Dictionary Learning for Coupled Feature Spaces with Application to Single Image Super-Resolution 耦合特征空间的Beta过程联合字典学习及其在单幅超分辨率图像中的应用
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.51
Li He, H. Qi, R. Zaretzki
This paper addresses the problem of learning over-complete dictionaries for the coupled feature spaces, where the learned dictionaries also reflect the relationship between the two spaces. A Bayesian method using a beta process prior is applied to learn the over-complete dictionaries. Compared to previous couple feature spaces dictionary learning algorithms, our algorithm not only provides dictionaries that customized to each feature space, but also adds more consistent and accurate mapping between the two feature spaces. This is due to the unique property of the beta process model that the sparse representation can be decomposed to values and dictionary atom indicators. The proposed algorithm is able to learn sparse representations that correspond to the same dictionary atoms with the same sparsity but different values in coupled feature spaces, thus bringing consistent and accurate mapping between coupled feature spaces. Another advantage of the proposed method is that the number of dictionary atoms and their relative importance may be inferred non-parametrically. We compare the proposed approach to several state-of-the-art dictionary learning methods by applying this method to single image super-resolution. The experimental results show that dictionaries learned by our method produces the best super-resolution results compared to other state-of-the-art methods.
本文解决了耦合特征空间的过完备字典学习问题,其中学习到的字典也反映了两个空间之间的关系。采用贝叶斯先验方法学习过完备字典。与之前的两个特征空间字典学习算法相比,我们的算法不仅提供了针对每个特征空间的定制字典,而且在两个特征空间之间增加了更加一致和准确的映射。这是由于beta过程模型的独特属性,即稀疏表示可以分解为值和字典原子指示符。该算法能够学习到在耦合特征空间中具有相同稀疏度但值不同的相同字典原子对应的稀疏表示,从而实现耦合特征空间之间一致、准确的映射。该方法的另一个优点是字典原子的数量和它们的相对重要性可以非参数地推断出来。我们通过将该方法应用于单个图像超分辨率,将所提出的方法与几种最先进的字典学习方法进行比较。实验结果表明,与其他最先进的方法相比,我们的方法学习的字典产生了最好的超分辨率结果。
{"title":"Beta Process Joint Dictionary Learning for Coupled Feature Spaces with Application to Single Image Super-Resolution","authors":"Li He, H. Qi, R. Zaretzki","doi":"10.1109/CVPR.2013.51","DOIUrl":"https://doi.org/10.1109/CVPR.2013.51","url":null,"abstract":"This paper addresses the problem of learning over-complete dictionaries for the coupled feature spaces, where the learned dictionaries also reflect the relationship between the two spaces. A Bayesian method using a beta process prior is applied to learn the over-complete dictionaries. Compared to previous couple feature spaces dictionary learning algorithms, our algorithm not only provides dictionaries that customized to each feature space, but also adds more consistent and accurate mapping between the two feature spaces. This is due to the unique property of the beta process model that the sparse representation can be decomposed to values and dictionary atom indicators. The proposed algorithm is able to learn sparse representations that correspond to the same dictionary atoms with the same sparsity but different values in coupled feature spaces, thus bringing consistent and accurate mapping between coupled feature spaces. Another advantage of the proposed method is that the number of dictionary atoms and their relative importance may be inferred non-parametrically. We compare the proposed approach to several state-of-the-art dictionary learning methods by applying this method to single image super-resolution. The experimental results show that dictionaries learned by our method produces the best super-resolution results compared to other state-of-the-art methods.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"14 1","pages":"345-352"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84913267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 173
Finding Things: Image Parsing with Regions and Per-Exemplar Detectors 寻找事物:使用区域和每样例检测器进行图像解析
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.386
Joseph Tighe, S. Lazebnik
This paper presents a system for image parsing, or labeling each pixel in an image with its semantic category, aimed at achieving broad coverage across hundreds of object categories, many of them sparsely sampled. The system combines region-level features with per-exemplar sliding window detectors. Per-exemplar detectors are better suited for our parsing task than traditional bounding box detectors: they perform well on classes with little training data and high intra-class variation, and they allow object masks to be transferred into the test image for pixel-level segmentation. The proposed system achieves state-of-the-art accuracy on three challenging datasets, the largest of which contains 45,676 images and 232 labels.
本文提出了一个用于图像解析的系统,或用其语义类别标记图像中的每个像素,旨在实现数百个对象类别的广泛覆盖,其中许多是稀疏采样的。该系统结合了区域级特征和每样例滑动窗口检测器。每样例检测器比传统的边界盒检测器更适合我们的解析任务:它们在训练数据少、类内变化大的类上表现良好,并且它们允许将对象掩码转移到测试图像中进行像素级分割。该系统在三个具有挑战性的数据集上实现了最先进的精度,其中最大的数据集包含45,676张图像和232个标签。
{"title":"Finding Things: Image Parsing with Regions and Per-Exemplar Detectors","authors":"Joseph Tighe, S. Lazebnik","doi":"10.1109/CVPR.2013.386","DOIUrl":"https://doi.org/10.1109/CVPR.2013.386","url":null,"abstract":"This paper presents a system for image parsing, or labeling each pixel in an image with its semantic category, aimed at achieving broad coverage across hundreds of object categories, many of them sparsely sampled. The system combines region-level features with per-exemplar sliding window detectors. Per-exemplar detectors are better suited for our parsing task than traditional bounding box detectors: they perform well on classes with little training data and high intra-class variation, and they allow object masks to be transferred into the test image for pixel-level segmentation. The proposed system achieves state-of-the-art accuracy on three challenging datasets, the largest of which contains 45,676 images and 232 labels.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"11 1","pages":"3001-3008"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84972961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 227
A Non-parametric Framework for Document Bleed-through Removal 一种非参数化的文档透漏删除框架
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.380
Róisín Rowley-Brooke, François Pitié, A. Kokaram
This paper presents recent work on a new framework for non-blind document bleed-through removal. The framework includes image preprocessing to remove local intensity variations, pixel region classification based on a segmentation of the joint recto-verso intensity histogram and connected component analysis on the subsequent image labelling. Finally restoration of the degraded regions is performed using exemplar-based image in painting. The proposed method is evaluated visually and numerically on a freely available database of 25 scanned manuscript image pairs with ground truth, and is shown to outperform recent non-blind bleed-through removal techniques.
本文介绍了一种新的非盲文档透漏删除框架的最新工作。该框架包括去除局部强度变化的图像预处理、基于联合直-反强度直方图分割的像素区域分类以及随后图像标记的连接分量分析。最后,利用基于实例的绘画图像对退化区域进行恢复。该方法在一个包含25个扫描手稿图像对的免费数据库上进行了视觉和数值评估,结果表明该方法优于最近的非盲透血去除技术。
{"title":"A Non-parametric Framework for Document Bleed-through Removal","authors":"Róisín Rowley-Brooke, François Pitié, A. Kokaram","doi":"10.1109/CVPR.2013.380","DOIUrl":"https://doi.org/10.1109/CVPR.2013.380","url":null,"abstract":"This paper presents recent work on a new framework for non-blind document bleed-through removal. The framework includes image preprocessing to remove local intensity variations, pixel region classification based on a segmentation of the joint recto-verso intensity histogram and connected component analysis on the subsequent image labelling. Finally restoration of the degraded regions is performed using exemplar-based image in painting. The proposed method is evaluated visually and numerically on a freely available database of 25 scanned manuscript image pairs with ground truth, and is shown to outperform recent non-blind bleed-through removal techniques.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"15 1","pages":"2954-2960"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85238444","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Single Image Calibration of Multi-axial Imaging Systems 多轴成像系统的单图像校准
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.184
Amit K. Agrawal, S. Ramalingam
Imaging systems consisting of a camera looking at multiple spherical mirrors (reflection) or multiple refractive spheres (refraction) have been used for wide-angle imaging applications. We describe such setups as multi-axial imaging systems, since a single sphere results in an axial system. Assuming an internally calibrated camera, calibration of such multi-axial systems involves estimating the sphere radii and locations in the camera coordinate system. However, previous calibration approaches require manual intervention or constrained setups. We present a fully automatic approach using a single photo of a 2D calibration grid. The pose of the calibration grid is assumed to be unknown and is also recovered. Our approach can handle unconstrained setups, where the mirrors/refractive balls can be arranged in any fashion, not necessarily on a grid. The axial nature of rays allows us to compute the axis of each sphere separately. We then show that by choosing rays from two or more spheres, the unknown pose of the calibration grid can be obtained linearly and independently of sphere radii and locations. Knowing the pose, we derive analytical solutions for obtaining the sphere radius and location. This leads to an interesting result that 6-DOF pose estimation of a multi-axial camera can be done without the knowledge of full calibration. Simulations and real experiments demonstrate the applicability of our algorithm.
由相机组成的成像系统可以观察多个球面镜(反射)或多个折射球(折射),用于广角成像应用。我们将这种设置描述为多轴成像系统,因为单个球体导致轴向系统。假设一个内部校准的相机,这种多轴系统的校准涉及估计球面半径和相机坐标系中的位置。然而,以前的校准方法需要人工干预或限制设置。我们提出了一种使用二维校准网格的单张照片的全自动方法。假设标定网格的位姿是未知的,并对其进行恢复。我们的方法可以处理不受约束的设置,其中镜子/折射球可以以任何方式排列,不一定是在网格上。射线的轴向性质允许我们分别计算每个球体的轴。然后,我们证明了通过选择来自两个或多个球体的射线,可以线性地独立于球体半径和位置获得校准网格的未知姿态。在已知姿态的基础上,推导出求解球面半径和位置的解析解。这导致了一个有趣的结果,多轴相机的6自由度姿态估计可以在没有完全校准知识的情况下完成。仿真和实际实验证明了该算法的适用性。
{"title":"Single Image Calibration of Multi-axial Imaging Systems","authors":"Amit K. Agrawal, S. Ramalingam","doi":"10.1109/CVPR.2013.184","DOIUrl":"https://doi.org/10.1109/CVPR.2013.184","url":null,"abstract":"Imaging systems consisting of a camera looking at multiple spherical mirrors (reflection) or multiple refractive spheres (refraction) have been used for wide-angle imaging applications. We describe such setups as multi-axial imaging systems, since a single sphere results in an axial system. Assuming an internally calibrated camera, calibration of such multi-axial systems involves estimating the sphere radii and locations in the camera coordinate system. However, previous calibration approaches require manual intervention or constrained setups. We present a fully automatic approach using a single photo of a 2D calibration grid. The pose of the calibration grid is assumed to be unknown and is also recovered. Our approach can handle unconstrained setups, where the mirrors/refractive balls can be arranged in any fashion, not necessarily on a grid. The axial nature of rays allows us to compute the axis of each sphere separately. We then show that by choosing rays from two or more spheres, the unknown pose of the calibration grid can be obtained linearly and independently of sphere radii and locations. Knowing the pose, we derive analytical solutions for obtaining the sphere radius and location. This leads to an interesting result that 6-DOF pose estimation of a multi-axial camera can be done without the knowledge of full calibration. Simulations and real experiments demonstrate the applicability of our algorithm.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"291 1","pages":"1399-1406"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85401547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
In Defense of Sparsity Based Face Recognition 基于稀疏度的人脸识别
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.58
Weihong Deng, Jiani Hu, Jun Guo
The success of sparse representation based classification (SRC) has largely boosted the research of sparsity based face recognition in recent years. A prevailing view is that the sparsity based face recognition performs well only when the training images have been carefully controlled and the number of samples per class is sufficiently large. This paper challenges the prevailing view by proposing a ``prototype plus variation'' representation model for sparsity based face recognition. Based on the new model, a Superposed SRC (SSRC), in which the dictionary is assembled by the class centroids and the sample-to-centroid differences, leads to a substantial improvement on SRC. The experiments results on AR, FERET and FRGC databases validate that, if the proposed prototype plus variation representation model is applied, sparse coding plays a crucial role in face recognition, and performs well even when the dictionary bases are collected under uncontrolled conditions and only a single sample per classes is available.
基于稀疏表示的分类(SRC)的成功极大地推动了近年来基于稀疏表示的人脸识别研究。一种流行的观点是,基于稀疏度的人脸识别只有在训练图像被仔细控制并且每个类的样本数量足够大的情况下才能表现良好。本文提出了一种基于稀疏性的人脸识别的“原型加变异”表示模型,挑战了目前流行的观点。在此基础上,提出了一种基于类质心和样本间质心差异对字典进行组合的Superposed SRC (SSRC)模型。在AR、FERET和FRGC数据库上的实验结果表明,如果采用本文提出的原型加变异表示模型,稀疏编码在人脸识别中发挥了至关重要的作用,即使在不受控制的条件下收集字典库,每个类只有一个样本,稀疏编码也能取得很好的效果。
{"title":"In Defense of Sparsity Based Face Recognition","authors":"Weihong Deng, Jiani Hu, Jun Guo","doi":"10.1109/CVPR.2013.58","DOIUrl":"https://doi.org/10.1109/CVPR.2013.58","url":null,"abstract":"The success of sparse representation based classification (SRC) has largely boosted the research of sparsity based face recognition in recent years. A prevailing view is that the sparsity based face recognition performs well only when the training images have been carefully controlled and the number of samples per class is sufficiently large. This paper challenges the prevailing view by proposing a ``prototype plus variation'' representation model for sparsity based face recognition. Based on the new model, a Superposed SRC (SSRC), in which the dictionary is assembled by the class centroids and the sample-to-centroid differences, leads to a substantial improvement on SRC. The experiments results on AR, FERET and FRGC databases validate that, if the proposed prototype plus variation representation model is applied, sparse coding plays a crucial role in face recognition, and performs well even when the dictionary bases are collected under uncontrolled conditions and only a single sample per classes is available.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"11 1","pages":"399-406"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81932675","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 186
Towards Pose Robust Face Recognition 面向姿态鲁棒人脸识别
Pub Date : 2013-06-23 DOI: 10.1109/CVPR.2013.454
Dong Yi, Zhen Lei, S. Li
Most existing pose robust methods are too computational complex to meet practical applications and their performance under unconstrained environments are rarely evaluated. In this paper, we propose a novel method for pose robust face recognition towards practical applications, which is fast, pose robust and can work well under unconstrained environments. Firstly, a 3D deformable model is built and a fast 3D model fitting algorithm is proposed to estimate the pose of face image. Secondly, a group of Gabor filters are transformed according to the pose and shape of face image for feature extraction. Finally, PCA is applied on the pose adaptive Gabor features to remove the redundances and Cosine metric is used to evaluate the similarity. The proposed method has three advantages: (1) The pose correction is applied in the filter space rather than image space, which makes our method less affected by the precision of the 3D model, (2) By combining the holistic pose transformation and local Gabor filtering, the final feature is robust to pose and other negative factors in face recognition, (3) The 3D structure and facial symmetry are successfully used to deal with self-occlusion. Extensive experiments on FERET and PIE show the proposed method outperforms state-of-the-art methods significantly, meanwhile, the method works well on LFW.
大多数现有的位姿鲁棒方法计算量太大,无法满足实际应用,而且在无约束环境下的性能很少得到评估。本文提出了一种面向实际应用的姿态鲁棒人脸识别新方法,该方法快速、姿态鲁棒且能在无约束环境下很好地工作。首先,建立三维可变形模型,提出一种快速的三维模型拟合算法来估计人脸图像的姿态;其次,根据人脸图像的姿态和形状变换一组Gabor滤波器进行特征提取;最后,对姿态自适应Gabor特征进行主成分分析去除冗余,并用余弦度量评估相似度。该方法具有三个优点:(1)姿态校正应用于滤波空间而不是图像空间,使我们的方法受三维模型精度的影响较小;(2)结合整体姿态变换和局部Gabor滤波,最终特征对姿态等人脸识别中的负面因素具有鲁棒性;(3)成功地利用三维结构和面部对称性来处理自遮挡。在FERET和PIE上的大量实验表明,该方法明显优于现有的方法,同时在LFW上也取得了良好的效果。
{"title":"Towards Pose Robust Face Recognition","authors":"Dong Yi, Zhen Lei, S. Li","doi":"10.1109/CVPR.2013.454","DOIUrl":"https://doi.org/10.1109/CVPR.2013.454","url":null,"abstract":"Most existing pose robust methods are too computational complex to meet practical applications and their performance under unconstrained environments are rarely evaluated. In this paper, we propose a novel method for pose robust face recognition towards practical applications, which is fast, pose robust and can work well under unconstrained environments. Firstly, a 3D deformable model is built and a fast 3D model fitting algorithm is proposed to estimate the pose of face image. Secondly, a group of Gabor filters are transformed according to the pose and shape of face image for feature extraction. Finally, PCA is applied on the pose adaptive Gabor features to remove the redundances and Cosine metric is used to evaluate the similarity. The proposed method has three advantages: (1) The pose correction is applied in the filter space rather than image space, which makes our method less affected by the precision of the 3D model, (2) By combining the holistic pose transformation and local Gabor filtering, the final feature is robust to pose and other negative factors in face recognition, (3) The 3D structure and facial symmetry are successfully used to deal with self-occlusion. Extensive experiments on FERET and PIE show the proposed method outperforms state-of-the-art methods significantly, meanwhile, the method works well on LFW.","PeriodicalId":6343,"journal":{"name":"2013 IEEE Conference on Computer Vision and Pattern Recognition","volume":"18 1","pages":"3539-3545"},"PeriodicalIF":0.0,"publicationDate":"2013-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82200654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 213
期刊
2013 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1