首页 > 最新文献

2009 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Disambiguating the recognition of 3D objects 消除三维物体识别的歧义
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206683
Gutemberg Guerra-Filho
We propose novel algorithms for the detection, segmentation, recognition, and pose estimation of three-dimensional objects. Our approach initially infers geometric primitives to describe the set of 3D objects. A hierarchical structure is constructed to organize the objects in terms of shared primitives and relations between different primitives in the same object. This structure is shown to disambiguate the object models and to improve recognition rates. The primitives are obtained through our new Invariant Hough Transform. This algorithm uses geometric invariants to compute relations for subsets of points in a specific object. Each relation is stored in a hash table according to the invariant value. The hash table is used to find potential corresponding points between objects. With point matches, pose estimation is achieved by building a probability distribution of transformations. We evaluate our methods with experiments using synthetic and real 3D objects.
我们提出了三维物体的检测、分割、识别和姿态估计的新算法。我们的方法首先推断几何原语来描述3D对象集。层次结构是根据共享原语和同一对象中不同原语之间的关系来组织对象的。这种结构可以消除目标模型的歧义,提高识别率。这些原语是通过新的不变霍夫变换得到的。该算法使用几何不变量来计算特定对象中点子集的关系。每个关系根据不变值存储在哈希表中。哈希表用于查找对象之间可能对应的点。对于点匹配,姿态估计是通过建立变换的概率分布来实现的。我们用合成的和真实的三维物体来评估我们的方法。
{"title":"Disambiguating the recognition of 3D objects","authors":"Gutemberg Guerra-Filho","doi":"10.1109/CVPR.2009.5206683","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206683","url":null,"abstract":"We propose novel algorithms for the detection, segmentation, recognition, and pose estimation of three-dimensional objects. Our approach initially infers geometric primitives to describe the set of 3D objects. A hierarchical structure is constructed to organize the objects in terms of shared primitives and relations between different primitives in the same object. This structure is shown to disambiguate the object models and to improve recognition rates. The primitives are obtained through our new Invariant Hough Transform. This algorithm uses geometric invariants to compute relations for subsets of points in a specific object. Each relation is stored in a hash table according to the invariant value. The hash table is used to find potential corresponding points between objects. With point matches, pose estimation is achieved by building a probability distribution of transformations. We evaluate our methods with experiments using synthetic and real 3D objects.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131302227","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Towards high-resolution large-scale multi-view stereo 迈向高分辨率大尺度多视点立体
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206617
Hoang-Hiep Vu, R. Keriven, Patrick Labatut, Jean-Philippe Pons
Boosted by the Middlebury challenge, the precision of dense multi-view stereovision methods has increased drastically in the past few years. Yet, most methods, although they perform well on this benchmark, are still inapplicable to large-scale data sets taken under uncontrolled conditions. In this paper, we propose a multi-view stereo pipeline able to deal at the same time with very large scenes while still producing highly detailed reconstructions within very reasonable time. The keys to these benefits are twofold: (i) a minimum s-t cut based global optimization that transforms a dense point cloud into a visibility consistent mesh, followed by (ii) a mesh-based variational refinement that captures small details, smartly handling photo-consistency, regularization and adaptive resolution. Our method has been tested on numerous large-scale outdoor scenes. The accuracy of our reconstructions is also measured on the recent dense multi-view benchmark proposed by Strecha et al., showing our results to compare more than favorably with the current state-of-the-art.
在米德尔伯里挑战的推动下,密集多视图立体视觉方法的精度在过去几年中急剧提高。然而,大多数方法虽然在这个基准上表现良好,但仍然不适用于在不受控制的条件下采集的大规模数据集。在本文中,我们提出了一种多视图立体管道,能够同时处理非常大的场景,同时在非常合理的时间内产生非常详细的重建。这些好处的关键是双重的:(i)基于最小s-t切割的全局优化,将密集的点云转换为可见性一致的网格,其次是(ii)基于网格的变分细化,捕获小细节,巧妙地处理照片一致性,正则化和自适应分辨率。我们的方法已经在许多大型户外场景中进行了测试。我们的重建精度也在Strecha等人最近提出的密集多视图基准上进行了测量,表明我们的结果与当前最先进的技术相比更加有利。
{"title":"Towards high-resolution large-scale multi-view stereo","authors":"Hoang-Hiep Vu, R. Keriven, Patrick Labatut, Jean-Philippe Pons","doi":"10.1109/CVPR.2009.5206617","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206617","url":null,"abstract":"Boosted by the Middlebury challenge, the precision of dense multi-view stereovision methods has increased drastically in the past few years. Yet, most methods, although they perform well on this benchmark, are still inapplicable to large-scale data sets taken under uncontrolled conditions. In this paper, we propose a multi-view stereo pipeline able to deal at the same time with very large scenes while still producing highly detailed reconstructions within very reasonable time. The keys to these benefits are twofold: (i) a minimum s-t cut based global optimization that transforms a dense point cloud into a visibility consistent mesh, followed by (ii) a mesh-based variational refinement that captures small details, smartly handling photo-consistency, regularization and adaptive resolution. Our method has been tested on numerous large-scale outdoor scenes. The accuracy of our reconstructions is also measured on the recent dense multi-view benchmark proposed by Strecha et al., showing our results to compare more than favorably with the current state-of-the-art.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131847950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 279
Learning rotational features for filament detection 学习旋转特征的灯丝检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206511
Germán González, F. Fleuret, P. Fua
State-of-the-art approaches for detecting filament-like structures in noisy images rely on filters optimized for signals of a particular shape, such as an ideal edge or ridge. While these approaches are optimal when the image conforms to these ideal shapes, their performance quickly degrades on many types of real data where the image deviates from the ideal model, and when noise processes violate a Gaussian assumption. In this paper, we show that by learning rotational features, we can outperform state-of-the-art filament detection techniques on many different kinds of imagery. More specifically, we demonstrate superior performance for the detection of blood vessel in retinal scans, neurons in brightfield microscopy imagery, and streets in satellite imagery.
在噪声图像中检测丝状结构的最先进方法依赖于针对特定形状的信号优化的滤波器,例如理想的边缘或脊。当图像符合这些理想形状时,这些方法是最优的,但是当图像偏离理想模型时,当噪声过程违反高斯假设时,它们的性能会迅速下降。在本文中,我们表明,通过学习旋转特征,我们可以在许多不同类型的图像上优于最先进的灯丝检测技术。更具体地说,我们在视网膜扫描中的血管检测、明场显微镜图像中的神经元检测和卫星图像中的街道检测方面展示了卓越的性能。
{"title":"Learning rotational features for filament detection","authors":"Germán González, F. Fleuret, P. Fua","doi":"10.1109/CVPR.2009.5206511","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206511","url":null,"abstract":"State-of-the-art approaches for detecting filament-like structures in noisy images rely on filters optimized for signals of a particular shape, such as an ideal edge or ridge. While these approaches are optimal when the image conforms to these ideal shapes, their performance quickly degrades on many types of real data where the image deviates from the ideal model, and when noise processes violate a Gaussian assumption. In this paper, we show that by learning rotational features, we can outperform state-of-the-art filament detection techniques on many different kinds of imagery. More specifically, we demonstrate superior performance for the detection of blood vessel in retinal scans, neurons in brightfield microscopy imagery, and streets in satellite imagery.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124383463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
A similarity measure between vector sequences with application to handwritten word image retrieval 向量序列之间的相似性度量及其在手写文字图像检索中的应用
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206783
José A. Rodríguez-Serrano, F. Perronnin, J. Lladós, Gemma Sánchez
This article proposes a novel similarity measure between vector sequences. Recently, a model-based approach was introduced to address this issue. It consists in modeling each sequence with a continuous Hidden Markov Model (CHMM) and computing a probabilistic measure of similarity between C-HMMs. In this paper we propose to model sequences with semi-continuous HMMs (SC-HMMs): the Gaussians of the SC-HMMs are constrained to belong to a shared pool of Gaussians. This constraint provides two major benefits. First, the a priori information contained in the common set of Gaussians leads to a more accurate estimate of the HMM parameters. Second, the computation of a probabilistic similarity between two SC-HMMs can be simplified to a Dynamic Time Warping (DTW) between their mixture weight vectors, which reduces significantly the computational cost. Experimental results on a handwritten word retrieval task show that the proposed similarity outperforms the traditional DTW between the original sequences, and the model-based approach which uses C-HMMs. We also show that this increase in accuracy can be traded against a significant reduction of the computational cost (up to 100 times).
本文提出了一种新的向量序列相似性度量方法。最近,引入了一种基于模型的方法来解决这个问题。它包括用连续隐马尔可夫模型(CHMM)对每个序列建模,并计算c - hmm之间相似性的概率度量。本文提出用半连续hmm (sc - hmm)对序列进行建模:sc - hmm的高斯函数被约束为属于一个共享的高斯函数池。这个约束提供了两个主要好处。首先,包含在公共高斯集合中的先验信息可以更准确地估计HMM参数。其次,将两个sc - hmm之间的概率相似度计算简化为混合权重向量之间的动态时间规整(DTW),大大降低了计算成本。手写体单词检索的实验结果表明,本文提出的相似度方法优于传统的原始序列之间的DTW方法和基于模型的使用c - hmm的方法。我们还表明,准确度的提高可以与计算成本的显著降低(高达100倍)相交换。
{"title":"A similarity measure between vector sequences with application to handwritten word image retrieval","authors":"José A. Rodríguez-Serrano, F. Perronnin, J. Lladós, Gemma Sánchez","doi":"10.1109/CVPR.2009.5206783","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206783","url":null,"abstract":"This article proposes a novel similarity measure between vector sequences. Recently, a model-based approach was introduced to address this issue. It consists in modeling each sequence with a continuous Hidden Markov Model (CHMM) and computing a probabilistic measure of similarity between C-HMMs. In this paper we propose to model sequences with semi-continuous HMMs (SC-HMMs): the Gaussians of the SC-HMMs are constrained to belong to a shared pool of Gaussians. This constraint provides two major benefits. First, the a priori information contained in the common set of Gaussians leads to a more accurate estimate of the HMM parameters. Second, the computation of a probabilistic similarity between two SC-HMMs can be simplified to a Dynamic Time Warping (DTW) between their mixture weight vectors, which reduces significantly the computational cost. Experimental results on a handwritten word retrieval task show that the proposed similarity outperforms the traditional DTW between the original sequences, and the model-based approach which uses C-HMMs. We also show that this increase in accuracy can be traded against a significant reduction of the computational cost (up to 100 times).","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114413139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Coded exposure deblurring: Optimized codes for PSF estimation and invertibility 编码曝光去模糊:PSF估计和可逆性的优化代码
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206685
Amit K. Agrawal, Yi Xu
We consider the problem of single image object motion deblurring from a static camera. It is well-known that deblurring of moving objects using a traditional camera is ill-posed, due to the loss of high spatial frequencies in the captured blurred image. A coded exposure camera modulates the integration pattern of light by opening and closing the shutter within the exposure time using a binary code. The code is chosen to make the resulting point spread function (PSF) invertible, for best deconvolution performance. However, for a successful deconvolution algorithm, PSF estimation is as important as PSF invertibility. We show that PSF estimation is easier if the resulting motion blur is smooth and the optimal code for PSF invertibility could worsen PSF estimation, since it leads to non-smooth blur. We show that both criterions of PSF invertibility and PSF estimation can be simultaneously met, albeit with a slight increase in the deconvolution noise. We propose design rules for a code to have good PSF estimation capability and outline two search criteria for finding the optimal code for a given length. We present theoretical analysis comparing the performance of the proposed code with the code optimized solely for PSF invertibility. We also show how to easily implement coded exposure on a consumer grade machine vision camera with no additional hardware. Real experimental results demonstrate the effectiveness of the proposed codes for motion deblurring.
研究了静态摄像机中单幅图像物体运动的去模糊问题。众所周知,由于在捕获的模糊图像中丢失了高空间频率,使用传统相机对运动物体进行去模糊是不恰当的。编码曝光相机通过使用二进制代码在曝光时间内打开和关闭快门来调制光的集成模式。为了获得最佳的反卷积性能,选择的代码使结果点扩展函数(PSF)可逆。然而,对于一个成功的反卷积算法,PSF估计和PSF可逆性同样重要。我们表明,如果产生的运动模糊是平滑的,PSF估计更容易,而PSF可逆性的最佳代码可能会恶化PSF估计,因为它会导致非平滑模糊。我们证明了PSF可逆性和PSF估计的两个准则可以同时满足,尽管反卷积噪声略有增加。我们提出了具有良好PSF估计能力的代码的设计规则,并概述了寻找给定长度的最优代码的两个搜索准则。我们提出了理论分析,比较了所提出的代码与仅针对PSF可逆性优化的代码的性能。我们还展示了如何在没有额外硬件的情况下在消费级机器视觉相机上轻松实现编码曝光。实际实验结果证明了所提代码对运动去模糊的有效性。
{"title":"Coded exposure deblurring: Optimized codes for PSF estimation and invertibility","authors":"Amit K. Agrawal, Yi Xu","doi":"10.1109/CVPR.2009.5206685","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206685","url":null,"abstract":"We consider the problem of single image object motion deblurring from a static camera. It is well-known that deblurring of moving objects using a traditional camera is ill-posed, due to the loss of high spatial frequencies in the captured blurred image. A coded exposure camera modulates the integration pattern of light by opening and closing the shutter within the exposure time using a binary code. The code is chosen to make the resulting point spread function (PSF) invertible, for best deconvolution performance. However, for a successful deconvolution algorithm, PSF estimation is as important as PSF invertibility. We show that PSF estimation is easier if the resulting motion blur is smooth and the optimal code for PSF invertibility could worsen PSF estimation, since it leads to non-smooth blur. We show that both criterions of PSF invertibility and PSF estimation can be simultaneously met, albeit with a slight increase in the deconvolution noise. We propose design rules for a code to have good PSF estimation capability and outline two search criteria for finding the optimal code for a given length. We present theoretical analysis comparing the performance of the proposed code with the code optimized solely for PSF invertibility. We also show how to easily implement coded exposure on a consumer grade machine vision camera with no additional hardware. Real experimental results demonstrate the effectiveness of the proposed codes for motion deblurring.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123445806","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 89
Motion pattern interpretation and detection for tracking moving vehicles in airborne video 机载视频中运动车辆跟踪的运动模式解释与检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206541
Qian Yu, G. Medioni
Detection and tracking of moving vehicles in airborne videos is a challenging problem. Many approaches have been proposed to improve motion segmentation on frame-by-frame and pixel-by-pixel bases, however, little attention has been paid to analyze the long-term motion pattern, which is a distinctive property for moving vehicles in airborne videos. In this paper, we provide a straightforward geometric interpretation of a general motion pattern in 4D space (x, y, vx, vy). We propose to use the tensor voting computational framework to detect and segment such motion patterns in 4D space. Specifically, in airborne videos, we analyze the essential difference in motion patterns caused by parallax and independent moving objects, which leads to a practical method for segmenting motion patterns (flows) created by moving vehicles in stabilized airborne videos. The flows are used in turn to facilitate detection and tracking of each individual object in the flow. Conceptually, this approach is similar to “track-before-detect” techniques, which involves temporal information in the process as early as possible. As shown in the experiments, many difficult cases in airborne videos, such as parallax, noisy background modeling and long term occlusions, can be addressed by our approach.
机载视频中运动飞行器的检测与跟踪是一个具有挑战性的问题。人们提出了许多方法来改进逐帧和逐像素的运动分割,然而,很少有人关注分析长期运动模式,这是机载视频中运动车辆的一个独特特性。在本文中,我们提供了四维空间(x, y, vx, vy)中一般运动模式的直接几何解释。我们建议使用张量投票计算框架来检测和分割四维空间中的这种运动模式。具体来说,在机载视频中,我们分析了视差和独立运动物体引起的运动模式的本质区别,从而得出了一种实用的方法来分割稳定机载视频中运动车辆产生的运动模式(流)。依次使用流来促进流中每个单独对象的检测和跟踪。从概念上讲,这种方法类似于“检测前跟踪”技术,它尽可能早地涉及到过程中的时间信息。实验表明,我们的方法可以解决机载视频中的许多困难情况,如视差、噪声背景建模和长期遮挡。
{"title":"Motion pattern interpretation and detection for tracking moving vehicles in airborne video","authors":"Qian Yu, G. Medioni","doi":"10.1109/CVPR.2009.5206541","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206541","url":null,"abstract":"Detection and tracking of moving vehicles in airborne videos is a challenging problem. Many approaches have been proposed to improve motion segmentation on frame-by-frame and pixel-by-pixel bases, however, little attention has been paid to analyze the long-term motion pattern, which is a distinctive property for moving vehicles in airborne videos. In this paper, we provide a straightforward geometric interpretation of a general motion pattern in 4D space (x, y, vx, vy). We propose to use the tensor voting computational framework to detect and segment such motion patterns in 4D space. Specifically, in airborne videos, we analyze the essential difference in motion patterns caused by parallax and independent moving objects, which leads to a practical method for segmenting motion patterns (flows) created by moving vehicles in stabilized airborne videos. The flows are used in turn to facilitate detection and tracking of each individual object in the flow. Conceptually, this approach is similar to “track-before-detect” techniques, which involves temporal information in the process as early as possible. As shown in the experiments, many difficult cases in airborne videos, such as parallax, noisy background modeling and long term occlusions, can be addressed by our approach.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116810879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 58
Fast multiple shape correspondence by pre-organizing shape instances 通过预先组织形状实例,快速实现多个形状对应
Pub Date : 2009-06-20 DOI: 10.1109/cvpr.2009.5206611
B. Munsell, Andrew Temlyakov, Song Wang
Accurately identifying corresponded landmarks from a population of shape instances is the major challenge in constructing statistical shape models. In general, shape-correspondence methods can be grouped into one of two categories: global methods and pair-wise methods. In this paper, we develop a new method that attempts to address the limitations of both the global and pair-wise methods. In particular, we reorganize the input population into a tree structure that incorporates global information about the population of shape instances, where each node in the tree represents a shape instance and each edge connects two very similar shape instances. Using this organized tree, neighboring shape instances can be corresponded efficiently and accurately by a pair-wise method. In the experiments, we evaluate the proposed method and compare its performance to five available shape correspondence methods and show the proposed method achieves the accuracy of a global method with speed of a pair-wise method.
在统计形状模型的构建中,准确地从一群形状实例中识别出相应的标志是一个主要的挑战。一般来说,形状对应方法可以分为两类:全局方法和成对方法。在本文中,我们开发了一种新的方法,试图解决全局方法和成对方法的局限性。特别是,我们将输入填充重新组织成一个包含形状实例填充全局信息的树结构,其中树中的每个节点代表一个形状实例,每个边连接两个非常相似的形状实例。利用这种组织树,相邻的形状实例可以通过配对方法高效准确地对应。在实验中,我们对所提出的方法进行了评价,并将其性能与五种可用的形状对应方法进行了比较,结果表明所提出的方法达到了全局方法的精度和成对方法的速度。
{"title":"Fast multiple shape correspondence by pre-organizing shape instances","authors":"B. Munsell, Andrew Temlyakov, Song Wang","doi":"10.1109/cvpr.2009.5206611","DOIUrl":"https://doi.org/10.1109/cvpr.2009.5206611","url":null,"abstract":"Accurately identifying corresponded landmarks from a population of shape instances is the major challenge in constructing statistical shape models. In general, shape-correspondence methods can be grouped into one of two categories: global methods and pair-wise methods. In this paper, we develop a new method that attempts to address the limitations of both the global and pair-wise methods. In particular, we reorganize the input population into a tree structure that incorporates global information about the population of shape instances, where each node in the tree represents a shape instance and each edge connects two very similar shape instances. Using this organized tree, neighboring shape instances can be corresponded efficiently and accurately by a pair-wise method. In the experiments, we evaluate the proposed method and compare its performance to five available shape correspondence methods and show the proposed method achieves the accuracy of a global method with speed of a pair-wise method.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123982331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Towards a practical face recognition system: Robust registration and illumination by sparse representation 一个实用的人脸识别系统:稀疏表示的鲁棒配准和照明
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206654
Andrew Wagner, John Wright, Arvind Ganesh, Zihan Zhou, Yi Ma
Most contemporary face recognition algorithms work well under laboratory conditions but degrade when tested in less-controlled environments. This is mostly due to the difficulty of simultaneously handling variations in illumination, alignment, pose, and occlusion. In this paper, we propose a simple and practical face recognition system that achieves a high degree of robustness and stability to all these variations. We demonstrate how to use tools from sparse representation to align a test face image with a set of frontal training images in the presence of significant registration error and occlusion. We thoroughly characterize the region of attraction for our alignment algorithm on public face datasets such as Multi-PIE. We further study how to obtain a sufficient set of training illuminations for linearly interpolating practical lighting conditions. We have implemented a complete face recognition system, including a projector-based training acquisition system, in order to evaluate how our algorithms work under practical testing conditions. We show that our system can efficiently and effectively recognize faces under a variety of realistic conditions, using only frontal images under the proposed illuminations as training.
大多数当代人脸识别算法在实验室条件下工作良好,但在控制较少的环境中进行测试时就会下降。这主要是由于同时处理照明,对齐,姿势和遮挡变化的困难。在本文中,我们提出了一个简单实用的人脸识别系统,该系统对所有这些变化都具有高度的鲁棒性和稳定性。我们演示了如何使用稀疏表示的工具在存在显着配准错误和遮挡的情况下将测试人脸图像与一组正面训练图像对齐。我们在公共人脸数据集(如Multi-PIE)上对我们的对齐算法的吸引区域进行了彻底的表征。我们进一步研究了如何获得一组足够的训练照明来线性插值实际照明条件。我们已经实现了一个完整的人脸识别系统,包括一个基于投影仪的训练采集系统,以评估我们的算法在实际测试条件下的工作情况。我们证明了我们的系统可以在各种现实条件下高效地识别人脸,仅使用提出的照明下的正面图像作为训练。
{"title":"Towards a practical face recognition system: Robust registration and illumination by sparse representation","authors":"Andrew Wagner, John Wright, Arvind Ganesh, Zihan Zhou, Yi Ma","doi":"10.1109/CVPR.2009.5206654","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206654","url":null,"abstract":"Most contemporary face recognition algorithms work well under laboratory conditions but degrade when tested in less-controlled environments. This is mostly due to the difficulty of simultaneously handling variations in illumination, alignment, pose, and occlusion. In this paper, we propose a simple and practical face recognition system that achieves a high degree of robustness and stability to all these variations. We demonstrate how to use tools from sparse representation to align a test face image with a set of frontal training images in the presence of significant registration error and occlusion. We thoroughly characterize the region of attraction for our alignment algorithm on public face datasets such as Multi-PIE. We further study how to obtain a sufficient set of training illuminations for linearly interpolating practical lighting conditions. We have implemented a complete face recognition system, including a projector-based training acquisition system, in order to evaluate how our algorithms work under practical testing conditions. We show that our system can efficiently and effectively recognize faces under a variety of realistic conditions, using only frontal images under the proposed illuminations as training.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123988885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 211
A nonparametric Riemannian framework for processing high angular resolution diffusion images (HARDI) 处理高角分辨率扩散图像的非参数黎曼框架
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206843
A. Goh, C. Lenglet, P. Thompson, R. Vidal
High angular resolution diffusion imaging has become an important magnetic resonance technique for in vivo imaging. Most current research in this field focuses on developing methods for computing the orientation distribution function (ODF), which is the probability distribution function of water molecule diffusion along any angle on the sphere. In this paper, we present a Riemannian framework to carry out computations on an ODF field. The proposed framework does not require that the ODFs be represented by any fixed parameterization, such as a mixture of von Mises-Fisher distributions or a spherical harmonic expansion. Instead, we use a non-parametric representation of the ODF, and exploit the fact that under the square-root re-parameterization, the space of ODFs forms a Riemannian manifold, namely the unit Hilbert sphere. Specifically, we use Riemannian operations to perform various geometric data processing algorithms, such as interpolation, convolution and linear and nonlinear filtering. We illustrate these concepts with numerical experiments on synthetic and real datasets.
高角分辨率扩散成像已成为磁共振体内成像的重要技术。目前该领域的研究主要集中在开发方向分布函数(ODF)的计算方法,方向分布函数是水分子沿球体上任意角度扩散的概率分布函数。在本文中,我们提出了一个黎曼框架来对ODF域进行计算。所提出的框架不要求odf用任何固定的参数化表示,例如von Mises-Fisher分布的混合或球谐展开。相反,我们使用ODF的非参数表示,并利用在平方根重新参数化下,ODF的空间形成黎曼流形,即单位希尔伯特球。具体来说,我们使用黎曼运算来执行各种几何数据处理算法,如插值,卷积以及线性和非线性滤波。我们用合成数据集和真实数据集的数值实验来说明这些概念。
{"title":"A nonparametric Riemannian framework for processing high angular resolution diffusion images (HARDI)","authors":"A. Goh, C. Lenglet, P. Thompson, R. Vidal","doi":"10.1109/CVPR.2009.5206843","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206843","url":null,"abstract":"High angular resolution diffusion imaging has become an important magnetic resonance technique for in vivo imaging. Most current research in this field focuses on developing methods for computing the orientation distribution function (ODF), which is the probability distribution function of water molecule diffusion along any angle on the sphere. In this paper, we present a Riemannian framework to carry out computations on an ODF field. The proposed framework does not require that the ODFs be represented by any fixed parameterization, such as a mixture of von Mises-Fisher distributions or a spherical harmonic expansion. Instead, we use a non-parametric representation of the ODF, and exploit the fact that under the square-root re-parameterization, the space of ODFs forms a Riemannian manifold, namely the unit Hilbert sphere. Specifically, we use Riemannian operations to perform various geometric data processing algorithms, such as interpolation, convolution and linear and nonlinear filtering. We illustrate these concepts with numerical experiments on synthetic and real datasets.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129725605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Efficient planar graph cuts with applications in Computer Vision 高效平面图形切割及其在计算机视觉中的应用
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206863
Frank R. Schmidt, Eno Töppe, D. Cremers
We present a fast graph cut algorithm for planar graphs. It is based on the graph theoretical work and leads to an efficient method that we apply on shape matching and image segmentation. In contrast to currently used methods in computer vision, the presented approach provides an upper bound for its runtime behavior that is almost linear. In particular, we are able to match two different planar shapes of N points in O(N2 log N) and segment a given image of N pixels in O(N log N). We present two experimental benchmark studies which demonstrate that the presented method is also in practice faster than previously proposed graph cut methods: On planar shape matching and image segmentation we observe a speed-up of an order of magnitude, depending on resolution.
提出了一种用于平面图形的快速图切算法。它是在图理论工作的基础上提出的一种有效的形状匹配和图像分割方法。与目前使用的计算机视觉方法相比,该方法为其运行时行为提供了一个几乎线性的上界。特别是,我们能够在O(N2 log N)内匹配N个点的两个不同平面形状,并在O(N log N)内分割给定图像的N个像素。我们提出了两个实验基准研究,表明所提出的方法在实践中也比以前提出的图切方法更快:在平面形状匹配和图像分割上,我们观察到速度提高了一个数量级,这取决于分辨率。
{"title":"Efficient planar graph cuts with applications in Computer Vision","authors":"Frank R. Schmidt, Eno Töppe, D. Cremers","doi":"10.1109/CVPR.2009.5206863","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206863","url":null,"abstract":"We present a fast graph cut algorithm for planar graphs. It is based on the graph theoretical work and leads to an efficient method that we apply on shape matching and image segmentation. In contrast to currently used methods in computer vision, the presented approach provides an upper bound for its runtime behavior that is almost linear. In particular, we are able to match two different planar shapes of N points in O(N2 log N) and segment a given image of N pixels in O(N log N). We present two experimental benchmark studies which demonstrate that the presented method is also in practice faster than previously proposed graph cut methods: On planar shape matching and image segmentation we observe a speed-up of an order of magnitude, depending on resolution.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127013163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
期刊
2009 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1