首页 > 最新文献

2009 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
A family of contextual measures of similarity between distributions with application to image retrieval 分布间相似性的一组上下文度量及其在图像检索中的应用
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206505
F. Perronnin, Yan Liu, J. Renders
We introduce a novel family of contextual measures of similarity between distributions: the similarity between two distributions q and p is measured in the context of a third distribution u. In our framework any traditional measure of similarity / dissimilarity has its contextual counterpart. We show that for two important families of divergences (Bregman and Csisz'ar), the contextual similarity computation consists in solving a convex optimization problem. We focus on the case of multinomials and explain how to compute in practice the similarity for several well-known measures. These contextual measures are then applied to the image retrieval problem. In such a case, the context u is estimated from the neighbors of a query q. One of the main benefits of our approach lies in the fact that using different contexts, and especially contexts at multiple scales (i.e. broad and narrow contexts), provides different views on the same problem. Combining the different views can improve retrieval accuracy. We will show on two very different datasets (one of photographs, the other of document images) that the proposed measures have a relatively small positive impact on macro Average Precision (which measures purely ranking) and a large positive impact on micro Average Precision (which measures both ranking and consistency of the scores across multiple queries).
我们引入了一组新的分布之间相似性的上下文度量:两个分布q和p之间的相似性是在第三个分布u的背景下测量的。在我们的框架中,任何传统的相似性/不相似性度量都有其上下文对应。我们表明,对于两个重要的散度族(Bregman和cissz 'ar),上下文相似性计算包括解决一个凸优化问题。我们将重点讨论多项式的情况,并解释如何在实践中计算几种众所周知的度量的相似度。然后将这些上下文度量应用于图像检索问题。在这种情况下,上下文u是从查询q的邻居中估计出来的。我们的方法的一个主要好处在于,使用不同的上下文,特别是在多个尺度上的上下文(即广义和狭义上下文),可以对同一个问题提供不同的观点。结合不同的视图可以提高检索的准确性。我们将在两个非常不同的数据集(一个是照片,另一个是文档图像)上展示,提议的度量对宏观平均精度(衡量纯粹的排名)有相对较小的积极影响,而对微观平均精度(衡量多个查询的排名和分数的一致性)有很大的积极影响。
{"title":"A family of contextual measures of similarity between distributions with application to image retrieval","authors":"F. Perronnin, Yan Liu, J. Renders","doi":"10.1109/CVPR.2009.5206505","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206505","url":null,"abstract":"We introduce a novel family of contextual measures of similarity between distributions: the similarity between two distributions q and p is measured in the context of a third distribution u. In our framework any traditional measure of similarity / dissimilarity has its contextual counterpart. We show that for two important families of divergences (Bregman and Csisz'ar), the contextual similarity computation consists in solving a convex optimization problem. We focus on the case of multinomials and explain how to compute in practice the similarity for several well-known measures. These contextual measures are then applied to the image retrieval problem. In such a case, the context u is estimated from the neighbors of a query q. One of the main benefits of our approach lies in the fact that using different contexts, and especially contexts at multiple scales (i.e. broad and narrow contexts), provides different views on the same problem. Combining the different views can improve retrieval accuracy. We will show on two very different datasets (one of photographs, the other of document images) that the proposed measures have a relatively small positive impact on macro Average Precision (which measures purely ranking) and a large positive impact on micro Average Precision (which measures both ranking and consistency of the scores across multiple queries).","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123443148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
Discriminative subvolume search for efficient action detection 判别子卷搜索,有效的动作检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206671
Junsong Yuan, Zicheng Liu, Ying Wu
Actions are spatio-temporal patterns which can be characterized by collections of spatio-temporal invariant features. Detection of actions is to find the re-occurrences (e.g. through pattern matching) of such spatio-temporal patterns. This paper addresses two critical issues in pattern matching-based action detection: (1) efficiency of pattern search in 3D videos and (2) tolerance of intra-pattern variations of actions. Our contributions are two-fold. First, we propose a discriminative pattern matching called naive-Bayes based mutual information maximization (NBMIM) for multi-class action categorization. It improves the state-of-the-art results on standard KTH dataset. Second, a novel search algorithm is proposed to locate the optimal subvolume in the 3D video space for efficient action detection. Our method is purely data-driven and does not rely on object detection, tracking or background subtraction. It can well handle the intra-pattern variations of actions such as scale and speed variations, and is insensitive to dynamic and clutter backgrounds and even partial occlusions. The experiments on versatile datasets including KTH and CMU action datasets demonstrate the effectiveness and efficiency of our method.
动作是一种时空模式,可以通过时空不变特征的集合来表征。动作检测就是发现这些时空模式的再次出现(例如通过模式匹配)。本文解决了基于模式匹配的动作检测中的两个关键问题:(1)3D视频中模式搜索的效率;(2)动作模式内变化的容忍度。我们的贡献是双重的。首先,我们提出了一种判别模式匹配方法——基于朴素贝叶斯的互信息最大化(NBMIM),用于多类动作分类。它改进了标准KTH数据集上的最新结果。其次,提出了一种新的搜索算法,在三维视频空间中定位最优子体以进行有效的动作检测。我们的方法是纯数据驱动的,不依赖于目标检测、跟踪或背景减除。它可以很好地处理动作的模式内变化,如规模和速度变化,并且对动态和杂乱背景甚至部分遮挡不敏感。在KTH和CMU动作数据集上的实验证明了该方法的有效性和高效性。
{"title":"Discriminative subvolume search for efficient action detection","authors":"Junsong Yuan, Zicheng Liu, Ying Wu","doi":"10.1109/CVPR.2009.5206671","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206671","url":null,"abstract":"Actions are spatio-temporal patterns which can be characterized by collections of spatio-temporal invariant features. Detection of actions is to find the re-occurrences (e.g. through pattern matching) of such spatio-temporal patterns. This paper addresses two critical issues in pattern matching-based action detection: (1) efficiency of pattern search in 3D videos and (2) tolerance of intra-pattern variations of actions. Our contributions are two-fold. First, we propose a discriminative pattern matching called naive-Bayes based mutual information maximization (NBMIM) for multi-class action categorization. It improves the state-of-the-art results on standard KTH dataset. Second, a novel search algorithm is proposed to locate the optimal subvolume in the 3D video space for efficient action detection. Our method is purely data-driven and does not rely on object detection, tracking or background subtraction. It can well handle the intra-pattern variations of actions such as scale and speed variations, and is insensitive to dynamic and clutter backgrounds and even partial occlusions. The experiments on versatile datasets including KTH and CMU action datasets demonstrate the effectiveness and efficiency of our method.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121698973","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 317
Multiphase geometric couplings for the segmentation of neural processes 神经过程分割的多相几何耦合
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206524
Amelio Vázquez Reina, E. Miller, H. Pfister
The ability to constrain the geometry of deformable models for image segmentation can be useful when information about the expected shape or positioning of the objects in a scene is known a priori. An example of this occurs when segmenting neural cross sections in electron microscopy. Such images often contain multiple nested boundaries separating regions of homogeneous intensities. For these applications, multiphase level sets provide a partitioning framework that allows for the segmentation of multiple deformable objects by combining several level set functions. Although there has been much effort in the study of statistical shape priors that can be used to constrain the geometry of each partition, none of these methods allow for the direct modeling of geometric arrangements of partitions. In this paper, we show how to define elastic couplings between multiple level set functions to model ribbon-like partitions. We build such couplings using dynamic force fields that can depend on the image content and relative location and shape of the level set functions. To the best of our knowledge, this is the first work that shows a direct way of geometrically constraining multiphase level sets for image segmentation. We demonstrate the robustness of our method by comparing it with previous level set segmentation methods.
约束可变形模型的几何形状以进行图像分割的能力在先验地知道场景中物体的预期形状或定位信息时是有用的。在电子显微镜中分割神经横截面就是一个例子。这样的图像通常包含多个嵌套的边界来分隔均匀强度的区域。对于这些应用,多相水平集提供了一个分区框架,允许通过组合几个水平集函数来分割多个可变形对象。尽管在统计形状先验的研究中已经付出了很多努力,这些先验可以用来约束每个分区的几何形状,但这些方法都不允许对分区的几何排列进行直接建模。在本文中,我们展示了如何定义多个水平集函数之间的弹性耦合来模拟带状分区。我们使用动态力场来构建这样的耦合,动态力场可以依赖于图像内容和水平集函数的相对位置和形状。据我们所知,这是第一个展示了用于图像分割的几何约束多相水平集的直接方法的工作。通过与以前的水平集分割方法进行比较,我们证明了该方法的鲁棒性。
{"title":"Multiphase geometric couplings for the segmentation of neural processes","authors":"Amelio Vázquez Reina, E. Miller, H. Pfister","doi":"10.1109/CVPR.2009.5206524","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206524","url":null,"abstract":"The ability to constrain the geometry of deformable models for image segmentation can be useful when information about the expected shape or positioning of the objects in a scene is known a priori. An example of this occurs when segmenting neural cross sections in electron microscopy. Such images often contain multiple nested boundaries separating regions of homogeneous intensities. For these applications, multiphase level sets provide a partitioning framework that allows for the segmentation of multiple deformable objects by combining several level set functions. Although there has been much effort in the study of statistical shape priors that can be used to constrain the geometry of each partition, none of these methods allow for the direct modeling of geometric arrangements of partitions. In this paper, we show how to define elastic couplings between multiple level set functions to model ribbon-like partitions. We build such couplings using dynamic force fields that can depend on the image content and relative location and shape of the level set functions. To the best of our knowledge, this is the first work that shows a direct way of geometrically constraining multiphase level sets for image segmentation. We demonstrate the robustness of our method by comparing it with previous level set segmentation methods.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124424459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
Joint and implicit registration for face recognition 人脸识别的联合和隐式配准
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206607
Peng Li, S. Prince
Contemporary face recognition algorithms rely on precise localization of keypoints (corner of eye, nose etc.). Unfortunately, finding keypoints reliably and accurately remains a hard problem. In this paper we pose two questions. First, is it possible to exploit the gallery image in order to find keypoints in the probe image? For instance, consider finding the left eye in the probe image. Rather than using a generic eye model, we use a model that is informed by the appearance of the eye in the gallery image. To this end we develop a probabilistic model which combines recognition and keypoint localization. Second, is it necessary to localize keypoints? Alternatively we can consider keypoint position as a hidden variable which we marginalize over in a Bayesian manner. We demonstrate that both of these innovations improve performance relative to conventional methods in both frontal and cross-pose face recognition.
当代人脸识别算法依赖于关键点(眼角、鼻子等)的精确定位。不幸的是,可靠而准确地找到关键点仍然是一个难题。在本文中,我们提出两个问题。首先,是否有可能利用图库图像来找到探测图像中的关键点?例如,考虑在探针图像中找到左眼。而不是使用一般的眼睛模型,我们使用的模型是由画廊图像中眼睛的外观所提供的信息。为此,我们开发了一种结合识别和关键点定位的概率模型。第二,是否有必要对关键点进行本地化?或者,我们可以将关键点位置视为一个隐变量,我们用贝叶斯方法将其边缘化。我们证明了这两种创新在正面和交叉姿态面部识别方面都比传统方法提高了性能。
{"title":"Joint and implicit registration for face recognition","authors":"Peng Li, S. Prince","doi":"10.1109/CVPR.2009.5206607","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206607","url":null,"abstract":"Contemporary face recognition algorithms rely on precise localization of keypoints (corner of eye, nose etc.). Unfortunately, finding keypoints reliably and accurately remains a hard problem. In this paper we pose two questions. First, is it possible to exploit the gallery image in order to find keypoints in the probe image? For instance, consider finding the left eye in the probe image. Rather than using a generic eye model, we use a model that is informed by the appearance of the eye in the gallery image. To this end we develop a probabilistic model which combines recognition and keypoint localization. Second, is it necessary to localize keypoints? Alternatively we can consider keypoint position as a hidden variable which we marginalize over in a Bayesian manner. We demonstrate that both of these innovations improve performance relative to conventional methods in both frontal and cross-pose face recognition.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124567460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Global optimization for alignment of generalized shapes 广义形状对齐的全局优化
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206548
Hongsheng Li, Tian Shen, Xiaolei Huang
In this paper, we introduce a novel algorithm to solve global shape registration problems. We use gray-scale “images” to represent source shapes, and propose a novel two-component Gaussian Mixtures (GM) distance map representation for target shapes. Based on this flexible asymmetric image-based representation, a new energy function is defined. It proves to be a more robust shape dissimilarity metric that can be computed efficiently. Such high efficiency is essential for global optimization methods. We adopt one of them, the Particle Swarm Optimization (PSO), to effectively estimate the global optimum of the new energy function. Experiments and comparison performed on generalized shape data including continuous shapes, unstructured sparse point sets, and gradient maps, demonstrate the robustness and effectiveness of the algorithm.
本文提出了一种求解全局形状配准问题的新算法。我们使用灰度“图像”来表示源形状,并提出了一种新的双分量高斯混合(GM)距离图表示目标形状。基于这种灵活的非对称图像表示,定义了一个新的能量函数。结果表明,该方法具有较强的鲁棒性和计算效率。这种高效率对于全局优化方法是必不可少的。我们采用其中的粒子群算法(PSO)来有效估计新能量函数的全局最优。在广义形状数据(包括连续形状、非结构化稀疏点集和梯度图)上进行的实验和比较,证明了该算法的鲁棒性和有效性。
{"title":"Global optimization for alignment of generalized shapes","authors":"Hongsheng Li, Tian Shen, Xiaolei Huang","doi":"10.1109/CVPR.2009.5206548","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206548","url":null,"abstract":"In this paper, we introduce a novel algorithm to solve global shape registration problems. We use gray-scale “images” to represent source shapes, and propose a novel two-component Gaussian Mixtures (GM) distance map representation for target shapes. Based on this flexible asymmetric image-based representation, a new energy function is defined. It proves to be a more robust shape dissimilarity metric that can be computed efficiently. Such high efficiency is essential for global optimization methods. We adopt one of them, the Particle Swarm Optimization (PSO), to effectively estimate the global optimum of the new energy function. Experiments and comparison performed on generalized shape data including continuous shapes, unstructured sparse point sets, and gradient maps, demonstrate the robustness and effectiveness of the algorithm.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124595576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
An instance selection approach to Multiple instance Learning 多实例学习的实例选择方法
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206655
Zhouyu Fu, A. Robles-Kelly
Multiple-instance learning (MIL) is a new paradigm of supervised learning that deals with the classification of bags. Each bag is presented as a collection of instances from which features are extracted. In MIL, we have usually confronted with a large instance space for even moderately sized data sets since each bag may contain many instances. Hence it is important to design efficient instance pruning and selection techniques to speed up the learning process without compromising on the performance. In this paper, we address the issue of instance selection in multiple instance learning and propose the IS-MIL, an instance selection framework for MIL, to tackle large-scale MIL problems. IS-MIL is based on an alternative optimisation framework by iteratively repeating the steps of instance selection/updating and classifier learning, which is guaranteed to converge. Experimental results demonstrate the utility and efficiency of the proposed approach compared to the alternatives.
多实例学习(Multiple-instance learning, MIL)是一种处理袋分类的监督学习新范式。每个包被表示为实例的集合,从中提取特征。在MIL中,即使是中等大小的数据集,我们通常也会面临较大的实例空间,因为每个包可能包含许多实例。因此,设计有效的实例修剪和选择技术以在不影响性能的情况下加快学习过程是非常重要的。在本文中,我们解决了多实例学习中的实例选择问题,并提出了一种多实例学习的实例选择框架IS-MIL来解决大规模的MIL问题。is - mil是基于一个备选优化框架,通过迭代重复实例选择/更新和分类器学习的步骤,保证收敛。实验结果证明了该方法的实用性和有效性。
{"title":"An instance selection approach to Multiple instance Learning","authors":"Zhouyu Fu, A. Robles-Kelly","doi":"10.1109/CVPR.2009.5206655","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206655","url":null,"abstract":"Multiple-instance learning (MIL) is a new paradigm of supervised learning that deals with the classification of bags. Each bag is presented as a collection of instances from which features are extracted. In MIL, we have usually confronted with a large instance space for even moderately sized data sets since each bag may contain many instances. Hence it is important to design efficient instance pruning and selection techniques to speed up the learning process without compromising on the performance. In this paper, we address the issue of instance selection in multiple instance learning and propose the IS-MIL, an instance selection framework for MIL, to tackle large-scale MIL problems. IS-MIL is based on an alternative optimisation framework by iteratively repeating the steps of instance selection/updating and classifier learning, which is guaranteed to converge. Experimental results demonstrate the utility and efficiency of the proposed approach compared to the alternatives.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"168 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123316711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Symmetry integrated region-based image segmentation 基于区域的对称集成图像分割
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206570
Yu Sun, B. Bhanu
Symmetry is an important cue for machine perception that involves high-level knowledge of image components. Unlike most of the previous research that only computes symmetry in an image, this paper integrates symmetry with image segmentation to improve the segmentation performance. The symmetry integration is used to optimize both the segmentation and the symmetry of regions simultaneously. Interesting points are initially extracted from an image and they are further refined for detecting symmetry axis. A symmetry affinity matrix is used explicitly as a constraint in a region growing algorithm in order to refine the symmetry of segmented regions. Experimental results and comparisons from a wide domain of images indicate a promising improvement by symmetry integrated image segmentation compared to other image segmentation methods that do not exploit symmetry.
对称是机器感知的一个重要线索,它涉及图像组件的高级知识。与以往大多数研究只计算图像中的对称性不同,本文将对称性与图像分割相结合,提高了图像分割的性能。采用对称积分法同时优化区域分割和区域对称性。首先从图像中提取感兴趣的点,然后对其进行进一步细化以检测对称轴。在区域生长算法中明确地使用对称亲和矩阵作为约束,以改进分割区域的对称性。从广泛的图像领域的实验结果和比较表明,与其他不利用对称性的图像分割方法相比,对称集成图像分割有很大的改善。
{"title":"Symmetry integrated region-based image segmentation","authors":"Yu Sun, B. Bhanu","doi":"10.1109/CVPR.2009.5206570","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206570","url":null,"abstract":"Symmetry is an important cue for machine perception that involves high-level knowledge of image components. Unlike most of the previous research that only computes symmetry in an image, this paper integrates symmetry with image segmentation to improve the segmentation performance. The symmetry integration is used to optimize both the segmentation and the symmetry of regions simultaneously. Interesting points are initially extracted from an image and they are further refined for detecting symmetry axis. A symmetry affinity matrix is used explicitly as a constraint in a region growing algorithm in order to refine the symmetry of segmented regions. Experimental results and comparisons from a wide domain of images indicate a promising improvement by symmetry integrated image segmentation compared to other image segmentation methods that do not exploit symmetry.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123959175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Nonparametric scene parsing: Label transfer via dense scene alignment 非参数场景解析:通过密集场景对齐进行标签转移
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206536
Ce Liu, Jenny Yuen, A. Torralba
In this paper we propose a novel nonparametric approach for object recognition and scene parsing using dense scene alignment. Given an input image, we retrieve its best matches from a large database with annotated images using our modified, coarse-to-fine SIFT flow algorithm that aligns the structures within two images. Based on the dense scene correspondence obtained from the SIFT flow, our system warps the existing annotations, and integrates multiple cues in a Markov random field framework to segment and recognize the query image. Promising experimental results have been achieved by our nonparametric scene parsing system on a challenging database. Compared to existing object recognition approaches that require training for each object category, our system is easy to implement, has few parameters, and embeds contextual information naturally in the retrieval/alignment procedure.
本文提出了一种基于密集场景对齐的非参数目标识别和场景分析方法。给定输入图像,我们使用改进的粗到细的SIFT流算法从带有注释图像的大型数据库中检索其最佳匹配项,该算法将两幅图像中的结构对齐。基于从SIFT流中获得的密集场景对应关系,我们的系统扭曲了现有的注释,并在马尔可夫随机场框架中集成多个线索来分割和识别查询图像。我们的非参数场景分析系统在具有挑战性的数据库上取得了良好的实验结果。与现有的需要对每个对象类别进行训练的对象识别方法相比,我们的系统易于实现,参数很少,并且在检索/对齐过程中自然地嵌入上下文信息。
{"title":"Nonparametric scene parsing: Label transfer via dense scene alignment","authors":"Ce Liu, Jenny Yuen, A. Torralba","doi":"10.1109/CVPR.2009.5206536","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206536","url":null,"abstract":"In this paper we propose a novel nonparametric approach for object recognition and scene parsing using dense scene alignment. Given an input image, we retrieve its best matches from a large database with annotated images using our modified, coarse-to-fine SIFT flow algorithm that aligns the structures within two images. Based on the dense scene correspondence obtained from the SIFT flow, our system warps the existing annotations, and integrates multiple cues in a Markov random field framework to segment and recognize the query image. Promising experimental results have been achieved by our nonparametric scene parsing system on a challenging database. Compared to existing object recognition approaches that require training for each object category, our system is easy to implement, has few parameters, and embeds contextual information naturally in the retrieval/alignment procedure.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124006901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 361
Visual tracking with online Multiple Instance Learning 视觉跟踪与在线多实例学习
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206737
Boris Babenko, Ming-Hsuan Yang, Serge J. Belongie
In this paper, we address the problem of learning an adaptive appearance model for object tracking. In particular, a class of tracking techniques called “tracking by detection” have been shown to give promising results at real-time speeds. These methods train a discriminative classifier in an online manner to separate the object from the background. This classifier bootstraps itself by using the current tracker state to extract positive and negative examples from the current frame. Slight inaccuracies in the tracker can therefore lead to incorrectly labeled training examples, which degrades the classifier and can cause further drift. In this paper we show that using Multiple Instance Learning (MIL) instead of traditional supervised learning avoids these problems, and can therefore lead to a more robust tracker with fewer parameter tweaks. We present a novel online MIL algorithm for object tracking that achieves superior results with real-time performance.
在本文中,我们解决了学习自适应外观模型用于目标跟踪的问题。特别是,一类被称为“检测跟踪”的跟踪技术已被证明可以在实时速度下提供有希望的结果。这些方法以在线的方式训练一个判别分类器来分离目标和背景。这个分类器通过使用当前跟踪器状态从当前帧中提取正例和负例来引导自己。因此,跟踪器中的轻微不准确可能导致错误标记的训练样例,从而降低分类器的性能,并可能导致进一步的漂移。在本文中,我们表明使用多实例学习(MIL)代替传统的监督学习可以避免这些问题,因此可以使用更少的参数调整产生更鲁棒的跟踪器。提出了一种新的在线MIL目标跟踪算法,该算法具有较好的实时性。
{"title":"Visual tracking with online Multiple Instance Learning","authors":"Boris Babenko, Ming-Hsuan Yang, Serge J. Belongie","doi":"10.1109/CVPR.2009.5206737","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206737","url":null,"abstract":"In this paper, we address the problem of learning an adaptive appearance model for object tracking. In particular, a class of tracking techniques called “tracking by detection” have been shown to give promising results at real-time speeds. These methods train a discriminative classifier in an online manner to separate the object from the background. This classifier bootstraps itself by using the current tracker state to extract positive and negative examples from the current frame. Slight inaccuracies in the tracker can therefore lead to incorrectly labeled training examples, which degrades the classifier and can cause further drift. In this paper we show that using Multiple Instance Learning (MIL) instead of traditional supervised learning avoids these problems, and can therefore lead to a more robust tracker with fewer parameter tweaks. We present a novel online MIL algorithm for object tracking that achieves superior results with real-time performance.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128490070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1986
Stacks of convolutional Restricted Boltzmann Machines for shift-invariant feature learning 移位不变特征学习的卷积受限玻尔兹曼机堆栈
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206577
Mohammad Norouzi, Mani Ranjbar, Greg Mori
In this paper we present a method for learning class-specific features for recognition. Recently a greedy layer-wise procedure was proposed to initialize weights of deep belief networks, by viewing each layer as a separate restricted Boltzmann machine (RBM). We develop the convolutional RBM (C-RBM), a variant of the RBM model in which weights are shared to respect the spatial structure of images. This framework learns a set of features that can generate the images of a specific object class. Our feature extraction model is a four layer hierarchy of alternating filtering and maximum subsampling. We learn feature parameters of the first and third layers viewing them as separate C-RBMs. The outputs of our feature extraction hierarchy are then fed as input to a discriminative classifier. It is experimentally demonstrated that the extracted features are effective for object detection, using them to obtain performance comparable to the state of the art on handwritten digit recognition and pedestrian detection.
在本文中,我们提出了一种学习类特定特征的方法。最近提出了一种贪婪的分层方法来初始化深度信念网络的权重,该方法将每一层视为一个单独的受限玻尔兹曼机(RBM)。我们开发了卷积RBM (C-RBM),这是RBM模型的一种变体,其中权重共享以尊重图像的空间结构。这个框架学习了一组可以生成特定对象类图像的特征。我们的特征提取模型是一个交替滤波和最大子采样的四层层次结构。我们学习第一层和第三层的特征参数,将它们视为单独的c - rbm。然后,我们的特征提取层次结构的输出作为判别分类器的输入。实验证明,提取的特征对目标检测是有效的,使用它们可以获得与手写体数字识别和行人检测相媲美的性能。
{"title":"Stacks of convolutional Restricted Boltzmann Machines for shift-invariant feature learning","authors":"Mohammad Norouzi, Mani Ranjbar, Greg Mori","doi":"10.1109/CVPR.2009.5206577","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206577","url":null,"abstract":"In this paper we present a method for learning class-specific features for recognition. Recently a greedy layer-wise procedure was proposed to initialize weights of deep belief networks, by viewing each layer as a separate restricted Boltzmann machine (RBM). We develop the convolutional RBM (C-RBM), a variant of the RBM model in which weights are shared to respect the spatial structure of images. This framework learns a set of features that can generate the images of a specific object class. Our feature extraction model is a four layer hierarchy of alternating filtering and maximum subsampling. We learn feature parameters of the first and third layers viewing them as separate C-RBMs. The outputs of our feature extraction hierarchy are then fed as input to a discriminative classifier. It is experimentally demonstrated that the extracted features are effective for object detection, using them to obtain performance comparable to the state of the art on handwritten digit recognition and pedestrian detection.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130673427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 211
期刊
2009 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1