首页 > 最新文献

2015 IEEE International Conference on Computer Vision (ICCV)最新文献

英文 中文
Deformable 3D Fusion: From Partial Dynamic 3D Observations to Complete 4D Models 可变形的3D融合:从部分动态3D观察到完整的4D模型
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.252
Weipeng Xu, M. Salzmann, Yongtian Wang, Yue Liu
Capturing the 3D motion of dynamic, non-rigid objects has attracted significant attention in computer vision. Existing methods typically require either complete 3D volumetric observations, or a shape template. In this paper, we introduce a template-less 4D reconstruction method that incrementally fuses highly-incomplete 3D observations of a deforming object, and generates a complete, temporally-coherent shape representation of the object. To this end, we design an online algorithm that alternatively registers new observations to the current model estimate and updates the model. We demonstrate the effectiveness of our approach at reconstructing non-rigidly moving objects from highly-incomplete measurements on both sequences of partial 3D point clouds and Kinect videos.
捕捉动态、非刚性物体的三维运动在计算机视觉领域引起了广泛的关注。现有的方法通常需要完整的三维体积观测或形状模板。在本文中,我们介绍了一种无模板的四维重建方法,该方法逐步融合变形物体的高度不完整的三维观测,并生成物体的完整的、时间连贯的形状表示。为此,我们设计了一种在线算法,该算法交替地将新的观测值注册到当前模型估计中并更新模型。我们证明了我们的方法在重建非刚性移动物体从高度不完整的测量对部分3D点云和Kinect视频的两个序列的有效性。
{"title":"Deformable 3D Fusion: From Partial Dynamic 3D Observations to Complete 4D Models","authors":"Weipeng Xu, M. Salzmann, Yongtian Wang, Yue Liu","doi":"10.1109/ICCV.2015.252","DOIUrl":"https://doi.org/10.1109/ICCV.2015.252","url":null,"abstract":"Capturing the 3D motion of dynamic, non-rigid objects has attracted significant attention in computer vision. Existing methods typically require either complete 3D volumetric observations, or a shape template. In this paper, we introduce a template-less 4D reconstruction method that incrementally fuses highly-incomplete 3D observations of a deforming object, and generates a complete, temporally-coherent shape representation of the object. To this end, we design an online algorithm that alternatively registers new observations to the current model estimate and updates the model. We demonstrate the effectiveness of our approach at reconstructing non-rigidly moving objects from highly-incomplete measurements on both sequences of partial 3D point clouds and Kinect videos.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"19 1","pages":"2183-2191"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75673381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Semantic Segmentation of RGBD Images with Mutex Constraints 基于互斥锁约束的RGBD图像语义分割
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.202
Zhuo Deng, S. Todorovic, Longin Jan Latecki
In this paper, we address the problem of semantic scene segmentation of RGB-D images of indoor scenes. We propose a novel image region labeling method which augments CRF formulation with hard mutual exclusion (mutex) constraints. This way our approach can make use of rich and accurate 3D geometric structure coming from Kinect in a principled manner. The final labeling result must satisfy all mutex constraints, which allows us to eliminate configurations that violate common sense physics laws like placing a floor above a night stand. Three classes of mutex constraints are proposed: global object co-occurrence constraint, relative height relationship constraint, and local support relationship constraint. We evaluate our approach on the NYU-Depth V2 dataset, which consists of 1449 cluttered indoor scenes, and also test generalization of our model trained on NYU-Depth V2 dataset directly on a recent SUN3D dataset without any new training. The experimental results show that we significantly outperform the state-of-the-art methods in scene labeling on both datasets.
本文研究了室内场景RGB-D图像的语义场景分割问题。我们提出了一种新的图像区域标记方法,该方法增强了带有硬互斥约束的CRF公式。这样我们的方法就可以利用Kinect丰富而精确的3D几何结构。最终的标记结果必须满足所有互斥锁约束,这使我们能够消除违反常识性物理定律的配置,例如将地板放在床头柜上方。提出了三种类型的互斥约束:全局对象共现约束、相对高度关系约束和局部支持关系约束。我们在NYU-Depth V2数据集上评估了我们的方法,该数据集由1449个杂乱的室内场景组成,并且直接在最近的SUN3D数据集上测试了我们在NYU-Depth V2数据集上训练的模型的泛化性,而不需要任何新的训练。实验结果表明,我们在两个数据集上的场景标记明显优于最先进的方法。
{"title":"Semantic Segmentation of RGBD Images with Mutex Constraints","authors":"Zhuo Deng, S. Todorovic, Longin Jan Latecki","doi":"10.1109/ICCV.2015.202","DOIUrl":"https://doi.org/10.1109/ICCV.2015.202","url":null,"abstract":"In this paper, we address the problem of semantic scene segmentation of RGB-D images of indoor scenes. We propose a novel image region labeling method which augments CRF formulation with hard mutual exclusion (mutex) constraints. This way our approach can make use of rich and accurate 3D geometric structure coming from Kinect in a principled manner. The final labeling result must satisfy all mutex constraints, which allows us to eliminate configurations that violate common sense physics laws like placing a floor above a night stand. Three classes of mutex constraints are proposed: global object co-occurrence constraint, relative height relationship constraint, and local support relationship constraint. We evaluate our approach on the NYU-Depth V2 dataset, which consists of 1449 cluttered indoor scenes, and also test generalization of our model trained on NYU-Depth V2 dataset directly on a recent SUN3D dataset without any new training. The experimental results show that we significantly outperform the state-of-the-art methods in scene labeling on both datasets.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"60 1","pages":"1733-1741"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74466977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras 径向畸变相机极面几何的有效解
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.266
Z. Kukelova, Jan Heller, Martin Bujnak, A. Fitzgibbon, T. Pajdla
The estimation of the epipolar geometry of two cameras from image matches is a fundamental problem of computer vision with many applications. While the closely related problem of estimating relative pose of two different uncalibrated cameras with radial distortion is of particular importance, none of the previously published methods is suitable for practical applications. These solutions are either numerically unstable, sensitive to noise, based on a large number of point correspondences, or simply too slow for real-time applications. In this paper, we present a new efficient solution to this problem that uses 10 image correspondences. By manipulating ten input polynomial equations, we derive a degree 10 polynomial equation in one variable. The solutions to this equation are efficiently found using the Sturm sequences method. In the experiments, we show that the proposed solution is stable, noise resistant, and fast, and as such efficiently usable in a practical Structure-from-Motion pipeline.
从图像匹配中估计两台相机的极极几何是计算机视觉的一个基本问题,具有广泛的应用。虽然与此密切相关的具有径向畸变的两个不同的未校准相机的相对姿态估计问题尤为重要,但之前发表的方法都不适合实际应用。这些解决方案要么在数值上不稳定,对噪声敏感,要么基于大量的点对应,要么对于实时应用来说太慢。在本文中,我们提出了一种新的有效的解决方案,即使用10个图像对应。通过处理10个输入多项式方程,我们导出了一个单变量的10次多项式方程。利用Sturm序列法有效地求出了该方程的解。实验结果表明,该方法具有稳定、抗噪、快速等特点,可有效地应用于实际的运动结构管道中。
{"title":"Efficient Solution to the Epipolar Geometry for Radially Distorted Cameras","authors":"Z. Kukelova, Jan Heller, Martin Bujnak, A. Fitzgibbon, T. Pajdla","doi":"10.1109/ICCV.2015.266","DOIUrl":"https://doi.org/10.1109/ICCV.2015.266","url":null,"abstract":"The estimation of the epipolar geometry of two cameras from image matches is a fundamental problem of computer vision with many applications. While the closely related problem of estimating relative pose of two different uncalibrated cameras with radial distortion is of particular importance, none of the previously published methods is suitable for practical applications. These solutions are either numerically unstable, sensitive to noise, based on a large number of point correspondences, or simply too slow for real-time applications. In this paper, we present a new efficient solution to this problem that uses 10 image correspondences. By manipulating ten input polynomial equations, we derive a degree 10 polynomial equation in one variable. The solutions to this equation are efficiently found using the Sturm sequences method. In the experiments, we show that the proposed solution is stable, noise resistant, and fast, and as such efficiently usable in a practical Structure-from-Motion pipeline.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"87 1","pages":"2309-2317"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74614925","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Joint Optimization of Segmentation and Color Clustering 分割与颜色聚类的联合优化
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.190
E. Lobacheva, O. Veksler, Yuri Boykov
Binary energy optimization is a popular approach for segmenting a color image into foreground/background regions. To model the appearance of the regions, color, a relatively high dimensional feature, should be handled effectively. A full color histogram is usually too sparse to be reliable. One approach is to explicitly reduce dimensionality by clustering or quantizing the color space. Another popular approach is to fit GMMs for soft implicit clustering of the color space. These approaches work well when the foreground/background are sufficiently distinct. In cases of more subtle difference in appearance, both approaches may reduce or even eliminate foreground/background distinction. This happens because either color clustering is performed completely independently from the segmentation process, as a preprocessing step (in clustering), or independently for the foreground and independently for the background (in GMM). We propose to make clustering an integral part of segmentation, by including a new clustering term in the energy function. Our energy function with a clustering term favours clusterings that make foreground/background appearance more distinct. Thus our energy function jointly optimizes over color clustering, foreground/background models, and segmentation. Exact optimization is not feasible, therefore we develop an approximate algorithm. We show the advantage of including the color clustering term into the energy function on camouflage images, as well as standard segmentation datasets.
二值能量优化是将彩色图像分割为前景/背景区域的常用方法。为了对区域的外观进行建模,应该有效地处理颜色这一相对高维的特征。全彩色直方图通常过于稀疏而不可靠。一种方法是通过聚类或量化颜色空间来显式地降低维数。另一种流行的方法是为色彩空间的软隐式聚类拟合gmm。当前景/背景足够明显时,这些方法效果很好。在更细微的外观差异的情况下,这两种方法都可以减少甚至消除前景/背景的区别。这是因为颜色聚类完全独立于分割过程,作为预处理步骤(在聚类中),或者独立于前景和独立于背景(在GMM中)。我们提出通过在能量函数中加入新的聚类项,使聚类成为分割的一个组成部分。我们带有聚类术语的能量函数有利于使前景/背景外观更加明显的聚类。因此,我们的能量函数在颜色聚类、前景/背景模型和分割上共同优化。精确的优化是不可行的,因此我们开发了一个近似算法。我们展示了在迷彩图像和标准分割数据集上将颜色聚类项纳入能量函数的优势。
{"title":"Joint Optimization of Segmentation and Color Clustering","authors":"E. Lobacheva, O. Veksler, Yuri Boykov","doi":"10.1109/ICCV.2015.190","DOIUrl":"https://doi.org/10.1109/ICCV.2015.190","url":null,"abstract":"Binary energy optimization is a popular approach for segmenting a color image into foreground/background regions. To model the appearance of the regions, color, a relatively high dimensional feature, should be handled effectively. A full color histogram is usually too sparse to be reliable. One approach is to explicitly reduce dimensionality by clustering or quantizing the color space. Another popular approach is to fit GMMs for soft implicit clustering of the color space. These approaches work well when the foreground/background are sufficiently distinct. In cases of more subtle difference in appearance, both approaches may reduce or even eliminate foreground/background distinction. This happens because either color clustering is performed completely independently from the segmentation process, as a preprocessing step (in clustering), or independently for the foreground and independently for the background (in GMM). We propose to make clustering an integral part of segmentation, by including a new clustering term in the energy function. Our energy function with a clustering term favours clusterings that make foreground/background appearance more distinct. Thus our energy function jointly optimizes over color clustering, foreground/background models, and segmentation. Exact optimization is not feasible, therefore we develop an approximate algorithm. We show the advantage of including the color clustering term into the energy function on camouflage images, as well as standard segmentation datasets.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"16 1","pages":"1626-1634"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75295645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Mining And-Or Graphs for Graph Matching and Object Discovery 用于图匹配和对象发现的and - or图挖掘
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.15
Quanshi Zhang, Y. Wu, Song-Chun Zhu
This paper reformulates the theory of graph mining on the technical basis of graph matching, and extends its scope of applications to computer vision. Given a set of attributed relational graphs (ARGs), we propose to use a hierarchical And-Or Graph (AoG) to model the pattern of maximal-size common subgraphs embedded in the ARGs, and we develop a general method to mine the AoG model from the unlabeled ARGs. This method provides a general solution to the problem of mining hierarchical models from unannotated visual data without exhaustive search of objects. We apply our method to RGB/RGB-D images and videos to demonstrate its generality and the wide range of applicability. The code will be available at https://sites.google.com/site/quanshizhang/mining-and-or-graphs.
本文在图匹配的技术基础上重新阐述了图挖掘理论,并将其应用范围扩展到计算机视觉。在给定一组属性关系图(arg)的基础上,我们提出了一种分层的and - or图(AoG)来建模嵌入在arg中的最大尺寸公共子图的模式,并开发了一种从未标记的arg中挖掘AoG模型的通用方法。该方法为从无注释的可视化数据中挖掘层次模型问题提供了一种通用的解决方案,而无需对对象进行穷举搜索。我们将该方法应用于RGB/RGB- d图像和视频,以证明其通用性和广泛的适用性。代码可在https://sites.google.com/site/quanshizhang/mining-and-or-graphs上获得。
{"title":"Mining And-Or Graphs for Graph Matching and Object Discovery","authors":"Quanshi Zhang, Y. Wu, Song-Chun Zhu","doi":"10.1109/ICCV.2015.15","DOIUrl":"https://doi.org/10.1109/ICCV.2015.15","url":null,"abstract":"This paper reformulates the theory of graph mining on the technical basis of graph matching, and extends its scope of applications to computer vision. Given a set of attributed relational graphs (ARGs), we propose to use a hierarchical And-Or Graph (AoG) to model the pattern of maximal-size common subgraphs embedded in the ARGs, and we develop a general method to mine the AoG model from the unlabeled ARGs. This method provides a general solution to the problem of mining hierarchical models from unannotated visual data without exhaustive search of objects. We apply our method to RGB/RGB-D images and videos to demonstrate its generality and the wide range of applicability. The code will be available at https://sites.google.com/site/quanshizhang/mining-and-or-graphs.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"44 1","pages":"55-63"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74187516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Multi-view Subspace Clustering 多视图子空间聚类
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.482
Hongchang Gao, F. Nie, Xuelong Li, Heng Huang
For many computer vision applications, the data sets distribute on certain low-dimensional subspaces. Subspace clustering is to find such underlying subspaces and cluster the data points correctly. In this paper, we propose a novel multi-view subspace clustering method. The proposed method performs clustering on the subspace representation of each view simultaneously. Meanwhile, we propose to use a common cluster structure to guarantee the consistence among different views. In addition, an efficient algorithm is proposed to solve the problem. Experiments on four benchmark data sets have been performed to validate our proposed method. The promising results demonstrate the effectiveness of our method.
对于许多计算机视觉应用,数据集分布在一定的低维子空间上。子空间聚类就是找到这样的底层子空间,并对数据点进行正确聚类。本文提出了一种新的多视图子空间聚类方法。该方法同时对每个视图的子空间表示进行聚类。同时,我们提出使用一个通用的聚类结构来保证不同视图之间的一致性。此外,还提出了一种有效的算法来解决这一问题。在四个基准数据集上进行了实验,验证了我们提出的方法。结果表明了该方法的有效性。
{"title":"Multi-view Subspace Clustering","authors":"Hongchang Gao, F. Nie, Xuelong Li, Heng Huang","doi":"10.1109/ICCV.2015.482","DOIUrl":"https://doi.org/10.1109/ICCV.2015.482","url":null,"abstract":"For many computer vision applications, the data sets distribute on certain low-dimensional subspaces. Subspace clustering is to find such underlying subspaces and cluster the data points correctly. In this paper, we propose a novel multi-view subspace clustering method. The proposed method performs clustering on the subspace representation of each view simultaneously. Meanwhile, we propose to use a common cluster structure to guarantee the consistence among different views. In addition, an efficient algorithm is proposed to solve the problem. Experiments on four benchmark data sets have been performed to validate our proposed method. The promising results demonstrate the effectiveness of our method.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"10 1","pages":"4238-4246"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74851068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 361
Video Restoration Against Yin-Yang Phasing 针对阴阳相位的视频复原
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.70
Xiaolin Wu, Zhenhao Li, Xiaowei Deng
A common video degradation problem, which is largely untreated in literature, is what we call Yin-Yang Phasing (YYP). YYP is characterized by involuntary, dramatic flip-flop in the intensity and possibly chromaticity of an object as the video plays. Such temporal artifacts occur under ill illumination conditions and are triggered by object or/and camera motions, which mislead the settings of camera's auto-exposure and white point. In this paper, we investigate the problem and propose a video restoration technique to suppress YYP artifacts and retain temporal consistency of objects appearance via inter-frame, spatially-adaptive, optimal tone mapping. The video quality can be further improved by a novel image enhancer designed in Weber's perception principle and by exploiting the second-order statistics of the scene. Experimental results are encouraging, pointing to an effective, practical solution for a common but surprisingly understudied problem.
一个常见的视频退化问题,在文献中基本上没有得到处理,就是我们所说的阴阳相位(YYP)。yp的特点是在视频播放时,物体的强度和可能的色度会发生无意识的戏剧性变化。这种时间伪影发生在光线不足的条件下,由物体或/和相机运动触发,从而误导相机的自动曝光和白点设置。在本文中,我们研究了这个问题,并提出了一种视频恢复技术,通过帧间、空间自适应、最佳色调映射来抑制YYP伪影,并保持物体外观的时间一致性。利用韦伯感知原理设计的新型图像增强器和利用场景的二阶统计量,可以进一步提高视频质量。实验结果令人鼓舞,为一个常见但令人惊讶的未被充分研究的问题指出了一个有效、实用的解决方案。
{"title":"Video Restoration Against Yin-Yang Phasing","authors":"Xiaolin Wu, Zhenhao Li, Xiaowei Deng","doi":"10.1109/ICCV.2015.70","DOIUrl":"https://doi.org/10.1109/ICCV.2015.70","url":null,"abstract":"A common video degradation problem, which is largely untreated in literature, is what we call Yin-Yang Phasing (YYP). YYP is characterized by involuntary, dramatic flip-flop in the intensity and possibly chromaticity of an object as the video plays. Such temporal artifacts occur under ill illumination conditions and are triggered by object or/and camera motions, which mislead the settings of camera's auto-exposure and white point. In this paper, we investigate the problem and propose a video restoration technique to suppress YYP artifacts and retain temporal consistency of objects appearance via inter-frame, spatially-adaptive, optimal tone mapping. The video quality can be further improved by a novel image enhancer designed in Weber's perception principle and by exploiting the second-order statistics of the scene. Experimental results are encouraging, pointing to an effective, practical solution for a common but surprisingly understudied problem.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"196 1","pages":"549-557"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77266687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Continuous Pose Estimation with a Spatial Ensemble of Fisher Regressors 基于Fisher回归量空间集合的连续姿态估计
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.124
Michele Fenzi, L. Leal-Taixé, J. Ostermann, T. Tuytelaars
In this paper, we treat the problem of continuous pose estimation for object categories as a regression problem on the basis of only 2D training information. While regression is a natural framework for continuous problems, regression methods so far achieved inferior results with respect to 3D-based and 2D-based classification-and-refinement approaches. This may be attributed to their weakness to high intra-class variability as well as to noisy matching procedures and lack of geometrical constraints. We propose to apply regression to Fisher-encoded vectors computed from large cells by learning an array of Fisher regressors. Fisher encoding makes our algorithm flexible to variations in class appearance, while the array structure permits to indirectly introduce spatial context information in the approach. We formulate our problem as a MAP inference problem, where the likelihood function is composed of a generative term based on the prediction error generated by the ensemble of Fisher regressors as well as a discriminative term based on SVM classifiers. We test our algorithm on three publicly available datasets that envisage several difficulties, such as high intra-class variability, truncations, occlusions, and motion blur, obtaining state-of-the-art results.
在本文中,我们将对象类别的连续姿态估计问题视为仅基于二维训练信息的回归问题。虽然回归是处理连续问题的自然框架,但迄今为止,相对于基于3d和基于2d的分类和细化方法,回归方法的效果较差。这可能是由于它们的弱点是高类内变异性,以及嘈杂的匹配过程和缺乏几何约束。我们建议通过学习一组Fisher回归量,将回归应用于从大细胞计算的Fisher编码向量。Fisher编码使我们的算法能够灵活地适应类外观的变化,而数组结构允许在方法中间接引入空间上下文信息。我们将我们的问题表述为MAP推理问题,其中似然函数由基于Fisher回归集合生成的预测误差的生成项和基于SVM分类器的判别项组成。我们在三个公开可用的数据集上测试了我们的算法,这些数据集设想了几个困难,如高类内可变性、截断、闭塞和运动模糊,获得了最先进的结果。
{"title":"Continuous Pose Estimation with a Spatial Ensemble of Fisher Regressors","authors":"Michele Fenzi, L. Leal-Taixé, J. Ostermann, T. Tuytelaars","doi":"10.1109/ICCV.2015.124","DOIUrl":"https://doi.org/10.1109/ICCV.2015.124","url":null,"abstract":"In this paper, we treat the problem of continuous pose estimation for object categories as a regression problem on the basis of only 2D training information. While regression is a natural framework for continuous problems, regression methods so far achieved inferior results with respect to 3D-based and 2D-based classification-and-refinement approaches. This may be attributed to their weakness to high intra-class variability as well as to noisy matching procedures and lack of geometrical constraints. We propose to apply regression to Fisher-encoded vectors computed from large cells by learning an array of Fisher regressors. Fisher encoding makes our algorithm flexible to variations in class appearance, while the array structure permits to indirectly introduce spatial context information in the approach. We formulate our problem as a MAP inference problem, where the likelihood function is composed of a generative term based on the prediction error generated by the ensemble of Fisher regressors as well as a discriminative term based on SVM classifiers. We test our algorithm on three publicly available datasets that envisage several difficulties, such as high intra-class variability, truncations, occlusions, and motion blur, obtaining state-of-the-art results.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"164 1","pages":"1035-1043"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77270753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Guaranteed Outlier Removal for Rotation Search 保证离群值去除旋转搜索
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.250
Álvaro Parra, Tat-Jun Chin
Rotation search has become a core routine for solving many computer vision problems. The aim is to rotationally align two input point sets with correspondences. Recently, there is significant interest in developing globally optimal rotation search algorithms. A notable weakness of global algorithms, however, is their relatively high computational cost, especially on large problem sizes and data with a high proportion of outliers. In this paper, we propose a novel outlier removal technique for rotation search. Our method guarantees that any correspondence it discards as an outlier does not exist in the inlier set of the globally optimal rotation for the original data. Based on simple geometric operations, our algorithm is deterministic and fast. Used as a preprocessor to prune a large portion of the outliers from the input data, our method enables substantial speed-up of rotation search algorithms without compromising global optimality. We demonstrate the efficacy of our method in various synthetic and real data experiments.
旋转搜索已经成为解决许多计算机视觉问题的核心程序。目的是旋转对齐两个输入点集对应。最近,人们对开发全局最优旋转搜索算法非常感兴趣。然而,全局算法的一个显著缺点是其相对较高的计算成本,特别是在大问题规模和具有高比例异常值的数据时。本文提出了一种新的旋转搜索异常值去除技术。我们的方法保证它作为离群值丢弃的任何对应都不存在于原始数据的全局最优旋转的内嵌集中。该算法基于简单的几何运算,具有确定性和快速的特点。我们的方法用作预处理器,从输入数据中剔除大部分异常值,从而大大加快了旋转搜索算法的速度,同时又不影响全局最优性。在各种合成实验和实际数据实验中证明了该方法的有效性。
{"title":"Guaranteed Outlier Removal for Rotation Search","authors":"Álvaro Parra, Tat-Jun Chin","doi":"10.1109/ICCV.2015.250","DOIUrl":"https://doi.org/10.1109/ICCV.2015.250","url":null,"abstract":"Rotation search has become a core routine for solving many computer vision problems. The aim is to rotationally align two input point sets with correspondences. Recently, there is significant interest in developing globally optimal rotation search algorithms. A notable weakness of global algorithms, however, is their relatively high computational cost, especially on large problem sizes and data with a high proportion of outliers. In this paper, we propose a novel outlier removal technique for rotation search. Our method guarantees that any correspondence it discards as an outlier does not exist in the inlier set of the globally optimal rotation for the original data. Based on simple geometric operations, our algorithm is deterministic and fast. Used as a preprocessor to prune a large portion of the outliers from the input data, our method enables substantial speed-up of rotation search algorithms without compromising global optimality. We demonstrate the efficacy of our method in various synthetic and real data experiments.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"10 1","pages":"2165-2173"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75017051","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Just Noticeable Differences in Visual Attributes 只是视觉属性的明显差异
Pub Date : 2015-12-07 DOI: 10.1109/ICCV.2015.278
Aron Yu, K. Grauman
We explore the problem of predicting "just noticeable differences" in a visual attribute. While some pairs of images have a clear ordering for an attribute (e.g., A is more sporty than B), for others the difference may be indistinguishable to human observers. However, existing relative attribute models are unequipped to infer partial orders on novel data. Attempting to map relative attribute ranks to equality predictions is non-trivial, particularly since the span of indistinguishable pairs in attribute space may vary in different parts of the feature space. We develop a Bayesian local learning strategy to infer when images are indistinguishable for a given attribute. On the UT-Zap50K shoes and LFW-10 faces datasets, we outperform a variety of alternative methods. In addition, we show the practical impact on fine-grained visual search.
我们探讨了在视觉属性中预测“只是明显差异”的问题。虽然有些图像对属性有明确的顺序(例如,a比B更运动),但对于其他图像,人类观察者可能无法区分差异。然而,现有的相对属性模型不具备在新数据上推断偏序的能力。试图将相对属性等级映射到相等性预测是非常重要的,特别是因为属性空间中不可区分的对的跨度可能在特征空间的不同部分变化。我们开发了一种贝叶斯局部学习策略来推断图像何时对给定属性不可区分。在UT-Zap50K鞋和LFW-10人脸数据集上,我们的表现优于各种替代方法。此外,我们还展示了对细粒度视觉搜索的实际影响。
{"title":"Just Noticeable Differences in Visual Attributes","authors":"Aron Yu, K. Grauman","doi":"10.1109/ICCV.2015.278","DOIUrl":"https://doi.org/10.1109/ICCV.2015.278","url":null,"abstract":"We explore the problem of predicting \"just noticeable differences\" in a visual attribute. While some pairs of images have a clear ordering for an attribute (e.g., A is more sporty than B), for others the difference may be indistinguishable to human observers. However, existing relative attribute models are unequipped to infer partial orders on novel data. Attempting to map relative attribute ranks to equality predictions is non-trivial, particularly since the span of indistinguishable pairs in attribute space may vary in different parts of the feature space. We develop a Bayesian local learning strategy to infer when images are indistinguishable for a given attribute. On the UT-Zap50K shoes and LFW-10 faces datasets, we outperform a variety of alternative methods. In addition, we show the practical impact on fine-grained visual search.","PeriodicalId":6633,"journal":{"name":"2015 IEEE International Conference on Computer Vision (ICCV)","volume":"82 1","pages":"2416-2424"},"PeriodicalIF":0.0,"publicationDate":"2015-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77651405","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
期刊
2015 IEEE International Conference on Computer Vision (ICCV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1