首页 > 最新文献

2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)最新文献

英文 中文
Stereo Vision in Structured Environments by Consistent Semi-Global Matching 基于一致半全局匹配的结构化环境立体视觉
H. Hirschmüller
This paper considers the use of stereo vision in structured environments. Sharp discontinuities and large untextured areas must be anticipated, but complex or natural shapes of objects and fine structures should be handled as well. Additionally, radiometric differences of input images often occur in practice. Finally, computation time is an issue for handling large or many images in acceptable time. The Semi-Global Matching method is chosen as it fulfills already many of the requirements. Remaining problems in structured environments are carefully analyzed and two novel extensions suggested. Firstly, intensity consistent disparity selection is proposed for handling untextured areas. Secondly, discontinuity preserving interpolation is suggested for filling holes in the disparity images that are caused by some filters. It is shown that the performance of the new method on test images with ground truth is comparable to the currently best stereo methods, but the complexity and runtime is much lower.
本文考虑了立体视觉在结构化环境中的应用。必须预料到尖锐的不连续和大面积的无纹理区域,但也应处理复杂或自然形状的物体和精细结构。此外,在实际应用中,经常会出现输入图像的辐射差异。最后,计算时间是在可接受的时间内处理大型或许多图像的一个问题。选择半全局匹配方法是因为它已经满足了许多要求。仔细分析了结构化环境中存在的问题,并提出了两个新的扩展。首先,提出了处理非纹理区域的强度一致视差选择方法。其次,针对某些滤波器在视差图像中造成的孔洞,提出了保持不连续的插值方法。实验结果表明,该方法在具有地面真值的测试图像上的性能与目前最好的立体图像方法相当,但复杂度和运行时间大大降低。
{"title":"Stereo Vision in Structured Environments by Consistent Semi-Global Matching","authors":"H. Hirschmüller","doi":"10.1109/CVPR.2006.294","DOIUrl":"https://doi.org/10.1109/CVPR.2006.294","url":null,"abstract":"This paper considers the use of stereo vision in structured environments. Sharp discontinuities and large untextured areas must be anticipated, but complex or natural shapes of objects and fine structures should be handled as well. Additionally, radiometric differences of input images often occur in practice. Finally, computation time is an issue for handling large or many images in acceptable time. The Semi-Global Matching method is chosen as it fulfills already many of the requirements. Remaining problems in structured environments are carefully analyzed and two novel extensions suggested. Firstly, intensity consistent disparity selection is proposed for handling untextured areas. Secondly, discontinuity preserving interpolation is suggested for filling holes in the disparity images that are caused by some filters. It is shown that the performance of the new method on test images with ground truth is comparable to the currently best stereo methods, but the complexity and runtime is much lower.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127879353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 182
CSIFT: A SIFT Descriptor with Color Invariant Characteristics 具有颜色不变性特征的SIFT描述符
Alaa E. Abdel-Hakim, A. Farag
SIFT has been proven to be the most robust local invariant feature descriptor. SIFT is designed mainly for gray images. However, color provides valuable information in object description and matching tasks. Many objects can be misclassified if their color contents are ignored. This paper addresses this problem and proposes a novel colored local invariant feature descriptor. Instead of using the gray space to represent the input image, the proposed approach builds the SIFT descriptors in a color invariant space. The built Colored SIFT (CSIFT) is more robust than the conventional SIFT with respect to color and photometrical variations. The evaluation results support the potential of the proposed approach.
SIFT已被证明是最鲁棒的局部不变特征描述子。SIFT主要是针对灰度图像设计的。然而,颜色在物体描述和匹配任务中提供了有价值的信息。如果忽略许多物体的颜色内容,它们可能会被错误分类。针对这一问题,本文提出了一种新的彩色局部不变特征描述子。该方法不是使用灰度空间来表示输入图像,而是在颜色不变空间中构建SIFT描述子。所构建的彩色SIFT (CSIFT)在颜色和光度变化方面比传统SIFT具有更强的鲁棒性。评价结果支持了该方法的潜力。
{"title":"CSIFT: A SIFT Descriptor with Color Invariant Characteristics","authors":"Alaa E. Abdel-Hakim, A. Farag","doi":"10.1109/CVPR.2006.95","DOIUrl":"https://doi.org/10.1109/CVPR.2006.95","url":null,"abstract":"SIFT has been proven to be the most robust local invariant feature descriptor. SIFT is designed mainly for gray images. However, color provides valuable information in object description and matching tasks. Many objects can be misclassified if their color contents are ignored. This paper addresses this problem and proposes a novel colored local invariant feature descriptor. Instead of using the gray space to represent the input image, the proposed approach builds the SIFT descriptors in a color invariant space. The built Colored SIFT (CSIFT) is more robust than the conventional SIFT with respect to color and photometrical variations. The evaluation results support the potential of the proposed approach.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129194162","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 617
Affine Invariance Revisited 回顾仿射不变性
Evgeni Begelfor, M. Werman
This paper proposes a Riemannian geometric framework to compute averages and distributions of point configurations so that different configurations up to affine transformations are considered to be the same. The algorithms are fast and proven to be robust both theoretically and empirically. The utility of this framework is shown in a number of affine invariant clustering algorithms on image point data.
本文提出了一种计算点构型的平均和分布的黎曼几何框架,使不同的构型直到仿射变换都被认为是相同的。该算法速度快,具有理论和经验上的鲁棒性。该框架的实用性体现在图像点数据的仿射不变聚类算法中。
{"title":"Affine Invariance Revisited","authors":"Evgeni Begelfor, M. Werman","doi":"10.1109/CVPR.2006.50","DOIUrl":"https://doi.org/10.1109/CVPR.2006.50","url":null,"abstract":"This paper proposes a Riemannian geometric framework to compute averages and distributions of point configurations so that different configurations up to affine transformations are considered to be the same. The algorithms are fast and proven to be robust both theoretically and empirically. The utility of this framework is shown in a number of affine invariant clustering algorithms on image point data.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126599475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 158
Scalable Recognition with a Vocabulary Tree 可扩展的识别与词汇树
D. Nistér, Henrik Stewénius
A recognition scheme that scales efficiently to a large number of objects is presented. The efficiency and quality is exhibited in a live demonstration that recognizes CD-covers from a database of 40000 images of popular music CD’s. The scheme builds upon popular techniques of indexing descriptors extracted from local regions, and is robust to background clutter and occlusion. The local region descriptors are hierarchically quantized in a vocabulary tree. The vocabulary tree allows a larger and more discriminatory vocabulary to be used efficiently, which we show experimentally leads to a dramatic improvement in retrieval quality. The most significant property of the scheme is that the tree directly defines the quantization. The quantization and the indexing are therefore fully integrated, essentially being one and the same. The recognition quality is evaluated through retrieval on a database with ground truth, showing the power of the vocabulary tree approach, going as high as 1 million images.
提出了一种适用于大量目标的有效识别方案。在一个现场演示中,从一个包含40000张流行音乐CD图像的数据库中识别CD封面,展示了这种方法的效率和质量。该方案建立在从局部区域提取索引描述符的流行技术的基础上,并且对背景杂波和遮挡具有鲁棒性。局部区域描述符在词汇树中分层量化。词汇树允许更大、更具歧视性的词汇被有效地使用,我们通过实验证明,这导致了检索质量的显著提高。该方案最重要的性质是树直接定义了量化。因此,量化和索引是完全集成的,本质上是相同的。识别质量通过在数据库中检索来评估,显示了词汇树方法的强大功能,最高可达100万张图像。
{"title":"Scalable Recognition with a Vocabulary Tree","authors":"D. Nistér, Henrik Stewénius","doi":"10.1109/CVPR.2006.264","DOIUrl":"https://doi.org/10.1109/CVPR.2006.264","url":null,"abstract":"A recognition scheme that scales efficiently to a large number of objects is presented. The efficiency and quality is exhibited in a live demonstration that recognizes CD-covers from a database of 40000 images of popular music CD’s. The scheme builds upon popular techniques of indexing descriptors extracted from local regions, and is robust to background clutter and occlusion. The local region descriptors are hierarchically quantized in a vocabulary tree. The vocabulary tree allows a larger and more discriminatory vocabulary to be used efficiently, which we show experimentally leads to a dramatic improvement in retrieval quality. The most significant property of the scheme is that the tree directly defines the quantization. The quantization and the indexing are therefore fully integrated, essentially being one and the same. The recognition quality is evaluated through retrieval on a database with ground truth, showing the power of the vocabulary tree approach, going as high as 1 million images.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"331 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123230046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4019
Region-Tree Based Stereo Using Dynamic Programming Optimization 基于区域树的立体视觉动态规划优化
C. Lei, Jason M. Selzer, Herbert Yang
In this paper, we present a novel stereo algorithm that combines the strengths of region-based stereo and dynamic programming on a tree approaches. Instead of formulating an image as individual scan-lines or as a pixel tree, a new region tree structure, which is built as a minimum spanning tree on the adjacency-graph of an over-segmented image, is used for the global dynamic programming optimization. The resulting disparity maps do not contain any streaking problem as is common in scanline-based algorithms because of the tree structure. The performance evaluation using the Middlebury benchmark datasets shows that the performance of our algorithm is comparable in accuracy and efficiency with top ranking algorithms.
在本文中,我们提出了一种新的立体算法,它结合了基于区域的立体和基于树的动态规划方法的优点。采用一种新的区域树结构,在过分割图像的邻接图上建立最小生成树,而不是将图像表述为单个扫描线或像素树,用于全局动态规划优化。由于树形结构,所产生的视差图不包含任何条纹问题,这在基于扫描线的算法中很常见。使用Middlebury基准数据集进行的性能评估表明,我们的算法在精度和效率方面与排名靠前的算法相当。
{"title":"Region-Tree Based Stereo Using Dynamic Programming Optimization","authors":"C. Lei, Jason M. Selzer, Herbert Yang","doi":"10.1109/CVPR.2006.251","DOIUrl":"https://doi.org/10.1109/CVPR.2006.251","url":null,"abstract":"In this paper, we present a novel stereo algorithm that combines the strengths of region-based stereo and dynamic programming on a tree approaches. Instead of formulating an image as individual scan-lines or as a pixel tree, a new region tree structure, which is built as a minimum spanning tree on the adjacency-graph of an over-segmented image, is used for the global dynamic programming optimization. The resulting disparity maps do not contain any streaking problem as is common in scanline-based algorithms because of the tree structure. The performance evaluation using the Middlebury benchmark datasets shows that the performance of our algorithm is comparable in accuracy and efficiency with top ranking algorithms.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114129565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 140
Learning Temporal Sequence Model from Partially Labeled Data 从部分标记数据中学习时间序列模型
Yifan Shi, A. Bobick, Irfan Essa
Graphical models are often used to represent and recognize activities. Purely unsupervised methods (such as HMMs) can be trained automatically but yield models whose internal structure - the nodes - are difficult to interpret semantically. Manually constructed networks typically have nodes corresponding to sub-events, but the programming and training of these networks is tedious and requires extensive domain expertise. In this paper, we propose a semi-supervised approach in which a manually structured, Propagation Network (a form of a DBN) is initialized from a small amount of fully annotated data, and then refined by an EM-based learning method in an unsupervised fashion. During node refinement (the M step) a boosting-based algorithm is employed to train the evidence detectors of individual nodes. Experiments on a variety of data types - vision and inertial measurements - in several tasks demonstrate the ability to learn from as little as one fully annotated example accompanied by a small number of positive but non-annotated training examples. The system is applied to both recognition and anomaly detection tasks.
图形模型通常用于表示和识别活动。纯粹的无监督方法(比如hmm)可以自动训练,但是产生的模型的内部结构——节点——很难从语义上解释。人工构建的网络通常具有对应于子事件的节点,但是这些网络的编程和训练是繁琐的,并且需要广泛的领域专业知识。在本文中,我们提出了一种半监督方法,其中手动结构化的传播网络(DBN的一种形式)从少量完全注释的数据初始化,然后通过基于em的学习方法以无监督的方式进行改进。在节点细化(M步)过程中,采用基于增强的算法来训练单个节点的证据检测器。在几个任务中,对各种数据类型(视觉和惯性测量)进行的实验证明了从一个完全注释的示例中学习的能力,并伴随着少量正面但未注释的训练示例。该系统可用于识别和异常检测任务。
{"title":"Learning Temporal Sequence Model from Partially Labeled Data","authors":"Yifan Shi, A. Bobick, Irfan Essa","doi":"10.1109/CVPR.2006.174","DOIUrl":"https://doi.org/10.1109/CVPR.2006.174","url":null,"abstract":"Graphical models are often used to represent and recognize activities. Purely unsupervised methods (such as HMMs) can be trained automatically but yield models whose internal structure - the nodes - are difficult to interpret semantically. Manually constructed networks typically have nodes corresponding to sub-events, but the programming and training of these networks is tedious and requires extensive domain expertise. In this paper, we propose a semi-supervised approach in which a manually structured, Propagation Network (a form of a DBN) is initialized from a small amount of fully annotated data, and then refined by an EM-based learning method in an unsupervised fashion. During node refinement (the M step) a boosting-based algorithm is employed to train the evidence detectors of individual nodes. Experiments on a variety of data types - vision and inertial measurements - in several tasks demonstrate the ability to learn from as little as one fully annotated example accompanied by a small number of positive but non-annotated training examples. The system is applied to both recognition and anomaly detection tasks.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116681052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
A Geodesic Active Contour Framework for Finding Glass 一种用于寻找玻璃的测地线主动轮廓框架
Kenton McHenry, J. Ponce
This paper addresses the problem of finding objects made of glass (or other transparent materials) in images. Since the appearance of glass objects depends for the most part on what lies behind them, we propose to use binary criteria ("are these two regions made of the same material?") rather than unary ones ("is this glass?") to guide the segmentation process. Concretely, we combine two complementary measures of affinity between regions made of the same material and discrepancy between regions made of different ones into a single objective function, and use the geodesic active contour framework to minimize this function over pixel labels. The proposed approach has been implemented, and qualitative and quantitative experimental results are presented.
本文解决了在图像中寻找由玻璃(或其他透明材料)制成的物体的问题。由于玻璃物体的外观在很大程度上取决于它们背后的东西,我们建议使用二元标准(“这两个区域是由相同的材料制成的吗?”)而不是一元标准(“这是玻璃吗?”)来指导分割过程。具体而言,我们将相同材料区域之间的亲和力和不同材料区域之间的差异这两个互补的度量结合到一个单一的目标函数中,并使用测地线活动轮廓框架在像素标签上最小化该函数。所提出的方法已经实现,并给出了定性和定量的实验结果。
{"title":"A Geodesic Active Contour Framework for Finding Glass","authors":"Kenton McHenry, J. Ponce","doi":"10.1109/CVPR.2006.28","DOIUrl":"https://doi.org/10.1109/CVPR.2006.28","url":null,"abstract":"This paper addresses the problem of finding objects made of glass (or other transparent materials) in images. Since the appearance of glass objects depends for the most part on what lies behind them, we propose to use binary criteria (\"are these two regions made of the same material?\") rather than unary ones (\"is this glass?\") to guide the segmentation process. Concretely, we combine two complementary measures of affinity between regions made of the same material and discrepancy between regions made of different ones into a single objective function, and use the geodesic active contour framework to minimize this function over pixel labels. The proposed approach has been implemented, and qualitative and quantitative experimental results are presented.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116701319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Specular Flow and the Recovery of Surface Structure 镜面流和表面结构的恢复
S. Roth, Michael J. Black
In scenes containing specular objects, the image motion observed by a moving camera may be an intermixed combination of optical flow resulting from diffuse reflectance (diffuse flow) and specular reflection (specular flow). Here, with few assumptions, we formalize the notion of specular flow, show how it relates to the 3D structure of the world, and develop an algorithm for estimating scene structure from 2D image motion. Unlike previous work on isolated specular highlights we use two image frames and estimate the semi-dense flow arising from the specular reflections of textured scenes. We parametrically model the image motion of a quadratic surface patch viewed from a moving camera. The flow is modeled as a probabilistic mixture of diffuse and specular components and the 3D shape is recovered using an Expectation-Maximization algorithm. Rather than treating specular reflections as noise to be removed or ignored, we show that the specular flow provides additional constraints on scene geometry that improve estimation of 3D structure when compared with reconstruction from diffuse flow alone. We demonstrate this for a set of synthetic and real sequences of mixed specular-diffuse objects.
在包含高光物体的场景中,移动摄像机观察到的图像运动可能是由漫反射(漫反射)和高光反射(高光流)产生的光流的混合组合。在这里,通过一些假设,我们形式化了镜面流的概念,展示了它与世界的3D结构的关系,并开发了一种从2D图像运动中估计场景结构的算法。不像以前的工作在孤立的高光,我们使用两个图像帧和估计由纹理场景的高光反射产生的半密集流。我们对从运动相机中观察到的二次曲面斑块的像运动进行了参数化建模。该流被建模为漫射和镜面成分的概率混合,并使用期望最大化算法恢复三维形状。我们没有将镜面反射视为需要去除或忽略的噪声,而是表明,与仅通过漫射流进行重建相比,镜面流为场景几何提供了额外的约束,从而提高了对3D结构的估计。我们为一组合成和真实的混合镜面漫射物体序列证明了这一点。
{"title":"Specular Flow and the Recovery of Surface Structure","authors":"S. Roth, Michael J. Black","doi":"10.1109/CVPR.2006.290","DOIUrl":"https://doi.org/10.1109/CVPR.2006.290","url":null,"abstract":"In scenes containing specular objects, the image motion observed by a moving camera may be an intermixed combination of optical flow resulting from diffuse reflectance (diffuse flow) and specular reflection (specular flow). Here, with few assumptions, we formalize the notion of specular flow, show how it relates to the 3D structure of the world, and develop an algorithm for estimating scene structure from 2D image motion. Unlike previous work on isolated specular highlights we use two image frames and estimate the semi-dense flow arising from the specular reflections of textured scenes. We parametrically model the image motion of a quadratic surface patch viewed from a moving camera. The flow is modeled as a probabilistic mixture of diffuse and specular components and the 3D shape is recovered using an Expectation-Maximization algorithm. Rather than treating specular reflections as noise to be removed or ignored, we show that the specular flow provides additional constraints on scene geometry that improve estimation of 3D structure when compared with reconstruction from diffuse flow alone. We demonstrate this for a set of synthetic and real sequences of mixed specular-diffuse objects.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123757365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 95
Combined Depth and Outlier Estimation in Multi-View Stereo 多视点立体图像中深度与离群值的组合估计
C. Strecha, R. Fransens, L. Gool
In this paper, we present a generative model based approach to solve the multi-view stereo problem. The input images are considered to be generated by either one of two processes: (i) an inlier process, which generates the pixels which are visible from the reference camera and which obey the constant brightness assumption, and (ii) an outlier process which generates all other pixels. Depth and visibility are jointly modelled as a hiddenMarkov Random Field, and the spatial correlations of both are explicitly accounted for. Inference is made tractable by an EM-algorithm, which alternates between estimation of visibility and depth, and optimisation of model parameters. We describe and compare two implementations of the E-step of the algorithm, which correspond to the Mean Field and Bethe approximations of the free energy. The approach is validated by experiments on challenging real-world scenes, of which two are contaminated by independently moving objects.
在本文中,我们提出了一种基于生成模型的方法来解决多视图立体问题。输入图像被认为是由以下两个过程之一生成的:(i)一个内部过程,它生成从参考相机可见的像素,并且服从恒定亮度假设,以及(ii)一个异常过程,它生成所有其他像素。深度和能见度被联合建模为一个隐马尔可夫随机场,并且两者的空间相关性被明确地解释。通过em算法使推理变得易于处理,该算法在可见性和深度估计以及模型参数优化之间交替进行。我们描述并比较了该算法的e步的两种实现,它们对应于自由能的平均场近似和贝特近似。该方法在具有挑战性的真实场景中得到了验证,其中两个场景被独立移动的物体污染。
{"title":"Combined Depth and Outlier Estimation in Multi-View Stereo","authors":"C. Strecha, R. Fransens, L. Gool","doi":"10.1109/CVPR.2006.78","DOIUrl":"https://doi.org/10.1109/CVPR.2006.78","url":null,"abstract":"In this paper, we present a generative model based approach to solve the multi-view stereo problem. The input images are considered to be generated by either one of two processes: (i) an inlier process, which generates the pixels which are visible from the reference camera and which obey the constant brightness assumption, and (ii) an outlier process which generates all other pixels. Depth and visibility are jointly modelled as a hiddenMarkov Random Field, and the spatial correlations of both are explicitly accounted for. Inference is made tractable by an EM-algorithm, which alternates between estimation of visibility and depth, and optimisation of model parameters. We describe and compare two implementations of the E-step of the algorithm, which correspond to the Mean Field and Bethe approximations of the free energy. The approach is validated by experiments on challenging real-world scenes, of which two are contaminated by independently moving objects.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123886964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 209
Kernel Uncorrelated and Orthogonal Discriminant Analysis: A Unified Approach 核不相关和正交判别分析:一种统一的方法
T. Xiong, Jieping Ye, V. Cherkassky
Several kernel algorithms have recently been proposed for nonlinear discriminant analysis. However, these methods mainly address the singularity problem in the high dimensional feature space. Less attention has been focused on the properties of the resulting discriminant vectors and feature vectors in the reduced dimensional space. In this paper, we present a new formulation for kernel discriminant analysis. The proposed formulation includes, as special cases, kernel uncorrelated discriminant analysis (KUDA) and kernel orthogonal discriminant analysis (KODA). The feature vectors of KUDA are uncorrelated, while the discriminant vectors of KODA are orthogonal to each other in the feature space. We present theoretical derivations of proposed KUDA and KODA algorithms. The experimental results show that both KUDA and KODA are very competitive in comparison with other nonlinear discriminant algorithms in terms of classification accuracy.
最近提出了几种用于非线性判别分析的核算法。然而,这些方法主要解决高维特征空间中的奇异性问题。对于在降维空间中得到的判别向量和特征向量的性质关注较少。本文给出了核判别分析的一个新公式。作为特殊情况,提出的公式包括核不相关判别分析(KUDA)和核正交判别分析(KODA)。KODA的特征向量是不相关的,而KODA的判别向量在特征空间中是相互正交的。我们提出了KUDA和KODA算法的理论推导。实验结果表明,与其他非线性判别算法相比,KUDA和KODA在分类精度方面都具有很强的竞争力。
{"title":"Kernel Uncorrelated and Orthogonal Discriminant Analysis: A Unified Approach","authors":"T. Xiong, Jieping Ye, V. Cherkassky","doi":"10.1109/CVPR.2006.161","DOIUrl":"https://doi.org/10.1109/CVPR.2006.161","url":null,"abstract":"Several kernel algorithms have recently been proposed for nonlinear discriminant analysis. However, these methods mainly address the singularity problem in the high dimensional feature space. Less attention has been focused on the properties of the resulting discriminant vectors and feature vectors in the reduced dimensional space. In this paper, we present a new formulation for kernel discriminant analysis. The proposed formulation includes, as special cases, kernel uncorrelated discriminant analysis (KUDA) and kernel orthogonal discriminant analysis (KODA). The feature vectors of KUDA are uncorrelated, while the discriminant vectors of KODA are orthogonal to each other in the feature space. We present theoretical derivations of proposed KUDA and KODA algorithms. The experimental results show that both KUDA and KODA are very competitive in comparison with other nonlinear discriminant algorithms in terms of classification accuracy.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122418090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1