首页 > 最新文献

Proceedings Ninth IEEE International Conference on Computer Vision最新文献

英文 中文
Image statistics and anisotropic diffusion 图像统计和各向异性扩散
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238435
H. Scharr, Michael J. Black, H. Haussecker
Many sensing techniques and image processing applications are characterized by noisy, or corrupted, image data. Anisotropic diffusion is a popular, and theoretically well understood, technique for denoising such images. Diffusion approaches however require the selection of an "edge stopping" function, the definition of which is typically ad hoc. We exploit and extend recent work on the statistics of natural images to define principled edge stopping functions for different types of imagery. We consider a variety of anisotropic diffusion schemes and note that they compute spatial derivatives at fixed scales from which we estimate the appropriate algorithm-specific image statistics. Going beyond traditional work on image statistics, we also model the statistics of the eigenvalues of the local structure tensor. Novel edge-stopping functions are derived from these image statistics giving a principled way of formulating anisotropic diffusion problems in which all edge-stopping parameters are learned from training data.
许多传感技术和图像处理应用的特点是有噪声或损坏的图像数据。各向异性扩散是一种流行的,理论上很好理解的图像去噪技术。然而,扩散方法需要选择一个“边缘停止”函数,其定义通常是特别的。我们利用并扩展了最近在自然图像统计方面的工作,为不同类型的图像定义原则性的边缘停止函数。我们考虑了各种各向异性扩散方案,并注意到它们在固定尺度上计算空间导数,我们从中估计适当的算法特定的图像统计。超越传统的图像统计工作,我们还建立了局部结构张量特征值的统计模型。从这些图像统计中导出了新的止边函数,给出了一种表达各向异性扩散问题的原则方法,其中所有止边参数都是从训练数据中学习的。
{"title":"Image statistics and anisotropic diffusion","authors":"H. Scharr, Michael J. Black, H. Haussecker","doi":"10.1109/ICCV.2003.1238435","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238435","url":null,"abstract":"Many sensing techniques and image processing applications are characterized by noisy, or corrupted, image data. Anisotropic diffusion is a popular, and theoretically well understood, technique for denoising such images. Diffusion approaches however require the selection of an \"edge stopping\" function, the definition of which is typically ad hoc. We exploit and extend recent work on the statistics of natural images to define principled edge stopping functions for different types of imagery. We consider a variety of anisotropic diffusion schemes and note that they compute spatial derivatives at fixed scales from which we estimate the appropriate algorithm-specific image statistics. Going beyond traditional work on image statistics, we also model the statistics of the eigenvalues of the local structure tensor. Novel edge-stopping functions are derived from these image statistics giving a principled way of formulating anisotropic diffusion problems in which all edge-stopping parameters are learned from training data.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"405 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123523002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 71
Discriminative random fields: a discriminative framework for contextual interaction in classification 判别随机场:分类中上下文交互的判别框架
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238478
Sanjiv Kumar, M. Hebert
In this work we present discriminative random fields (DRFs), a discriminative framework for the classification of image regions by incorporating neighborhood interactions in the labels as well as the observed data. The discriminative random fields offer several advantages over the conventional Markov random field (MRF) framework. First, the DRFs allow to relax the strong assumption of conditional independence of the observed data generally used in the MRF framework for tractability. This assumption is too restrictive for a large number of applications in vision. Second, the DRFs derive their classification power by exploiting the probabilistic discriminative models instead of the generative models used in the MRF framework. Finally, all the parameters in the DRF model are estimated simultaneously from the training data unlike the MRF framework where likelihood parameters are usually learned separately from the field parameters. We illustrate the advantages of the DRFs over the MRF framework in an application of man-made structure detection in natural images taken from the Corel database.
在这项工作中,我们提出了判别随机场(DRFs),这是一种判别框架,通过在标签和观测数据中结合邻域相互作用来对图像区域进行分类。与传统的马尔可夫随机场(MRF)框架相比,判别式随机场具有许多优点。首先,drf允许放松通常在MRF框架中用于可追溯性的观测数据的条件独立性的强假设。这个假设对于视觉领域的大量应用来说过于严格。其次,drf通过利用概率判别模型而不是MRF框架中使用的生成模型来获得分类能力。最后,DRF模型中的所有参数都是同时从训练数据中估计出来的,而MRF框架中的似然参数通常是与现场参数分开学习的。我们在Corel数据库中提取的自然图像的人工结构检测应用中说明了drf比MRF框架的优势。
{"title":"Discriminative random fields: a discriminative framework for contextual interaction in classification","authors":"Sanjiv Kumar, M. Hebert","doi":"10.1109/ICCV.2003.1238478","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238478","url":null,"abstract":"In this work we present discriminative random fields (DRFs), a discriminative framework for the classification of image regions by incorporating neighborhood interactions in the labels as well as the observed data. The discriminative random fields offer several advantages over the conventional Markov random field (MRF) framework. First, the DRFs allow to relax the strong assumption of conditional independence of the observed data generally used in the MRF framework for tractability. This assumption is too restrictive for a large number of applications in vision. Second, the DRFs derive their classification power by exploiting the probabilistic discriminative models instead of the generative models used in the MRF framework. Finally, all the parameters in the DRF model are estimated simultaneously from the training data unlike the MRF framework where likelihood parameters are usually learned separately from the field parameters. We illustrate the advantages of the DRFs over the MRF framework in an application of man-made structure detection in natural images taken from the Corel database.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125307833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 535
Dense matching of multiple wide-baseline views 多个宽基线视图的密集匹配
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238627
C. Strecha, T. Tuytelaars, L. Gool
This paper describes a PDE-based method for dense depth extraction from multiple wide-baseline images. Emphasis lies on the usage of only a small amount of images. The integration of these multiple wide-baseline views is guided by the relative confidence that the system has in the matching to different views. This weighting is fine-grained in that it is determined for every pixel at every iteration. Reliable information spreads fast at the expense of less reliable data, both in terms of spatial communications within a view and in terms of information exchange between the views. Changes in intensity between images can be handled in a similar fine grained fashion.
本文描述了一种基于偏微分方程的多幅宽基线图像密集深度提取方法。重点在于只使用少量的图像。这些多个宽基线视图的集成是由系统在匹配不同视图方面的相对置信度指导的。这种加权是细粒度的,因为它是在每次迭代中为每个像素确定的。在视图内的空间通信和视图之间的信息交换方面,可靠的信息以不可靠的数据为代价迅速传播。图像之间的强度变化可以用类似的细粒度方式处理。
{"title":"Dense matching of multiple wide-baseline views","authors":"C. Strecha, T. Tuytelaars, L. Gool","doi":"10.1109/ICCV.2003.1238627","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238627","url":null,"abstract":"This paper describes a PDE-based method for dense depth extraction from multiple wide-baseline images. Emphasis lies on the usage of only a small amount of images. The integration of these multiple wide-baseline views is guided by the relative confidence that the system has in the matching to different views. This weighting is fine-grained in that it is determined for every pixel at every iteration. Reliable information spreads fast at the expense of less reliable data, both in terms of spatial communications within a view and in terms of information exchange between the views. Changes in intensity between images can be handled in a similar fine grained fashion.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128247170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 197
A class of photometric invariants: separating material from shape and illumination 一类光度不变量:从形状和光照中分离物质
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238652
S. Narasimhan, Visvanathan Ramesh, S. Nayar
We derive a new class of photometric invariants that can be used for a variety of vision tasks including lighting invariant material segmentation, change detection and tracking, as well as material invariant shape recognition. The key idea is the formulation of a scene radiance model for the class of "separable" BRDFs, that can be decomposed into material related terms and object shape and lighting related terms. All the proposed invariants are simple rational functions of the appearance parameters (say, material or shape and lighting). The invariants in this class differ from one another in the number and type of image measurements they require. Most of the invariants in this class need changes in illumination or object position between image acquisitions. The invariants can handle large changes in lighting which pose problems for most existing vision algorithms. We demonstrate the power of these invariants using scenes with complex shapes, materials, textures, shadows and specularities.
我们推导了一类新的光度不变量,可用于各种视觉任务,包括照明不变量材料分割,变化检测和跟踪,以及材料不变量形状识别。关键思想是为“可分离”brdf类建立场景亮度模型,该模型可以分解为与材料相关的术语和与物体形状和照明相关的术语。所有提出的不变量都是外观参数(例如,材料或形状和照明)的简单有理函数。这类中的不变量在它们需要的图像测量的数量和类型上彼此不同。这类中的大多数不变量需要在图像获取之间改变光照或物体位置。不变量可以处理光照的大变化,这给大多数现有的视觉算法带来了问题。我们使用具有复杂形状,材料,纹理,阴影和镜面的场景来展示这些不变量的力量。
{"title":"A class of photometric invariants: separating material from shape and illumination","authors":"S. Narasimhan, Visvanathan Ramesh, S. Nayar","doi":"10.1109/ICCV.2003.1238652","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238652","url":null,"abstract":"We derive a new class of photometric invariants that can be used for a variety of vision tasks including lighting invariant material segmentation, change detection and tracking, as well as material invariant shape recognition. The key idea is the formulation of a scene radiance model for the class of \"separable\" BRDFs, that can be decomposed into material related terms and object shape and lighting related terms. All the proposed invariants are simple rational functions of the appearance parameters (say, material or shape and lighting). The invariants in this class differ from one another in the number and type of image measurements they require. Most of the invariants in this class need changes in illumination or object position between image acquisitions. The invariants can handle large changes in lighting which pose problems for most existing vision algorithms. We demonstrate the power of these invariants using scenes with complex shapes, materials, textures, shadows and specularities.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122309552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
Highlight removal by illumination-constrained inpainting 通过光照约束的绘画去除高光
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238333
P. Tan, Stephen Lin, Long Quan, H. Shum
We present a single-image highlight removal method that incorporates illumination-based constraints into image inpainting. Unlike occluded image regions filled by traditional inpainting, highlight pixels contain some useful information for guiding the inpainting process. Constraints provided by observed pixel colors, highlight color analysis and illumination color uniformity are employed in our method to improve estimation of the underlying diffuse color. The inclusion of these illumination constraints allows for better recovery of shading and textures by inpainting. Experimental results are given to demonstrate the performance of our method.
我们提出了一种单幅图像高光去除方法,该方法将基于光照的约束结合到图像绘制中。与传统补漆填充的闭塞图像区域不同,高亮像素包含一些有用的信息,可以指导补漆过程。我们的方法利用了观测像素颜色、高光颜色分析和照明颜色均匀性提供的约束来改进底层漫反射颜色的估计。这些光照约束的包含允许更好地恢复阴影和纹理通过油漆。实验结果验证了该方法的有效性。
{"title":"Highlight removal by illumination-constrained inpainting","authors":"P. Tan, Stephen Lin, Long Quan, H. Shum","doi":"10.1109/ICCV.2003.1238333","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238333","url":null,"abstract":"We present a single-image highlight removal method that incorporates illumination-based constraints into image inpainting. Unlike occluded image regions filled by traditional inpainting, highlight pixels contain some useful information for guiding the inpainting process. Constraints provided by observed pixel colors, highlight color analysis and illumination color uniformity are employed in our method to improve estimation of the underlying diffuse color. The inclusion of these illumination constraints allows for better recovery of shading and textures by inpainting. Experimental results are given to demonstrate the performance of our method.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115032696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 121
Video Google: a text retrieval approach to object matching in videos Video谷歌:视频中对象匹配的文本检索方法
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238663
Josef Sivic, Andrew Zisserman
We describe an approach to object and scene retrieval which searches for and localizes all the occurrences of a user outlined object in a video. The object is represented by a set of viewpoint invariant region descriptors so that recognition can proceed successfully despite changes in viewpoint, illumination and partial occlusion. The temporal continuity of the video within a shot is used to track the regions in order to reject unstable regions and reduce the effects of noise in the descriptors. The analogy with text retrieval is in the implementation where matches on descriptors are pre-computed (using vector quantization), and inverted file systems and document rankings are used. The result is that retrieved is immediate, returning a ranked list of key frames/shots in the manner of Google. The method is illustrated for matching in two full length feature films.
我们描述了一种对象和场景检索方法,该方法搜索并定位视频中用户概述对象的所有出现情况。目标由一组视点不变区域描述符表示,以便在视点、光照和部分遮挡变化的情况下仍能成功识别。在一个镜头内视频的时间连续性被用来跟踪区域,以拒绝不稳定的区域和减少噪声在描述符的影响。与文本检索类似的是在实现中预先计算描述符上的匹配(使用矢量量化),并使用反向文件系统和文档排名。结果是检索是即时的,以谷歌的方式返回关键帧/镜头的排名列表。以两长片的匹配为例说明了该方法的有效性。
{"title":"Video Google: a text retrieval approach to object matching in videos","authors":"Josef Sivic, Andrew Zisserman","doi":"10.1109/ICCV.2003.1238663","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238663","url":null,"abstract":"We describe an approach to object and scene retrieval which searches for and localizes all the occurrences of a user outlined object in a video. The object is represented by a set of viewpoint invariant region descriptors so that recognition can proceed successfully despite changes in viewpoint, illumination and partial occlusion. The temporal continuity of the video within a shot is used to track the regions in order to reject unstable regions and reduce the effects of noise in the descriptors. The analogy with text retrieval is in the implementation where matches on descriptors are pre-computed (using vector quantization), and inverted file systems and document rankings are used. The result is that retrieved is immediate, returning a ranked list of key frames/shots in the manner of Google. The method is illustrated for matching in two full length feature films.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129514931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7002
Globally convergent autocalibration 全局收敛的自动校准
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238657
A. Benedetti, Alessandro Busti, M. Farenzena, Andrea Fusiello
Existing autocalibration techniques use numerical optimization algorithms that are prone to the problem of local minima. To address this problem, we have developed a method where an interval branch-and-bound method is employed for numerical minimization. Thanks to the properties of interval analysis this method is guaranteed to converge to the global solution with mathematical certainty and arbitrary accuracy, and the only input information it requires from the user is a set of point correspondences and a search box. The cost function is based on the Huang-Faugeras constraint of the fundamental matrix. A recently proposed interval extension based on Bernstein polynomial forms has been investigated to speed up the search for the solution. Finally, some experimental results on synthetic images are presented.
现有的自动校准技术使用数值优化算法,容易出现局部极小值问题。为了解决这个问题,我们开发了一种采用区间分支定界法进行数值最小化的方法。由于区间分析的性质,保证了该方法收敛到具有数学确定性和任意精度的全局解,并且只需要用户输入一组点对应和一个搜索框。代价函数基于基本矩阵的Huang-Faugeras约束。研究了最近提出的一种基于Bernstein多项式形式的区间扩展,以加快求解速度。最后给出了在合成图像上的一些实验结果。
{"title":"Globally convergent autocalibration","authors":"A. Benedetti, Alessandro Busti, M. Farenzena, Andrea Fusiello","doi":"10.1109/ICCV.2003.1238657","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238657","url":null,"abstract":"Existing autocalibration techniques use numerical optimization algorithms that are prone to the problem of local minima. To address this problem, we have developed a method where an interval branch-and-bound method is employed for numerical minimization. Thanks to the properties of interval analysis this method is guaranteed to converge to the global solution with mathematical certainty and arbitrary accuracy, and the only input information it requires from the user is a set of point correspondences and a search box. The cost function is based on the Huang-Faugeras constraint of the fundamental matrix. A recently proposed interval extension based on Bernstein polynomial forms has been investigated to speed up the search for the solution. Finally, some experimental results on synthetic images are presented.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134513615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
SVM-based nonparametric discriminant analysis, an application to face detection 基于svm的非参数判别分析,在人脸检测中的应用
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238639
R. Fransens, J. D. Prins, L. Gool
Detecting the dominant normal directions to the decision surface is an established technique for feature selection in high dimensional classification problems. Several approaches have been proposed to render this strategy more amenable to practice, but they still show a number of important shortcomings from a pragmatic point of view. This paper introduces a novel such approach, which combines the normal directions idea with support vector machine classifiers. The two make a natural and powerful match, as SVs are located nearby, and fully describe the decision surfaces. The approach can be included elegantly into the training of performant classifiers from extensive datasets. The potential is corroborated by experiments, both on synthetic and real data, the latter on a face detection experiment. In this experiment we demonstrate how our approach can lead to a significant reduction of CPU-time, with neglectable loss of classification performance.
检测决策面的优势法线方向是一种成熟的高维分类问题特征选择技术。为了使这一战略更易于实施,已经提出了几种方法,但从实用主义的观点来看,它们仍然显示出一些重要的缺点。本文提出了一种将法向思想与支持向量机分类器相结合的方法。这两者形成了一种自然而有力的匹配,因为SVs位于附近,并且充分描述了决策面。该方法可以从广泛的数据集优雅地纳入高性能分类器的训练中。这种潜力得到了实验的证实,包括合成数据和真实数据,后者是人脸检测实验。在这个实验中,我们演示了我们的方法如何显著减少cpu时间,而分类性能的损失可以忽略不计。
{"title":"SVM-based nonparametric discriminant analysis, an application to face detection","authors":"R. Fransens, J. D. Prins, L. Gool","doi":"10.1109/ICCV.2003.1238639","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238639","url":null,"abstract":"Detecting the dominant normal directions to the decision surface is an established technique for feature selection in high dimensional classification problems. Several approaches have been proposed to render this strategy more amenable to practice, but they still show a number of important shortcomings from a pragmatic point of view. This paper introduces a novel such approach, which combines the normal directions idea with support vector machine classifiers. The two make a natural and powerful match, as SVs are located nearby, and fully describe the decision surfaces. The approach can be included elegantly into the training of performant classifiers from extensive datasets. The potential is corroborated by experiments, both on synthetic and real data, the latter on a face detection experiment. In this experiment we demonstrate how our approach can lead to a significant reduction of CPU-time, with neglectable loss of classification performance.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130853459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
Tales of shape and radiance in multiview stereo 在多视角立体的形状和光芒的故事
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238454
Stefano Soatto, A. Yezzi, Hailin Jin
To what extent can three-dimensional shape and radiance be inferred from a collection of images? Can the two be estimated separately while retaining optimality? How should the optimality criterion be computed? When is it necessary to employ an explicit model of the reflectance properties of a scene? In this paper we introduce a separation principle for shape and radiance estimation that applies to Lambertian scenes and holds for any choice of norm. When the scene is not Lambertian, however, shape cannot be decoupled from radiance, and therefore matching image-to-image is not possible directly. We employ a rank constraint on the radiance tensor, which is commonly used in computer graphics, and construct a novel cost functional whose minimization leads to an estimate of both shape and radiance for nonLambertian objects, which we validate experimentally.
在多大程度上可以从一组图像中推断出三维形状和亮度?能否在保持最优性的情况下分别对两者进行估计?如何计算最优性准则?什么时候需要使用场景反射属性的显式模型?在本文中,我们介绍了一种分离原理,用于形状和亮度估计,适用于兰伯特场景,并适用于任何范数的选择。然而,当场景不是朗伯时,形状不能与亮度解耦,因此不可能直接匹配图像到图像。我们在亮度张量上采用秩约束,这是计算机图形学中常用的,并构建了一个新的成本函数,其最小化导致非兰伯物体的形状和亮度的估计,我们实验验证了这一点。
{"title":"Tales of shape and radiance in multiview stereo","authors":"Stefano Soatto, A. Yezzi, Hailin Jin","doi":"10.1109/ICCV.2003.1238454","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238454","url":null,"abstract":"To what extent can three-dimensional shape and radiance be inferred from a collection of images? Can the two be estimated separately while retaining optimality? How should the optimality criterion be computed? When is it necessary to employ an explicit model of the reflectance properties of a scene? In this paper we introduce a separation principle for shape and radiance estimation that applies to Lambertian scenes and holds for any choice of norm. When the scene is not Lambertian, however, shape cannot be decoupled from radiance, and therefore matching image-to-image is not possible directly. We employ a rank constraint on the radiance tensor, which is commonly used in computer graphics, and construct a novel cost functional whose minimization leads to an estimate of both shape and radiance for nonLambertian objects, which we validate experimentally.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134124041","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
Fast intensity-based 2D-3D image registration of clinical data using light 基于光的临床数据快速2D-3D图像配准
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238376
Daniel B. Russakoff, T. Rohlfing, C. Maurer
Registration of a preoperative CT (3D) image to one or more X-ray projection (2D) images, a special case of the pose estimation problem, has been attempted in a variety of ways with varying degrees of success. Recently, there has been a great deal of interest in intensity-based methods. One of the drawbacks to such methods is the need to create digitally reconstructed radiographs (DRRs) at each step of the optimization process. DRRs are typically generated by ray casting, an operation that requires O(n/sup 3/) time, where we assume that n is approximately the size (in voxels) of one side of the DRR as well as one side of the CT volume. We address this issue by extending light field rendering techniques from the computer graphics community to generate DRRs instead of conventional rendered images. Using light fields allows most of the computation to be performed in a preprocessing step; after this precomputation, very accurate DRRs can be generated in O(n/sup 2/) time. Another important issue for 2D-3D registration algorithms is validation. Previously reported 2D-3D registration algorithms were validated using synthetic data or phantoms but not clinical data. We present an intensity-based 2D-3D registration system that generates DRRs using light fields; we validate its performance using clinical data with a known gold standard transformation.
将术前CT (3D)图像配准到一个或多个x射线投影(2D)图像,这是姿态估计问题的一个特殊情况,已经尝试了各种方法,并取得了不同程度的成功。最近,人们对基于强度的方法产生了极大的兴趣。这种方法的缺点之一是在优化过程的每一步都需要创建数字重建射线照片(DRRs)。DRR通常由光线投射生成,该操作需要O(n/sup 3/)时间,其中我们假设n大约是DRR一侧的大小(以体素为单位)以及CT体积的一侧。我们通过扩展计算机图形界的光场渲染技术来生成drr而不是传统的渲染图像来解决这个问题。使用光场允许在预处理步骤中执行大部分计算;经过这种预计算,可以在0 (n/sup 2/)时间内生成非常精确的drr。2D-3D配准算法的另一个重要问题是验证。先前报道的2D-3D配准算法使用合成数据或模型进行验证,而不是临床数据。我们提出了一种基于强度的2D-3D配准系统,该系统使用光场生成drr;我们使用已知的金标准转换的临床数据验证其性能。
{"title":"Fast intensity-based 2D-3D image registration of clinical data using light","authors":"Daniel B. Russakoff, T. Rohlfing, C. Maurer","doi":"10.1109/ICCV.2003.1238376","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238376","url":null,"abstract":"Registration of a preoperative CT (3D) image to one or more X-ray projection (2D) images, a special case of the pose estimation problem, has been attempted in a variety of ways with varying degrees of success. Recently, there has been a great deal of interest in intensity-based methods. One of the drawbacks to such methods is the need to create digitally reconstructed radiographs (DRRs) at each step of the optimization process. DRRs are typically generated by ray casting, an operation that requires O(n/sup 3/) time, where we assume that n is approximately the size (in voxels) of one side of the DRR as well as one side of the CT volume. We address this issue by extending light field rendering techniques from the computer graphics community to generate DRRs instead of conventional rendered images. Using light fields allows most of the computation to be performed in a preprocessing step; after this precomputation, very accurate DRRs can be generated in O(n/sup 2/) time. Another important issue for 2D-3D registration algorithms is validation. Previously reported 2D-3D registration algorithms were validated using synthetic data or phantoms but not clinical data. We present an intensity-based 2D-3D registration system that generates DRRs using light fields; we validate its performance using clinical data with a known gold standard transformation.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133470708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
期刊
Proceedings Ninth IEEE International Conference on Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1