首页 > 最新文献

2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)最新文献

英文 中文
Cosegmentation of Image Pairs by Histogram Matching - Incorporating a Global Constraint into MRFs 基于直方图匹配的图像对共分割——将全局约束纳入mrf
C. Rother, T. Minka, A. Blake, V. Kolmogorov
We introduce the term cosegmentation which denotes the task of segmenting simultaneously the common parts of an image pair. A generative model for cosegmentation is presented. Inference in the model leads to minimizing an energy with an MRF term encoding spatial coherency and a global constraint which attempts to match the appearance histograms of the common parts. This energy has not been proposed previously and its optimization is challenging and NP-hard. For this problem a novel optimization scheme which we call trust region graph cuts is presented. We demonstrate that this framework has the potential to improve a wide range of research: Object driven image retrieval, video tracking and segmentation, and interactive image editing. The power of the framework lies in its generality, the common part can be a rigid/non-rigid object (or scene), observed from different viewpoints or even similar objects of the same class.
我们引入了术语共分割,它表示同时分割图像对的公共部分的任务。提出了一种生成式共分割模型。模型中的推理导致最小化能量与编码空间相干性的MRF项和试图匹配公共部分的外观直方图的全局约束。这种能量以前没有被提出过,它的优化具有挑战性和NP-hard。针对这一问题,提出了一种新的优化方案——信赖域图切。我们证明,这个框架有潜力改善广泛的研究:对象驱动的图像检索,视频跟踪和分割,以及交互式图像编辑。框架的强大之处在于它的通用性,公共部分可以是刚性/非刚性对象(或场景),从不同的角度观察,甚至是同一类的类似对象。
{"title":"Cosegmentation of Image Pairs by Histogram Matching - Incorporating a Global Constraint into MRFs","authors":"C. Rother, T. Minka, A. Blake, V. Kolmogorov","doi":"10.1109/CVPR.2006.91","DOIUrl":"https://doi.org/10.1109/CVPR.2006.91","url":null,"abstract":"We introduce the term cosegmentation which denotes the task of segmenting simultaneously the common parts of an image pair. A generative model for cosegmentation is presented. Inference in the model leads to minimizing an energy with an MRF term encoding spatial coherency and a global constraint which attempts to match the appearance histograms of the common parts. This energy has not been proposed previously and its optimization is challenging and NP-hard. For this problem a novel optimization scheme which we call trust region graph cuts is presented. We demonstrate that this framework has the potential to improve a wide range of research: Object driven image retrieval, video tracking and segmentation, and interactive image editing. The power of the framework lies in its generality, the common part can be a rigid/non-rigid object (or scene), observed from different viewpoints or even similar objects of the same class.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129177797","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 588
Regression-based Hand Pose Estimation from Multiple Cameras 基于回归的多相机手部姿态估计
T. D. Campos, D. W. Murray
The RVM-based learning method for whole body pose estimation proposed by Agarwal and Triggs is adapted to hand pose recovery. To help overcome the difficulties presented by the greater degree of self-occlusion and the wider range of poses exhibited in hand imagery, the adaptation proposes a method for combining multiple views. Comparisons of performance using single versus multiple views are reported for both synthesized and real imagery, and the effects of the number of image measurements and the number of training samples on performance are explored.
Agarwal和Triggs提出的基于rvm的全身姿态估计学习方法适用于手部姿态恢复。为了帮助克服手图像中更大程度的自遮挡和更大范围的姿势所带来的困难,该适应提出了一种结合多个视图的方法。报告了合成图像和真实图像使用单视图与多视图的性能比较,并探讨了图像测量数量和训练样本数量对性能的影响。
{"title":"Regression-based Hand Pose Estimation from Multiple Cameras","authors":"T. D. Campos, D. W. Murray","doi":"10.1109/CVPR.2006.252","DOIUrl":"https://doi.org/10.1109/CVPR.2006.252","url":null,"abstract":"The RVM-based learning method for whole body pose estimation proposed by Agarwal and Triggs is adapted to hand pose recovery. To help overcome the difficulties presented by the greater degree of self-occlusion and the wider range of poses exhibited in hand imagery, the adaptation proposes a method for combining multiple views. Comparisons of performance using single versus multiple views are reported for both synthesized and real imagery, and the effects of the number of image measurements and the number of training samples on performance are explored.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123884748","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 92
Multiview Geometry for Texture Mapping 2D Images Onto 3D Range Data 多视图几何纹理映射2D图像到3D范围数据
Lingyun Liu, G. Yu, G. Wolberg, Siavash Zokai
The photorealistic modeling of large-scale scenes, such as urban structures, requires a fusion of range sensing technology and traditional digital photography. This paper presents a system that integrates multiview geometry and automated 3D registration techniques for texture mapping 2D images onto 3D range data. The 3D range scans and the 2D photographs are respectively used to generate a pair of 3D models of the scene. The first model consists of a dense 3D point cloud, produced by using a 3D-to-3D registration method that matches 3D lines in the range images. The second model consists of a sparse 3D point cloud, produced by applying a multiview geometry (structure-from-motion) algorithm directly on a sequence of 2D photographs. This paper introduces a novel algorithm for automatically recovering the rotation, scale, and translation that best aligns the dense and sparse models. This alignment is necessary to enable the photographs to be optimally texture mapped onto the dense model. The contribution of this work is that it merges the benefits of multiview geometry with automated registration of 3D range scans to produce photorealistic models with minimal human interaction. We present results from experiments in large-scale urban scenes.
城市建筑等大型场景的逼真建模需要将距离传感技术与传统数码摄影技术相融合。本文提出了一个集成多视点几何和自动三维配准技术的系统,用于将二维图像纹理映射到三维距离数据上。分别使用三维距离扫描和二维照片生成一对场景的三维模型。第一个模型由密集的3D点云组成,使用3D到3D配准方法匹配范围图像中的3D线。第二个模型由一个稀疏的3D点云组成,该点云是通过直接对一系列2D照片应用多视图几何(运动结构)算法产生的。本文介绍了一种自动恢复旋转、缩放和平移的新算法,该算法可以使密集模型和稀疏模型最好地对齐。这种对齐是必要的,以使照片被最佳纹理映射到密集的模型。这项工作的贡献在于,它将多视图几何的好处与3D范围扫描的自动注册相结合,以最少的人为交互产生逼真的模型。我们展示了大规模城市场景的实验结果。
{"title":"Multiview Geometry for Texture Mapping 2D Images Onto 3D Range Data","authors":"Lingyun Liu, G. Yu, G. Wolberg, Siavash Zokai","doi":"10.1109/CVPR.2006.204","DOIUrl":"https://doi.org/10.1109/CVPR.2006.204","url":null,"abstract":"The photorealistic modeling of large-scale scenes, such as urban structures, requires a fusion of range sensing technology and traditional digital photography. This paper presents a system that integrates multiview geometry and automated 3D registration techniques for texture mapping 2D images onto 3D range data. The 3D range scans and the 2D photographs are respectively used to generate a pair of 3D models of the scene. The first model consists of a dense 3D point cloud, produced by using a 3D-to-3D registration method that matches 3D lines in the range images. The second model consists of a sparse 3D point cloud, produced by applying a multiview geometry (structure-from-motion) algorithm directly on a sequence of 2D photographs. This paper introduces a novel algorithm for automatically recovering the rotation, scale, and translation that best aligns the dense and sparse models. This alignment is necessary to enable the photographs to be optimally texture mapped onto the dense model. The contribution of this work is that it merges the benefits of multiview geometry with automated registration of 3D range scans to produce photorealistic models with minimal human interaction. We present results from experiments in large-scale urban scenes.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"68 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120873723","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 94
Unsupervised Bayesian Detection of Independent Motion in Crowds 群体中独立运动的无监督贝叶斯检测
G. Brostow, R. Cipolla
While crowds of various subjects may offer applicationspecific cues to detect individuals, we demonstrate that for the general case, motion itself contains more information than previously exploited. This paper describes an unsupervised data driven Bayesian clustering algorithm which has detection of individual entities as its primary goal. We track simple image features and probabilistically group them into clusters representing independently moving entities. The numbers of clusters and the grouping of constituent features are determined without supervised learning or any subject-specific model. The new approach is instead, that space-time proximity and trajectory coherence through image space are used as the only probabilistic criteria for clustering. An important contribution of this work is how these criteria are used to perform a one-shot data association without iterating through combinatorial hypotheses of cluster assignments. Our proposed general detection algorithm can be augmented with subject-specific filtering, but is shown to already be effective at detecting individual entities in crowds of people, insects, and animals. This paper and the associated video examine the implementation and experiments of our motion clustering framework.
虽然不同对象的群体可能会提供特定于应用程序的线索来检测个体,但我们证明,对于一般情况,运动本身包含的信息比以前所利用的更多。本文描述了一种以检测单个实体为主要目标的无监督数据驱动的贝叶斯聚类算法。我们跟踪简单的图像特征,并概率地将它们分组到代表独立移动实体的簇中。集群的数量和组成特征的分组是在没有监督学习或任何特定主题模型的情况下确定的。新的方法是将时空接近性和通过图像空间的轨迹相干性作为聚类的唯一概率标准。这项工作的一个重要贡献是如何使用这些标准来执行一次性数据关联,而不需要迭代聚类分配的组合假设。我们提出的通用检测算法可以通过特定主题的过滤来增强,但已经被证明在检测人群、昆虫和动物中的个体实体方面是有效的。本文和相关视频研究了我们的运动聚类框架的实现和实验。
{"title":"Unsupervised Bayesian Detection of Independent Motion in Crowds","authors":"G. Brostow, R. Cipolla","doi":"10.1109/CVPR.2006.320","DOIUrl":"https://doi.org/10.1109/CVPR.2006.320","url":null,"abstract":"While crowds of various subjects may offer applicationspecific cues to detect individuals, we demonstrate that for the general case, motion itself contains more information than previously exploited. This paper describes an unsupervised data driven Bayesian clustering algorithm which has detection of individual entities as its primary goal. We track simple image features and probabilistically group them into clusters representing independently moving entities. The numbers of clusters and the grouping of constituent features are determined without supervised learning or any subject-specific model. The new approach is instead, that space-time proximity and trajectory coherence through image space are used as the only probabilistic criteria for clustering. An important contribution of this work is how these criteria are used to perform a one-shot data association without iterating through combinatorial hypotheses of cluster assignments. Our proposed general detection algorithm can be augmented with subject-specific filtering, but is shown to already be effective at detecting individual entities in crowds of people, insects, and animals. This paper and the associated video examine the implementation and experiments of our motion clustering framework.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116506599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 461
Transformation invariant component analysis for binary images 二值图像的变换不变分量分析
Zoran Zivkovic, Jakob Verbeek
There are various situations where image data is binary: character recognition, result of image segmentation etc. As a first contribution, we compare Gaussian based principal component analysis (PCA), which is often used to model images, and "binary PCA" which models the binary data more naturally using Bernoulli distributions. Furthermore, we address the problem of data alignment. Image data is often perturbed by some global transformations such as shifting, rotation, scaling etc. In such cases the data needs to be transformed to some canonical aligned form. As a second contribution, we extend the binary PCA to the "transformation invariant mixture of binary PCAs" which simultaneously corrects the data for a set of global transformations and learns the binary PCA model on the aligned data.
图像数据是二值化的情况有很多:字符识别、图像分割结果等。作为第一个贡献,我们比较了基于高斯的主成分分析(PCA)和“二进制PCA”,前者通常用于图像建模,后者更自然地使用伯努利分布对二进制数据建模。此外,我们还解决了数据对齐的问题。图像数据经常受到一些全局变换的干扰,如移动、旋转、缩放等。在这种情况下,需要将数据转换为某种规范对齐的形式。作为第二个贡献,我们将二元主成分分析扩展到“二元主成分分析的变换不变混合”,它同时校正一组全局变换的数据,并在对齐的数据上学习二元主成分分析模型。
{"title":"Transformation invariant component analysis for binary images","authors":"Zoran Zivkovic, Jakob Verbeek","doi":"10.1109/CVPR.2006.316","DOIUrl":"https://doi.org/10.1109/CVPR.2006.316","url":null,"abstract":"There are various situations where image data is binary: character recognition, result of image segmentation etc. As a first contribution, we compare Gaussian based principal component analysis (PCA), which is often used to model images, and \"binary PCA\" which models the binary data more naturally using Bernoulli distributions. Furthermore, we address the problem of data alignment. Image data is often perturbed by some global transformations such as shifting, rotation, scaling etc. In such cases the data needs to be transformed to some canonical aligned form. As a second contribution, we extend the binary PCA to the \"transformation invariant mixture of binary PCAs\" which simultaneously corrects the data for a set of global transformations and learns the binary PCA model on the aligned data.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113967573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Sparse and Semi-supervised Visual Mapping with the S^3GP 基于S^3GP的稀疏半监督视觉映射
Oliver Williams, A. Blake, R. Cipolla
This paper is about mapping images to continuous output spaces using powerful Bayesian learning techniques. A sparse, semi-supervised Gaussian process regression model (S3GP) is introduced which learns a mapping using only partially labelled training data. We show that sparsity bestows efficiency on the S3GP which requires minimal CPU utilization for real-time operation; the predictions of uncertainty made by the S3GP are more accurate than those of other models leading to considerable performance improvements when combined with a probabilistic filter; and the ability to learn from semi-supervised data simplifies the process of collecting training data. The S3GP uses a mixture of different image features: this is also shown to improve the accuracy and consistency of the mapping. A major application of this work is its use as a gaze tracking system in which images of a human eye are mapped to screen coordinates: in this capacity our approach is efficient, accurate and versatile.
本文是关于使用强大的贝叶斯学习技术将图像映射到连续输出空间。介绍了一种稀疏的半监督高斯过程回归模型(S3GP),该模型仅使用部分标记的训练数据学习映射。我们表明,稀疏性为S3GP带来了效率,这需要最小的CPU利用率来进行实时操作;S3GP对不确定性的预测比其他模型更准确,当与概率过滤器结合使用时,可以显著提高性能;从半监督数据中学习的能力简化了收集训练数据的过程。S3GP使用不同图像特征的混合:这也表明可以提高映射的准确性和一致性。这项工作的一个主要应用是将其用作注视跟踪系统,在该系统中,人眼的图像被映射到屏幕坐标上:在这种情况下,我们的方法是高效、准确和通用的。
{"title":"Sparse and Semi-supervised Visual Mapping with the S^3GP","authors":"Oliver Williams, A. Blake, R. Cipolla","doi":"10.1109/CVPR.2006.285","DOIUrl":"https://doi.org/10.1109/CVPR.2006.285","url":null,"abstract":"This paper is about mapping images to continuous output spaces using powerful Bayesian learning techniques. A sparse, semi-supervised Gaussian process regression model (S3GP) is introduced which learns a mapping using only partially labelled training data. We show that sparsity bestows efficiency on the S3GP which requires minimal CPU utilization for real-time operation; the predictions of uncertainty made by the S3GP are more accurate than those of other models leading to considerable performance improvements when combined with a probabilistic filter; and the ability to learn from semi-supervised data simplifies the process of collecting training data. The S3GP uses a mixture of different image features: this is also shown to improve the accuracy and consistency of the mapping. A major application of this work is its use as a gaze tracking system in which images of a human eye are mapped to screen coordinates: in this capacity our approach is efficient, accurate and versatile.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114806731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 145
Multi-Resolution Patch Tensor for Facial Expression Hallucination 面部表情幻觉的多分辨率Patch张量
K. Jia, S. Gong
In this paper, we propose a sequential approach to hallucinate/ synthesize high-resolution images of multiple facial expressions. We propose an idea of multi-resolution tensor for super-resolution, and decompose facial expression images into small local patches. We build a multi-resolution patch tensor across different facial expressions. By unifying the identity parameters and learning the subspace mappings across different resolutions and expressions, we simplify the facial expression hallucination as a problem of parameter recovery in a patch tensor space. We further add a high-frequency component residue using nonparametric patch learning from high-resolution training data. We integrate the sequential statistical modelling into a Bayesian framework, so that given any low-resolution facial image of a single expression, we are able to synthesize multiple facial expression images in high-resolution. We show promising experimental results from both facial expression database and live video sequences.
在本文中,我们提出了一种序列方法来幻觉/合成多个面部表情的高分辨率图像。我们提出了多分辨率张量的超分辨率思想,并将面部表情图像分解成小的局部小块。我们在不同的面部表情之间建立了一个多分辨率的patch张量。通过统一身份参数和学习不同分辨率和表达式的子空间映射,将面部表情幻觉简化为一个斑块张量空间中的参数恢复问题。我们进一步使用非参数patch学习从高分辨率训练数据中添加高频成分残差。我们将序列统计建模整合到贝叶斯框架中,因此,给定任何低分辨率的单一表情面部图像,我们都能够合成高分辨率的多个面部表情图像。我们从面部表情数据库和实时视频序列中展示了有希望的实验结果。
{"title":"Multi-Resolution Patch Tensor for Facial Expression Hallucination","authors":"K. Jia, S. Gong","doi":"10.1109/CVPR.2006.196","DOIUrl":"https://doi.org/10.1109/CVPR.2006.196","url":null,"abstract":"In this paper, we propose a sequential approach to hallucinate/ synthesize high-resolution images of multiple facial expressions. We propose an idea of multi-resolution tensor for super-resolution, and decompose facial expression images into small local patches. We build a multi-resolution patch tensor across different facial expressions. By unifying the identity parameters and learning the subspace mappings across different resolutions and expressions, we simplify the facial expression hallucination as a problem of parameter recovery in a patch tensor space. We further add a high-frequency component residue using nonparametric patch learning from high-resolution training data. We integrate the sequential statistical modelling into a Bayesian framework, so that given any low-resolution facial image of a single expression, we are able to synthesize multiple facial expression images in high-resolution. We show promising experimental results from both facial expression database and live video sequences.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127752067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A New Formulation for Shape from Shading for Non-Lambertian Surfaces 非朗伯曲面阴影形状的新公式
Abdelrehim H. Ahmed, A. Farag
Lambert’s model for diffuse reflection is a main assumption in most of shape from shading (SFS) literature. Even with this simplified model, the SFS is still a difficult problem. Nevertheless, Lambert’s model has been proven to be an inaccurate approximation of the diffuse component of the surface reflectance. In this paper, we propose a new solution of the SFS problem based on a more comprehensive diffuse reflectance model: the Oren and Nayar model. In this work, we slightly modify this more realistic model in order to take into account the attenuation of the illumination due to distance. Using the modified non-Lambertian reflectance, we design a new explicit Partial Differential Equation (PDE) and then solve it using Lax-Friedrichs Sweeping method. Our experiments on synthetic data show that the proposed modeling gives a unique solution without any information about the height at the singular points of the surface. Additional results for real data are presented to show the efficiency of the proposed method . To the best of our knowledge, this is the first non-Lambertian SFS formulation that eliminates the concave/convex ambiguity which is a well known problem in SFS.
漫反射的兰伯特模型是遮阳(SFS)文献中大多数形状的主要假设。即使有了这个简化的模型,SFS仍然是一个难题。然而,兰伯特的模型已被证明是表面反射率漫反射分量的不准确近似值。在本文中,我们提出了一种基于更全面漫反射模型的SFS问题的新解:Oren和Nayar模型。在这项工作中,我们稍微修改了这个更现实的模型,以便考虑到由于距离引起的照明衰减。利用改进的非朗伯反射率,设计了一个新的显式偏微分方程(PDE),然后用Lax-Friedrichs扫描法求解。我们在合成数据上的实验表明,所提出的模型给出了一个唯一的解,不需要任何关于表面奇异点高度的信息。实际数据的结果表明了该方法的有效性。据我们所知,这是第一个消除了凹凸模糊的非朗伯SFS公式,这是SFS中一个众所周知的问题。
{"title":"A New Formulation for Shape from Shading for Non-Lambertian Surfaces","authors":"Abdelrehim H. Ahmed, A. Farag","doi":"10.1109/CVPR.2006.35","DOIUrl":"https://doi.org/10.1109/CVPR.2006.35","url":null,"abstract":"Lambert’s model for diffuse reflection is a main assumption in most of shape from shading (SFS) literature. Even with this simplified model, the SFS is still a difficult problem. Nevertheless, Lambert’s model has been proven to be an inaccurate approximation of the diffuse component of the surface reflectance. In this paper, we propose a new solution of the SFS problem based on a more comprehensive diffuse reflectance model: the Oren and Nayar model. In this work, we slightly modify this more realistic model in order to take into account the attenuation of the illumination due to distance. Using the modified non-Lambertian reflectance, we design a new explicit Partial Differential Equation (PDE) and then solve it using Lax-Friedrichs Sweeping method. Our experiments on synthetic data show that the proposed modeling gives a unique solution without any information about the height at the singular points of the surface. Additional results for real data are presented to show the efficiency of the proposed method . To the best of our knowledge, this is the first non-Lambertian SFS formulation that eliminates the concave/convex ambiguity which is a well known problem in SFS.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128146936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Identifying Color in Motion in Video Sensors 在视频传感器中识别运动中的颜色
Gang Wu, Amir M. Rahimi, E. Chang, Kingshy Goh, Tomy Tsai, Ankur Jain, Yuan-fang Wang
Identifying or matching the surface color of a moving object in surveillance video is critical for achieving reliable object-tracking and searching. Traditional color models provide little help, since the surface of an object is usually not flat, the object’s motion can alter the surface’s orientation, and the lighting conditions can vary when the object moves. To tackle this research problem, we conduct extensive data mining on video clips collected under various lighting conditions and distances from several video-cameras. We observe how each of the eleven culture colors can drift in the color space when an object’s surface is in motion. In the color space, we then learn the drift pattern of each culture color for classifying unseen surface colors. Finally, we devise a distance function taking color drift into consideration to perform color identification and matching. Empirical studies show our approach to be very promising: achieving over 95% color-prediction accuracy.
在监控视频中,识别或匹配运动物体的表面颜色是实现可靠的目标跟踪和搜索的关键。传统的色彩模型提供的帮助很少,因为物体的表面通常不是平坦的,物体的运动可以改变表面的方向,并且当物体移动时,照明条件也会发生变化。为了解决这个研究问题,我们对在不同照明条件和距离下从几个摄像机收集的视频片段进行了广泛的数据挖掘。我们观察到,当物体的表面处于运动状态时,这11种文化色彩是如何在色彩空间中漂移的。然后在色彩空间中,我们学习每种文化颜色的漂移模式,用于分类未见过的表面颜色。最后,我们设计了一个考虑颜色漂移的距离函数来进行颜色识别和匹配。实证研究表明,我们的方法是非常有前途的:达到95%以上的颜色预测精度。
{"title":"Identifying Color in Motion in Video Sensors","authors":"Gang Wu, Amir M. Rahimi, E. Chang, Kingshy Goh, Tomy Tsai, Ankur Jain, Yuan-fang Wang","doi":"10.1109/CVPR.2006.139","DOIUrl":"https://doi.org/10.1109/CVPR.2006.139","url":null,"abstract":"Identifying or matching the surface color of a moving object in surveillance video is critical for achieving reliable object-tracking and searching. Traditional color models provide little help, since the surface of an object is usually not flat, the object’s motion can alter the surface’s orientation, and the lighting conditions can vary when the object moves. To tackle this research problem, we conduct extensive data mining on video clips collected under various lighting conditions and distances from several video-cameras. We observe how each of the eleven culture colors can drift in the color space when an object’s surface is in motion. In the color space, we then learn the drift pattern of each culture color for classifying unseen surface colors. Finally, we devise a distance function taking color drift into consideration to perform color identification and matching. Empirical studies show our approach to be very promising: achieving over 95% color-prediction accuracy.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125631551","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Three-Dimensional Volume Reconstruction Based on Trajectory Fusion from Confocal Laser Scanning Microscope Images 基于共聚焦激光扫描显微镜图像轨迹融合的三维体重建
Sang-chul Lee, P. Bajcsy
In this paper, we address the problem of 3D volume reconstruction from depth adjacent subvolumes (i.e., sets of image frames) acquired using a confocal laser scanning microscope (CLSM). Our goal is to align sub-volumes by estimating an optimal global image transformation which preserves morphological smoothness of medical structures (called features, e.g., blood vessels) inside of a reconstructed 3D volume. We approached the problem by learning morphological characteristics of structures inside of each sub-volume, i.e. centroid trajectories of features. Next, adjacent sub-volumes are aligned by fusing the morphological characteristics of structures using extrapolation or model fitting. Finally, a global sub-volume to subvolume transformation is computed based on the entire set of fused structures. The trajectory-based 3D volume reconstruction method described here is evaluated with a pair of consecutive physical sections using two evaluation metrics for morphological continu
在本文中,我们解决了使用共聚焦激光扫描显微镜(CLSM)从深度相邻子体(即图像帧集)获得的三维体重建问题。我们的目标是通过估计最优的全局图像变换来对齐子体积,该变换保留重建3D体积内医疗结构(称为特征,例如血管)的形态学平滑性。我们通过学习每个子体内部结构的形态特征,即特征的质心轨迹来解决这个问题。接下来,通过使用外推或模型拟合融合结构的形态特征来对齐相邻的子体。最后,基于整个融合结构集计算全局子体到子体的转换。本文描述的基于轨迹的三维体重建方法使用形态学连续性的两个评估指标对一对连续的物理切片进行评估
{"title":"Three-Dimensional Volume Reconstruction Based on Trajectory Fusion from Confocal Laser Scanning Microscope Images","authors":"Sang-chul Lee, P. Bajcsy","doi":"10.1109/CVPR.2006.308","DOIUrl":"https://doi.org/10.1109/CVPR.2006.308","url":null,"abstract":"In this paper, we address the problem of 3D volume reconstruction from depth adjacent subvolumes (i.e., sets of image frames) acquired using a confocal laser scanning microscope (CLSM). Our goal is to align sub-volumes by estimating an optimal global image transformation which preserves morphological smoothness of medical structures (called features, e.g., blood vessels) inside of a reconstructed 3D volume. We approached the problem by learning morphological characteristics of structures inside of each sub-volume, i.e. centroid trajectories of features. Next, adjacent sub-volumes are aligned by fusing the morphological characteristics of structures using extrapolation or model fitting. Finally, a global sub-volume to subvolume transformation is computed based on the entire set of fused structures. The trajectory-based 3D volume reconstruction method described here is evaluated with a pair of consecutive physical sections using two evaluation metrics for morphological continu","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122251067","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1