首页 > 最新文献

2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)最新文献

英文 中文
A Generative-Discriminative Hybrid Method for Multi-View Object Detection 一种多视图目标检测的生成-判别混合方法
Dongqing Zhang, Shih-Fu Chang
We present a novel discriminative-generative hybrid approach in this paper, with emphasis on application in multiview object detection. Our method includes a novel generative model called Random Attributed Relational Graph (RARG) which is able to capture the structural and appearance characteristics of parts extracted from objects. We develop new variational learning methods to compute the approximation of the detection likelihood ratio function. The variaitonal likelihood ratio function can be shown to be a linear combination of the individual generative classifiers defined at nodes and edges of the RARG. Such insight inspires us to replace the generative classifiers at nodes and edges with discriminative classifiers, such as support vector machines, to further improve the detection performance. Our experiments have shown the robustness of the hybrid approach - the combined detection method incorporating the SVM-based discriminative classifiers yields superior detection performances compared to prior works in multiview object detection.
本文提出了一种新的判别-生成混合方法,重点研究了该方法在多视图目标检测中的应用。我们的方法包括一种新的生成模型,称为随机属性关系图(RARG),它能够捕获从物体中提取的零件的结构和外观特征。我们开发了新的变分学习方法来计算检测似然比函数的近似值。变似然比函数可以显示为在RARG的节点和边缘定义的单个生成分类器的线性组合。这启发我们将节点和边缘的生成分类器替换为判别分类器,如支持向量机,以进一步提高检测性能。我们的实验已经证明了混合方法的鲁棒性-与先前的多视图目标检测工作相比,结合基于svm的判别分类器的组合检测方法产生了更好的检测性能。
{"title":"A Generative-Discriminative Hybrid Method for Multi-View Object Detection","authors":"Dongqing Zhang, Shih-Fu Chang","doi":"10.1109/CVPR.2006.27","DOIUrl":"https://doi.org/10.1109/CVPR.2006.27","url":null,"abstract":"We present a novel discriminative-generative hybrid approach in this paper, with emphasis on application in multiview object detection. Our method includes a novel generative model called Random Attributed Relational Graph (RARG) which is able to capture the structural and appearance characteristics of parts extracted from objects. We develop new variational learning methods to compute the approximation of the detection likelihood ratio function. The variaitonal likelihood ratio function can be shown to be a linear combination of the individual generative classifiers defined at nodes and edges of the RARG. Such insight inspires us to replace the generative classifiers at nodes and edges with discriminative classifiers, such as support vector machines, to further improve the detection performance. Our experiments have shown the robustness of the hybrid approach - the combined detection method incorporating the SVM-based discriminative classifiers yields superior detection performances compared to prior works in multiview object detection.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115082689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Performance Modeling and Prediction of Face Recognition Systems 人脸识别系统的性能建模与预测
Peng Wang, Q. Ji
It is a challenging task to accurately model the performance of a face recognition system, and to predict its individual recognition results under various environments. This paper presents generic methods to model and predict the face recognition performance based on analysis of similarity measurement. We first introduce a concept of "perfect recognition", which only depends on the intrinsic structure of a recognition system. A metric extracted from perfect recognition similarity scores (PRSS) allows modeling the face recognition performance without empirical testing. This paper also presents an EM algorithm to predict the recognition rate of a query set. Furthermore, features are extracted from similarity scores to predict recognition results of individual queries. The presented methods can select algorithm parameters offline, predict recognition performance online, and adjust face alignment online for better recognition. Experimental results show that the performance of recognition systems can be greatly improved using presented methods.
准确地对人脸识别系统的性能进行建模,并预测其在各种环境下的个体识别结果,是一项具有挑战性的任务。本文提出了基于相似性度量分析的人脸识别性能建模和预测的通用方法。我们首先引入“完美识别”的概念,它只依赖于识别系统的内在结构。从完美识别相似度分数(PRSS)中提取的度量可以在没有经验测试的情况下对人脸识别性能进行建模。本文还提出了一种预测查询集识别率的EM算法。此外,从相似度分数中提取特征来预测单个查询的识别结果。该方法可以离线选择算法参数,在线预测识别性能,在线调整人脸对齐以获得更好的识别效果。实验结果表明,采用该方法可以大大提高识别系统的性能。
{"title":"Performance Modeling and Prediction of Face Recognition Systems","authors":"Peng Wang, Q. Ji","doi":"10.1109/CVPR.2006.222","DOIUrl":"https://doi.org/10.1109/CVPR.2006.222","url":null,"abstract":"It is a challenging task to accurately model the performance of a face recognition system, and to predict its individual recognition results under various environments. This paper presents generic methods to model and predict the face recognition performance based on analysis of similarity measurement. We first introduce a concept of \"perfect recognition\", which only depends on the intrinsic structure of a recognition system. A metric extracted from perfect recognition similarity scores (PRSS) allows modeling the face recognition performance without empirical testing. This paper also presents an EM algorithm to predict the recognition rate of a query set. Furthermore, features are extracted from similarity scores to predict recognition results of individual queries. The presented methods can select algorithm parameters offline, predict recognition performance online, and adjust face alignment online for better recognition. Experimental results show that the performance of recognition systems can be greatly improved using presented methods.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116435029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Fully Automatic Registration of 3D Point Clouds 全自动注册3D点云
A. Makadia, Alexander Patterson, Kostas Daniilidis
We propose a novel technique for the registration of 3D point clouds which makes very few assumptions: we avoid any manual rough alignment or the use of landmarks, displacement can be arbitrarily large, and the two point sets can have very little overlap. Crude alignment is achieved by estimation of the 3D-rotation from two Extended Gaussian Images even when the data sets inducing them have partial overlap. The technique is based on the correlation of the two EGIs in the Fourier domain and makes use of the spherical and rotational harmonic transforms. For pairs with low overlap which fail a critical verification step, the rotational alignment can be obtained by the alignment of constellation images generated from the EGIs. Rotationally aligned sets are matched by correlation using the Fourier transform of volumetric functions. A fine alignment is acquired in the final step by running Iterative Closest Points with just few iterations.
我们提出了一种新的3D点云配准技术,它只做了很少的假设:我们避免了任何人工粗糙对准或使用地标,位移可以任意大,两个点集可以有很少的重叠。粗糙对齐是通过估计两个扩展高斯图像的三维旋转来实现的,即使诱导它们的数据集有部分重叠。该技术是基于两个egi在傅里叶域中的相关性,并利用球面和旋转谐波变换。对于没有通过关键验证步骤的低重叠对,可以通过对EGIs生成的星座图像进行对齐来获得旋转对准。旋转对齐的集合通过使用体积函数的傅里叶变换进行相关匹配。在最后一步中,通过运行迭代最近点(Iterative nearest Points)来获得精确的对齐。
{"title":"Fully Automatic Registration of 3D Point Clouds","authors":"A. Makadia, Alexander Patterson, Kostas Daniilidis","doi":"10.1109/CVPR.2006.122","DOIUrl":"https://doi.org/10.1109/CVPR.2006.122","url":null,"abstract":"We propose a novel technique for the registration of 3D point clouds which makes very few assumptions: we avoid any manual rough alignment or the use of landmarks, displacement can be arbitrarily large, and the two point sets can have very little overlap. Crude alignment is achieved by estimation of the 3D-rotation from two Extended Gaussian Images even when the data sets inducing them have partial overlap. The technique is based on the correlation of the two EGIs in the Fourier domain and makes use of the spherical and rotational harmonic transforms. For pairs with low overlap which fail a critical verification step, the rotational alignment can be obtained by the alignment of constellation images generated from the EGIs. Rotationally aligned sets are matched by correlation using the Fourier transform of volumetric functions. A fine alignment is acquired in the final step by running Iterative Closest Points with just few iterations.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"209 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123672940","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 352
Using Stationary-Dynamic Camera Assemblies for Wide-area Video Surveillance and Selective Attention 利用静止-动态摄像机组件进行广域视频监控和选择性注意
Ankur Jain, Dan Koppel, Kyle Kakligian, Yuan-fang Wang
In this paper, we present a prototype video surveillance system that uses stationary-dynamic (or master-slave) camera assemblies to achieve wide-area surveillance and selective focus-of-attention. We address two critical issues in deploying such camera assemblies in real-world applications: off-line camera calibration and on-line selective focus-ofattention. Our contributions over existing techniques are twofold: (1) in terms of camera calibration, our technique calibrates all degrees-of-freedom (DOFs) of both stationary and dynamic cameras, using a closed-form solution that is both efficient and accurate, and (2) in terms of selective focus-of-attention, our technique correctly handles dynamic changes in the scene and varying object depths. This is a significant improvement over existing techniques that use an expensive and non-adaptable table-look-up process.
在本文中,我们提出了一个原型视频监控系统,它使用静止-动态(或主从)摄像机组件来实现广域监控和选择性关注焦点。我们解决了在实际应用中部署此类相机组件的两个关键问题:离线相机校准和在线选择焦点。我们对现有技术的贡献是双重的:(1)在相机校准方面,我们的技术校准固定和动态相机的所有自由度(dof),使用既高效又准确的封闭形式解决方案;(2)在选择性焦点方面,我们的技术正确处理场景中的动态变化和不同的物体深度。这是对现有技术的重大改进,现有技术使用昂贵且不适应的表查找过程。
{"title":"Using Stationary-Dynamic Camera Assemblies for Wide-area Video Surveillance and Selective Attention","authors":"Ankur Jain, Dan Koppel, Kyle Kakligian, Yuan-fang Wang","doi":"10.1109/CVPR.2006.327","DOIUrl":"https://doi.org/10.1109/CVPR.2006.327","url":null,"abstract":"In this paper, we present a prototype video surveillance system that uses stationary-dynamic (or master-slave) camera assemblies to achieve wide-area surveillance and selective focus-of-attention. We address two critical issues in deploying such camera assemblies in real-world applications: off-line camera calibration and on-line selective focus-ofattention. Our contributions over existing techniques are twofold: (1) in terms of camera calibration, our technique calibrates all degrees-of-freedom (DOFs) of both stationary and dynamic cameras, using a closed-form solution that is both efficient and accurate, and (2) in terms of selective focus-of-attention, our technique correctly handles dynamic changes in the scene and varying object depths. This is a significant improvement over existing techniques that use an expensive and non-adaptable table-look-up process.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126173744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Face Recognition using 2.5D Shape Information 使用2.5D形状信息的人脸识别
W. Smith, E. Hancock
In this paper we investigate whether the 2.5D shape information delivered by a novel shape-from-shading algorithm can be used for illumination insensitive face recognition. We present a robust and efficient facial shape-fromshading algorithm which uses principal geodesic analysis to model the variation in surface orientation across a face. We show how this algorithm can be used to recover accurate facial shape and albedo from real world images. Our second contribution is to use the recovered 2.5D shape information in a variety of recognition methods. We present a novel recognition strategy in which similarity is measured in the space of the principal geodesic parameters. We also use the recovered shape information to generate illumination normalised prototype images on which recognition can be performed. Finally we show that, from a single input image, we are able to generate the basis images employed by a number of well known illumination-insensitive recognition algorithms. We also demonstrate that the principal geodesics provide an efficient parameterisation of the space of harmonic basis images.
本文研究了一种新的形状-阴影算法所传递的2.5D形状信息是否可以用于光照不敏感的人脸识别。我们提出了一种鲁棒且高效的面部形状-阴影算法,该算法使用主测地线分析来模拟面部表面方向的变化。我们展示了如何使用该算法从真实世界的图像中恢复准确的面部形状和反照率。我们的第二个贡献是在各种识别方法中使用恢复的2.5D形状信息。我们提出了一种新的识别策略,在主要测地线参数的空间中测量相似性。我们还使用恢复的形状信息来生成照明归一化的原型图像,可以对其进行识别。最后,我们表明,从一个单一的输入图像,我们能够产生的基础图像采用了许多众所周知的照明不敏感识别算法。我们还证明了主测地线提供了谐波基图像空间的有效参数化。
{"title":"Face Recognition using 2.5D Shape Information","authors":"W. Smith, E. Hancock","doi":"10.1109/CVPR.2006.117","DOIUrl":"https://doi.org/10.1109/CVPR.2006.117","url":null,"abstract":"In this paper we investigate whether the 2.5D shape information delivered by a novel shape-from-shading algorithm can be used for illumination insensitive face recognition. We present a robust and efficient facial shape-fromshading algorithm which uses principal geodesic analysis to model the variation in surface orientation across a face. We show how this algorithm can be used to recover accurate facial shape and albedo from real world images. Our second contribution is to use the recovered 2.5D shape information in a variety of recognition methods. We present a novel recognition strategy in which similarity is measured in the space of the principal geodesic parameters. We also use the recovered shape information to generate illumination normalised prototype images on which recognition can be performed. Finally we show that, from a single input image, we are able to generate the basis images employed by a number of well known illumination-insensitive recognition algorithms. We also demonstrate that the principal geodesics provide an efficient parameterisation of the space of harmonic basis images.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129702093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Incremental learning of object detectors using a visual shape alphabet 使用视觉形状字母的目标检测器的增量学习
A. Opelt, A. Pinz, Andrew Zisserman
We address the problem of multiclass object detection. Our aims are to enable models for new categories to benefit from the detectors built previously for other categories, and for the complexity of the multiclass system to grow sublinearly with the number of categories. To this end we introduce a visual alphabet representation which can be learnt incrementally, and explicitly shares boundary fragments (contours) and spatial configurations (relation to centroid) across object categories. We develop a learning algorithm with the following novel contributions: (i) AdaBoost is adapted to learn jointly, based on shape features; (ii) a new learning schedule enables incremental additions of new categories; and (iii) the algorithm learns to detect objects (instead of categorizing images). Furthermore, we show that category similarities can be predicted from the alphabet. We obtain excellent experimental results on a variety of complex categories over several visual aspects. We show that the sharing of shape features not only reduces the number of features required per category, but also often improves recognition performance, as compared to individual detectors which are trained on a per-class basis.
我们解决了多类目标的检测问题。我们的目标是使新类别的模型受益于先前为其他类别构建的检测器,并使多类别系统的复杂性随着类别数量的次线性增长。为此,我们引入了一种可以增量学习的视觉字母表表示,并明确地跨对象类别共享边界片段(轮廓)和空间配置(与质心的关系)。我们开发了一种具有以下新颖贡献的学习算法:(i) AdaBoost适应于基于形状特征的联合学习;(ii)新的学习时间表可以增加新类别;(3)算法学习检测物体(而不是对图像进行分类)。此外,我们表明类别相似性可以从字母表预测。我们在多个视觉方面对各种复杂类别获得了很好的实验结果。我们表明,与在每个类的基础上训练的单个检测器相比,形状特征的共享不仅减少了每个类别所需的特征数量,而且经常提高识别性能。
{"title":"Incremental learning of object detectors using a visual shape alphabet","authors":"A. Opelt, A. Pinz, Andrew Zisserman","doi":"10.1109/CVPR.2006.153","DOIUrl":"https://doi.org/10.1109/CVPR.2006.153","url":null,"abstract":"We address the problem of multiclass object detection. Our aims are to enable models for new categories to benefit from the detectors built previously for other categories, and for the complexity of the multiclass system to grow sublinearly with the number of categories. To this end we introduce a visual alphabet representation which can be learnt incrementally, and explicitly shares boundary fragments (contours) and spatial configurations (relation to centroid) across object categories. We develop a learning algorithm with the following novel contributions: (i) AdaBoost is adapted to learn jointly, based on shape features; (ii) a new learning schedule enables incremental additions of new categories; and (iii) the algorithm learns to detect objects (instead of categorizing images). Furthermore, we show that category similarities can be predicted from the alphabet. We obtain excellent experimental results on a variety of complex categories over several visual aspects. We show that the sharing of shape features not only reduces the number of features required per category, but also often improves recognition performance, as compared to individual detectors which are trained on a per-class basis.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129917868","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 206
Modeling and Classifying Breast Tissue Density in Mammograms 乳房x光照片中乳腺组织密度的建模和分类
Anna Bosch, X. Muñoz, A. Oliver, J. Martí
We present a new approach to model and classify breast parenchymal tissue. Given a mammogram, first, we will discover the distribution of the different tissue densities in an unsupervised manner, and second, we will use this tissue distribution to perform the classification. We achieve this using a classifier based on local descriptors and probabilistic Latent Semantic Analysis (pLSA), a generative model from the statistical text literature. We studied the influence of different descriptors like texture and SIFT features at the classification stage showing that textons outperform SIFT in all cases. Moreover we demonstrate that pLSA automatically extracts meaningful latent aspects generating a compact tissue representation based on their densities, useful for discriminating on mammogram classification. We show the results of tissue classification over the MIAS and DDSM datasets. We compare our method with approaches that classified these same datasets showing a better performance of our proposal.
我们提出了一种新的方法来建模和分类乳腺实质组织。给定一张乳房x光片,首先,我们将以一种无监督的方式发现不同组织密度的分布,其次,我们将使用这种组织分布来进行分类。我们使用基于局部描述符和概率潜在语义分析(pLSA)的分类器来实现这一点,pLSA是一种来自统计文本文献的生成模型。我们研究了不同描述符如纹理和SIFT特征在分类阶段的影响,结果表明纹理在所有情况下都优于SIFT。此外,我们证明了pLSA自动提取有意义的潜在方面,根据它们的密度生成紧凑的组织表示,有助于区分乳房x线照片分类。我们展示了在MIAS和DDSM数据集上的组织分类结果。我们将我们的方法与对这些相同数据集进行分类的方法进行比较,显示出我们的建议的更好性能。
{"title":"Modeling and Classifying Breast Tissue Density in Mammograms","authors":"Anna Bosch, X. Muñoz, A. Oliver, J. Martí","doi":"10.1109/CVPR.2006.188","DOIUrl":"https://doi.org/10.1109/CVPR.2006.188","url":null,"abstract":"We present a new approach to model and classify breast parenchymal tissue. Given a mammogram, first, we will discover the distribution of the different tissue densities in an unsupervised manner, and second, we will use this tissue distribution to perform the classification. We achieve this using a classifier based on local descriptors and probabilistic Latent Semantic Analysis (pLSA), a generative model from the statistical text literature. We studied the influence of different descriptors like texture and SIFT features at the classification stage showing that textons outperform SIFT in all cases. Moreover we demonstrate that pLSA automatically extracts meaningful latent aspects generating a compact tissue representation based on their densities, useful for discriminating on mammogram classification. We show the results of tissue classification over the MIAS and DDSM datasets. We compare our method with approaches that classified these same datasets showing a better performance of our proposal.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129919509","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 104
A Shape Representation for Planar Curves by Shape Signature Harmonic Embedding 基于形状特征谐波嵌入的平面曲线形状表示
Sang-Mook Lee, A. L. Abbott, Neil A. Clark, P. Araman
This paper introduces a new representation for planar curves. From the well-known Dirichlet problem for a disk, the harmonic function embedded in a circular disk is solely dependent on specified boundary values and can be obtained from Poisson’s integral formula. We derive a discrete version of Poisson’s formula and assess its harmonic properties. Various shape signatures can be used as boundary values, whereas only the corresponding Fourier descriptors are needed for the framework. The proposed approach is similar to a scale space representation but exhibits greater generality by accommodating using any type of shape signature. In addition, it is robust to noise and computationally efficient, and it is guaranteed to have a unique solution. In this paper, we demonstrate that the approach has strong potential for shape representation and matching applications.
介绍了平面曲线的一种新的表示方法。由圆盘的狄利克雷问题可知,嵌入在圆盘上的调和函数完全依赖于特定的边值,可以由泊松积分公式得到。我们推导了泊松公式的离散版本,并评估了它的调和性质。可以使用各种形状特征作为边界值,而框架只需要相应的傅里叶描述子。所提出的方法类似于尺度空间表示,但通过容纳任何类型的形状签名,显示出更大的通用性。此外,该方法对噪声具有鲁棒性,计算效率高,保证有唯一解。在本文中,我们证明了该方法在形状表示和匹配应用方面具有很强的潜力。
{"title":"A Shape Representation for Planar Curves by Shape Signature Harmonic Embedding","authors":"Sang-Mook Lee, A. L. Abbott, Neil A. Clark, P. Araman","doi":"10.1109/CVPR.2006.40","DOIUrl":"https://doi.org/10.1109/CVPR.2006.40","url":null,"abstract":"This paper introduces a new representation for planar curves. From the well-known Dirichlet problem for a disk, the harmonic function embedded in a circular disk is solely dependent on specified boundary values and can be obtained from Poisson’s integral formula. We derive a discrete version of Poisson’s formula and assess its harmonic properties. Various shape signatures can be used as boundary values, whereas only the corresponding Fourier descriptors are needed for the framework. The proposed approach is similar to a scale space representation but exhibits greater generality by accommodating using any type of shape signature. In addition, it is robust to noise and computationally efficient, and it is guaranteed to have a unique solution. In this paper, we demonstrate that the approach has strong potential for shape representation and matching applications.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130056620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Composite Templates for Cloth Modeling and Sketching 复合模板布料建模和素描
Hong Chen, Zijian Xu, Ziqiang Liu, Song-Chun Zhu
Cloth modeling and recognition is an important and challenging problem in both vision and graphics tasks, such as dressed human recognition and tracking, human sketch and portrait. In this paper, we present a context sensitive grammar in an And-Or graph representation which will produce a large set of composite graphical templates to account for the wide variabilities of cloth configurations, such as T-shirts, jackets, etc. In a supervised learning phase, we ask an artist to draw sketches on a set of dressed people, and we decompose the sketches into categories of cloth and body components: collars, shoulders, cuff, hands, pants, shoes etc. Each component has a number of distinct subtemplates (sub-graphs). These sub-templates serve as leafnodes in a big And-Or graph where an And-node represents a decomposition of the graph into sub-configurations with Markov relations for context and constraints (soft or hard), and an Or-node is a switch for choosing one out of a set of alternative And-nodes (sub-configurations) - similar to a node in stochastic context free grammar (SCFG). This representation integrates the SCFG for structural variability and the Markov (graphical) model for context. An algorithm which integrates the bottom-up proposals and the topdown information is proposed to infer the composite cloth template from the image.
在服装识别和跟踪、人体素描和肖像等视觉和图形任务中,服装建模和识别都是一个重要而具有挑战性的问题。在本文中,我们在And-Or图表示中提出了一种上下文敏感语法,该语法将产生一组大的复合图形模板,以解释布料配置的广泛变化,如t恤,夹克等。在监督学习阶段,我们请一位艺术家在一组穿着衣服的人身上画草图,我们将草图分解为布料和身体成分的类别:衣领、肩膀、袖口、手、裤子、鞋子等。每个组件都有许多不同的子模板(子图)。这些子模板充当大型and - or图中的叶节点,其中and -node表示将图分解为具有上下文和约束(软或硬)的马尔可夫关系的子配置,or -node是用于从一组可选and -node(子配置)中选择一个的开关——类似于随机上下文自由语法(SCFG)中的节点。这种表示集成了用于结构可变性的SCFG和用于上下文的马尔可夫(图形)模型。提出了一种融合自底向上和自顶向下信息的算法,从图像中推断出复合布模板。
{"title":"Composite Templates for Cloth Modeling and Sketching","authors":"Hong Chen, Zijian Xu, Ziqiang Liu, Song-Chun Zhu","doi":"10.1109/CVPR.2006.81","DOIUrl":"https://doi.org/10.1109/CVPR.2006.81","url":null,"abstract":"Cloth modeling and recognition is an important and challenging problem in both vision and graphics tasks, such as dressed human recognition and tracking, human sketch and portrait. In this paper, we present a context sensitive grammar in an And-Or graph representation which will produce a large set of composite graphical templates to account for the wide variabilities of cloth configurations, such as T-shirts, jackets, etc. In a supervised learning phase, we ask an artist to draw sketches on a set of dressed people, and we decompose the sketches into categories of cloth and body components: collars, shoulders, cuff, hands, pants, shoes etc. Each component has a number of distinct subtemplates (sub-graphs). These sub-templates serve as leafnodes in a big And-Or graph where an And-node represents a decomposition of the graph into sub-configurations with Markov relations for context and constraints (soft or hard), and an Or-node is a switch for choosing one out of a set of alternative And-nodes (sub-configurations) - similar to a node in stochastic context free grammar (SCFG). This representation integrates the SCFG for structural variability and the Markov (graphical) model for context. An algorithm which integrates the bottom-up proposals and the topdown information is proposed to infer the composite cloth template from the image.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"140 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132867556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 156
A Multi-Channel Algorithm for Edge Detection Under Varying Lighting
W. Xu, M. Jenkin, Y. Lespérance
In vision-based autonomous spacecraft docking multiple views of scene structure captured with the same camera and scene geometry is available under different lighting conditions. These "multiple-exposure" images must be processed to localize visual features to compute the pose of the target object. This paper describes a robust multi-channel edge detection algorithm that localizes the structure of the target object from the local gradient distribution computed over these multiple-exposure images. This approach reduces the effect of the illumination variation including the effect of shadow edges over the use of a single image. Experiments demonstrate that this approach has a lower false detection rate than the average response of the Canny edge detector applied to the individual images separately.
在基于视觉的自主航天器对接中,在不同的光照条件下,可以使用相同的相机和场景几何形状捕获多个场景结构视图。这些“多次曝光”的图像必须经过处理,以定位视觉特征,从而计算目标物体的姿态。本文描述了一种鲁棒的多通道边缘检测算法,该算法通过在这些多次曝光图像上计算的局部梯度分布来定位目标物体的结构。这种方法减少了光照变化的影响,包括使用单个图像时阴影边缘的影响。实验表明,该方法比Canny边缘检测器单独应用于单个图像的平均响应具有更低的误检率。
{"title":"A Multi-Channel Algorithm for Edge Detection Under Varying Lighting","authors":"W. Xu, M. Jenkin, Y. Lespérance","doi":"10.1109/CVPR.2006.33","DOIUrl":"https://doi.org/10.1109/CVPR.2006.33","url":null,"abstract":"In vision-based autonomous spacecraft docking multiple views of scene structure captured with the same camera and scene geometry is available under different lighting conditions. These \"multiple-exposure\" images must be processed to localize visual features to compute the pose of the target object. This paper describes a robust multi-channel edge detection algorithm that localizes the structure of the target object from the local gradient distribution computed over these multiple-exposure images. This approach reduces the effect of the illumination variation including the effect of shadow edges over the use of a single image. Experiments demonstrate that this approach has a lower false detection rate than the average response of the Canny edge detector applied to the individual images separately.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130599524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1