首页 > 最新文献

2009 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
A graph-based approach to skin mole matching incorporating template-normalized coordinates 结合模板归一化坐标的基于图的皮肤痣匹配方法
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206725
H. Mirzaalian, G. Hamarneh, Tim K. Lee
Density of moles is a strong predictor of malignant melanoma. Some dermatologists advocate periodic full-body scan for high-risk patients. In current practice, physicians compare images taken at different time instances to recognize changes. There is an important clinical need to follow changes in the number of moles and their appearance (size, color, texture, shape) in images from two different times. In this paper, we propose a method for finding corresponding moles in patient's skin back images at different scanning times. At first, a template is defined for the human back to calculate the moles' normalized spatial coordinates. Next, matching moles across images is modeled as a graph matching problem and algebraic relations between nodes and edges in the graphs are induced in the matching cost function, which contains terms reflecting proximity regularization, angular agreement between mole pairs, and agreement between the moles' normalized coordinates calculated in the unwarped back template. We propose and discuss alternative approaches for evaluating the goodness of matching. We evaluate our method on a large set of synthetic data (hundreds of pairs) as well as 56 pairs of real dermatological images. Our proposed method compares favorably with the state-of-the-art.
痣的密度是恶性黑色素瘤的一个强有力的预测指标。一些皮肤科医生提倡对高危患者进行定期全身扫描。在目前的实践中,医生比较在不同时间拍摄的图像来识别变化。有一个重要的临床需要跟踪变化的痣的数量和外观(大小,颜色,质地,形状)的图像从两个不同的时间。在本文中,我们提出了一种在不同扫描时间的患者皮肤背部图像中寻找相应痣的方法。首先为人体背部定义一个模板,计算鼹鼠的归一化空间坐标。接下来,将图像间的痣匹配建模为图匹配问题,并在匹配成本函数中归纳图中节点和边之间的代数关系,该函数包含反映邻近正则化、痣对之间的角度一致性以及在未弯曲的背模板中计算的痣的归一化坐标之间的一致性的项。我们提出并讨论了评估匹配优度的替代方法。我们在大量合成数据(数百对)以及56对真实皮肤病学图像上评估了我们的方法。我们提出的方法比最先进的方法要好。
{"title":"A graph-based approach to skin mole matching incorporating template-normalized coordinates","authors":"H. Mirzaalian, G. Hamarneh, Tim K. Lee","doi":"10.1109/CVPR.2009.5206725","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206725","url":null,"abstract":"Density of moles is a strong predictor of malignant melanoma. Some dermatologists advocate periodic full-body scan for high-risk patients. In current practice, physicians compare images taken at different time instances to recognize changes. There is an important clinical need to follow changes in the number of moles and their appearance (size, color, texture, shape) in images from two different times. In this paper, we propose a method for finding corresponding moles in patient's skin back images at different scanning times. At first, a template is defined for the human back to calculate the moles' normalized spatial coordinates. Next, matching moles across images is modeled as a graph matching problem and algebraic relations between nodes and edges in the graphs are induced in the matching cost function, which contains terms reflecting proximity regularization, angular agreement between mole pairs, and agreement between the moles' normalized coordinates calculated in the unwarped back template. We propose and discuss alternative approaches for evaluating the goodness of matching. We evaluate our method on a large set of synthetic data (hundreds of pairs) as well as 56 pairs of real dermatological images. Our proposed method compares favorably with the state-of-the-art.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124286214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Multiple instance fFeature for robust part-based object detection 多实例特征鲁棒的基于零件的目标检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206858
Zhe L. Lin, G. Hua, L. Davis
Feature misalignment in object detection refers to the phenomenon that features which fire up in some positive detection windows do not fire up in other positive detection windows. Most often it is caused by pose variation and local part deformation. Previous work either totally ignores this issue, or naively performs a local exhaustive search to better position each feature. We propose a learning framework to mitigate this problem, where a boosting algorithm is performed to seed the position of the object part, and a multiple instance boosting algorithm further pursues an aggregated feature for this part, namely multiple instance feature. Unlike most previous boosting based object detectors, where each feature value produces a single classification result, the value of the proposed multiple instance feature is the Noisy-OR integration of a bag of classification results. Our approach is applied to the task of human detection and is tested on two popular benchmarks. The proposed approach brings significant improvement in performance, i.e., smaller number of features used in the cascade and better detection accuracy.
目标检测中的特征错位是指在某个正检测窗口中触发的特征在其他正检测窗口中没有触发的现象。最常见的原因是位姿变化和局部变形。以前的工作要么完全忽略了这个问题,要么天真地执行局部穷举搜索来更好地定位每个特征。我们提出了一个学习框架来缓解这一问题,其中执行增强算法来播种对象部分的位置,多实例增强算法进一步追求该部分的聚合特征,即多实例特征。与之前大多数基于增强的目标检测器不同,其中每个特征值产生单个分类结果,而所提出的多实例特征值是一组分类结果的noise - or集成。我们的方法应用于人类检测任务,并在两个流行的基准上进行了测试。该方法在性能上有了显著的提高,即级联中使用的特征数量更少,检测精度更高。
{"title":"Multiple instance fFeature for robust part-based object detection","authors":"Zhe L. Lin, G. Hua, L. Davis","doi":"10.1109/CVPR.2009.5206858","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206858","url":null,"abstract":"Feature misalignment in object detection refers to the phenomenon that features which fire up in some positive detection windows do not fire up in other positive detection windows. Most often it is caused by pose variation and local part deformation. Previous work either totally ignores this issue, or naively performs a local exhaustive search to better position each feature. We propose a learning framework to mitigate this problem, where a boosting algorithm is performed to seed the position of the object part, and a multiple instance boosting algorithm further pursues an aggregated feature for this part, namely multiple instance feature. Unlike most previous boosting based object detectors, where each feature value produces a single classification result, the value of the proposed multiple instance feature is the Noisy-OR integration of a bag of classification results. Our approach is applied to the task of human detection and is tested on two popular benchmarks. The proposed approach brings significant improvement in performance, i.e., smaller number of features used in the cascade and better detection accuracy.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126195298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Shape of Gaussians as feature descriptors 高斯函数的形状作为特征描述符
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206506
Liyu Gong, Tianjiang Wang, Fang Liu
This paper introduces a feature descriptor called shape of Gaussian (SOG), which is based on a general feature descriptor design framework called shape of signal probability density function (SOSPDF). SOSPDF takes the shape of a signal's probability density function (pdf) as its feature. Under such a view, both histogram and region covariance often used in computer vision are SOSPDF features. Histogram describes SOSPDF by a discrete approximation way. Region covariance describes SOSPDF as an incomplete parameterized multivariate Gaussian distribution. Our proposed SOG descriptor is a full parameterized Gaussian, so it has all the advantages of region covariance and is more effective. Furthermore, we identify that SOGs form a Lie group. Based on Lie group theory, we propose a distance metric for SOG. We test SOG features in tracking problem. Experiments show better tracking results compared with region covariance. Moreover, experiment results indicate that SOG features attempt to harvest more useful information and are less sensitive against noise.
本文介绍了一种基于信号概率密度函数形状(SOSPDF)的通用特征描述子设计框架的高斯形状特征描述子(SOG)。SOSPDF以信号的概率密度函数(pdf)的形状为特征。在这种观点下,计算机视觉中常用的直方图和区域协方差都是SOSPDF特征。直方图以离散逼近的方式描述SOSPDF。区域协方差将SOSPDF描述为一个不完全参数化的多元高斯分布。我们提出的SOG描述符是一个全参数化的高斯描述符,因此它具有区域协方差的所有优点,并且更加有效。此外,我们发现sog形成李群。基于李群理论,提出了一种SOG的距离度量。我们在跟踪问题中测试SOG特性。实验表明,与区域协方差法相比,跟踪效果更好。此外,实验结果表明,SOG特征试图获取更多有用的信息,对噪声的敏感性较低。
{"title":"Shape of Gaussians as feature descriptors","authors":"Liyu Gong, Tianjiang Wang, Fang Liu","doi":"10.1109/CVPR.2009.5206506","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206506","url":null,"abstract":"This paper introduces a feature descriptor called shape of Gaussian (SOG), which is based on a general feature descriptor design framework called shape of signal probability density function (SOSPDF). SOSPDF takes the shape of a signal's probability density function (pdf) as its feature. Under such a view, both histogram and region covariance often used in computer vision are SOSPDF features. Histogram describes SOSPDF by a discrete approximation way. Region covariance describes SOSPDF as an incomplete parameterized multivariate Gaussian distribution. Our proposed SOG descriptor is a full parameterized Gaussian, so it has all the advantages of region covariance and is more effective. Furthermore, we identify that SOGs form a Lie group. Based on Lie group theory, we propose a distance metric for SOG. We test SOG features in tracking problem. Experiments show better tracking results compared with region covariance. Moreover, experiment results indicate that SOG features attempt to harvest more useful information and are less sensitive against noise.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126401348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Angular embedding: From jarring intensity differences to perceived luminance 角嵌入:从不和谐的强度差异到感知亮度
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206673
Stella X. Yu
Our goal is to turn an intensity image into its perceived luminance without parsing it into depths, surfaces, or scene illuminations. We start with jarring intensity differences at two scales mixed according to edges, identified by a pixel-centric edge detector. We propose angular embedding as a more robust, efficient, and versatile alternative to LS, LLE, and NCUTS for obtaining a global brightness ordering from local differences. Our model explains a variety of brightness illusions with a single algorithm. Brightness of a pixel can be understood locally as its intensity deviating in the gradient direction and globally as finding its rank relative to others, particularly the lightest and darkest ones.
我们的目标是将强度图像转换为其可感知的亮度,而无需将其解析为深度,表面或场景照明。我们从根据边缘混合的两个尺度上的强度差异开始,由像素中心边缘检测器识别。我们提出角嵌入作为一种比LS、LLE和NCUTS更鲁棒、高效和通用的替代方法,可以从局部差异中获得全局亮度排序。我们的模型用一个算法解释了各种各样的亮度错觉。一个像素的亮度在局部可以理解为它在梯度方向上的强度偏差,而在全局上可以理解为它相对于其他像素的等级,尤其是最亮和最暗的像素。
{"title":"Angular embedding: From jarring intensity differences to perceived luminance","authors":"Stella X. Yu","doi":"10.1109/CVPR.2009.5206673","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206673","url":null,"abstract":"Our goal is to turn an intensity image into its perceived luminance without parsing it into depths, surfaces, or scene illuminations. We start with jarring intensity differences at two scales mixed according to edges, identified by a pixel-centric edge detector. We propose angular embedding as a more robust, efficient, and versatile alternative to LS, LLE, and NCUTS for obtaining a global brightness ordering from local differences. Our model explains a variety of brightness illusions with a single algorithm. Brightness of a pixel can be understood locally as its intensity deviating in the gradient direction and globally as finding its rank relative to others, particularly the lightest and darkest ones.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130113898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Visual tracking with online Multiple Instance Learning 视觉跟踪与在线多实例学习
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206737
Boris Babenko, Ming-Hsuan Yang, Serge J. Belongie
In this paper, we address the problem of learning an adaptive appearance model for object tracking. In particular, a class of tracking techniques called “tracking by detection” have been shown to give promising results at real-time speeds. These methods train a discriminative classifier in an online manner to separate the object from the background. This classifier bootstraps itself by using the current tracker state to extract positive and negative examples from the current frame. Slight inaccuracies in the tracker can therefore lead to incorrectly labeled training examples, which degrades the classifier and can cause further drift. In this paper we show that using Multiple Instance Learning (MIL) instead of traditional supervised learning avoids these problems, and can therefore lead to a more robust tracker with fewer parameter tweaks. We present a novel online MIL algorithm for object tracking that achieves superior results with real-time performance.
在本文中,我们解决了学习自适应外观模型用于目标跟踪的问题。特别是,一类被称为“检测跟踪”的跟踪技术已被证明可以在实时速度下提供有希望的结果。这些方法以在线的方式训练一个判别分类器来分离目标和背景。这个分类器通过使用当前跟踪器状态从当前帧中提取正例和负例来引导自己。因此,跟踪器中的轻微不准确可能导致错误标记的训练样例,从而降低分类器的性能,并可能导致进一步的漂移。在本文中,我们表明使用多实例学习(MIL)代替传统的监督学习可以避免这些问题,因此可以使用更少的参数调整产生更鲁棒的跟踪器。提出了一种新的在线MIL目标跟踪算法,该算法具有较好的实时性。
{"title":"Visual tracking with online Multiple Instance Learning","authors":"Boris Babenko, Ming-Hsuan Yang, Serge J. Belongie","doi":"10.1109/CVPR.2009.5206737","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206737","url":null,"abstract":"In this paper, we address the problem of learning an adaptive appearance model for object tracking. In particular, a class of tracking techniques called “tracking by detection” have been shown to give promising results at real-time speeds. These methods train a discriminative classifier in an online manner to separate the object from the background. This classifier bootstraps itself by using the current tracker state to extract positive and negative examples from the current frame. Slight inaccuracies in the tracker can therefore lead to incorrectly labeled training examples, which degrades the classifier and can cause further drift. In this paper we show that using Multiple Instance Learning (MIL) instead of traditional supervised learning avoids these problems, and can therefore lead to a more robust tracker with fewer parameter tweaks. We present a novel online MIL algorithm for object tracking that achieves superior results with real-time performance.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128490070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1986
Resolution-Invariant Image Representation and its applications 分辨率不变图像表示及其应用
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206679
Jinjun Wang, Shenghuo Zhu, Yihong Gong
We present a resolution-invariant image representation (RIIR) framework in this paper. The RIIR framework includes the methods of building a set of multi-resolution bases from training images, estimating the optimal sparse resolution-invariant representation of any image, and reconstructing the missing patches of any resolution level. As the proposed RIIR framework has many potential resolution enhancement applications, we discuss three novel image magnification applications in this paper. In the first application, we apply the RIIR framework to perform Multi-Scale Image Magnification where we also introduced a training strategy to built a compact RIIR set. In the second application, the RIIR framework is extended to conduct Continuous Image Scaling where a new base at any resolution level can be generated using existing RIIR set on the fly. In the third application, we further apply the RIIR framework onto Content-Base Automatic Zooming applications. The experimental results show that in all these applications, our RIIR based method outperforms existing methods in various aspects.
本文提出了一种分辨率不变图像表示(RIIR)框架。RIIR框架包括从训练图像中构建一组多分辨率基,估计任意图像的最优稀疏分辨率不变表示以及重建任意分辨率水平的缺失补丁的方法。由于所提出的RIIR框架具有许多潜在的分辨率增强应用,因此本文讨论了三种新的图像放大应用。在第一个应用中,我们应用RIIR框架来执行多尺度图像放大,其中我们还引入了一个训练策略来构建一个紧凑的RIIR集。在第二个应用中,RIIR框架被扩展到进行连续图像缩放,其中可以使用现有的RIIR动态生成任何分辨率水平的新基础。在第三个应用程序中,我们进一步将RIIR框架应用到基于内容的自动缩放应用程序中。实验结果表明,在所有这些应用中,基于RIIR的方法在各个方面都优于现有方法。
{"title":"Resolution-Invariant Image Representation and its applications","authors":"Jinjun Wang, Shenghuo Zhu, Yihong Gong","doi":"10.1109/CVPR.2009.5206679","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206679","url":null,"abstract":"We present a resolution-invariant image representation (RIIR) framework in this paper. The RIIR framework includes the methods of building a set of multi-resolution bases from training images, estimating the optimal sparse resolution-invariant representation of any image, and reconstructing the missing patches of any resolution level. As the proposed RIIR framework has many potential resolution enhancement applications, we discuss three novel image magnification applications in this paper. In the first application, we apply the RIIR framework to perform Multi-Scale Image Magnification where we also introduced a training strategy to built a compact RIIR set. In the second application, the RIIR framework is extended to conduct Continuous Image Scaling where a new base at any resolution level can be generated using existing RIIR set on the fly. In the third application, we further apply the RIIR framework onto Content-Base Automatic Zooming applications. The experimental results show that in all these applications, our RIIR based method outperforms existing methods in various aspects.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124540445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Joint and implicit registration for face recognition 人脸识别的联合和隐式配准
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206607
Peng Li, S. Prince
Contemporary face recognition algorithms rely on precise localization of keypoints (corner of eye, nose etc.). Unfortunately, finding keypoints reliably and accurately remains a hard problem. In this paper we pose two questions. First, is it possible to exploit the gallery image in order to find keypoints in the probe image? For instance, consider finding the left eye in the probe image. Rather than using a generic eye model, we use a model that is informed by the appearance of the eye in the gallery image. To this end we develop a probabilistic model which combines recognition and keypoint localization. Second, is it necessary to localize keypoints? Alternatively we can consider keypoint position as a hidden variable which we marginalize over in a Bayesian manner. We demonstrate that both of these innovations improve performance relative to conventional methods in both frontal and cross-pose face recognition.
当代人脸识别算法依赖于关键点(眼角、鼻子等)的精确定位。不幸的是,可靠而准确地找到关键点仍然是一个难题。在本文中,我们提出两个问题。首先,是否有可能利用图库图像来找到探测图像中的关键点?例如,考虑在探针图像中找到左眼。而不是使用一般的眼睛模型,我们使用的模型是由画廊图像中眼睛的外观所提供的信息。为此,我们开发了一种结合识别和关键点定位的概率模型。第二,是否有必要对关键点进行本地化?或者,我们可以将关键点位置视为一个隐变量,我们用贝叶斯方法将其边缘化。我们证明了这两种创新在正面和交叉姿态面部识别方面都比传统方法提高了性能。
{"title":"Joint and implicit registration for face recognition","authors":"Peng Li, S. Prince","doi":"10.1109/CVPR.2009.5206607","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206607","url":null,"abstract":"Contemporary face recognition algorithms rely on precise localization of keypoints (corner of eye, nose etc.). Unfortunately, finding keypoints reliably and accurately remains a hard problem. In this paper we pose two questions. First, is it possible to exploit the gallery image in order to find keypoints in the probe image? For instance, consider finding the left eye in the probe image. Rather than using a generic eye model, we use a model that is informed by the appearance of the eye in the gallery image. To this end we develop a probabilistic model which combines recognition and keypoint localization. Second, is it necessary to localize keypoints? Alternatively we can consider keypoint position as a hidden variable which we marginalize over in a Bayesian manner. We demonstrate that both of these innovations improve performance relative to conventional methods in both frontal and cross-pose face recognition.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124567460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Global optimization for alignment of generalized shapes 广义形状对齐的全局优化
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206548
Hongsheng Li, Tian Shen, Xiaolei Huang
In this paper, we introduce a novel algorithm to solve global shape registration problems. We use gray-scale “images” to represent source shapes, and propose a novel two-component Gaussian Mixtures (GM) distance map representation for target shapes. Based on this flexible asymmetric image-based representation, a new energy function is defined. It proves to be a more robust shape dissimilarity metric that can be computed efficiently. Such high efficiency is essential for global optimization methods. We adopt one of them, the Particle Swarm Optimization (PSO), to effectively estimate the global optimum of the new energy function. Experiments and comparison performed on generalized shape data including continuous shapes, unstructured sparse point sets, and gradient maps, demonstrate the robustness and effectiveness of the algorithm.
本文提出了一种求解全局形状配准问题的新算法。我们使用灰度“图像”来表示源形状,并提出了一种新的双分量高斯混合(GM)距离图表示目标形状。基于这种灵活的非对称图像表示,定义了一个新的能量函数。结果表明,该方法具有较强的鲁棒性和计算效率。这种高效率对于全局优化方法是必不可少的。我们采用其中的粒子群算法(PSO)来有效估计新能量函数的全局最优。在广义形状数据(包括连续形状、非结构化稀疏点集和梯度图)上进行的实验和比较,证明了该算法的鲁棒性和有效性。
{"title":"Global optimization for alignment of generalized shapes","authors":"Hongsheng Li, Tian Shen, Xiaolei Huang","doi":"10.1109/CVPR.2009.5206548","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206548","url":null,"abstract":"In this paper, we introduce a novel algorithm to solve global shape registration problems. We use gray-scale “images” to represent source shapes, and propose a novel two-component Gaussian Mixtures (GM) distance map representation for target shapes. Based on this flexible asymmetric image-based representation, a new energy function is defined. It proves to be a more robust shape dissimilarity metric that can be computed efficiently. Such high efficiency is essential for global optimization methods. We adopt one of them, the Particle Swarm Optimization (PSO), to effectively estimate the global optimum of the new energy function. Experiments and comparison performed on generalized shape data including continuous shapes, unstructured sparse point sets, and gradient maps, demonstrate the robustness and effectiveness of the algorithm.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124595576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
A projector-based movable hand-held display system 一种基于投影仪的可移动手持显示系统
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206658
M. Leung, K. Lee, K. Wong, M. Chang
In this paper, we proposed a movable hand-held display system which uses a projector to project display content onto an ordinary cardboard which can move freely within the projection area. Such a system can give users greater freedom of control of the display such as the viewing angle and distance. At the same time, the size of the cardboard can be made to a size that fits one's application. A projector-camera pair is calibrated and used as the tracking and projection system. We present a vision based algorithm to detect an ordinary cardboard and track its subsequent motion. Display content is then pre-warped and projected onto the cardboard at the correct position. Experimental results show that our system can project onto the cardboard in reasonable precision.
在本文中,我们提出了一种可移动的手持显示系统,该系统使用投影仪将显示内容投影到可在投影区域内自由移动的普通纸板上。这样的系统可以给用户更大的自由控制显示,如视角和距离。与此同时,纸板的大小可以被制成适合一个人的应用的大小。一个投影仪-摄像机对被校准并用作跟踪和投影系统。提出了一种基于视觉的检测普通纸板并跟踪其后续运动的算法。然后将显示内容预翘曲并投影到正确位置的纸板上。实验结果表明,该系统能够以合理的精度投影到纸板上。
{"title":"A projector-based movable hand-held display system","authors":"M. Leung, K. Lee, K. Wong, M. Chang","doi":"10.1109/CVPR.2009.5206658","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206658","url":null,"abstract":"In this paper, we proposed a movable hand-held display system which uses a projector to project display content onto an ordinary cardboard which can move freely within the projection area. Such a system can give users greater freedom of control of the display such as the viewing angle and distance. At the same time, the size of the cardboard can be made to a size that fits one's application. A projector-camera pair is calibrated and used as the tracking and projection system. We present a vision based algorithm to detect an ordinary cardboard and track its subsequent motion. Display content is then pre-warped and projected onto the cardboard at the correct position. Experimental results show that our system can project onto the cardboard in reasonable precision.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121100659","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Motion capture using joint skeleton tracking and surface estimation 基于关节骨架跟踪和表面估计的运动捕捉
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206755
Juergen Gall, Carsten Stoll, Edilson de Aguiar, C. Theobalt, B. Rosenhahn, H. Seidel
This paper proposes a method for capturing the performance of a human or an animal from a multi-view video sequence. Given an articulated template model and silhouettes from a multi-view image sequence, our approach recovers not only the movement of the skeleton, but also the possibly non-rigid temporal deformation of the 3D surface. While large scale deformations or fast movements are captured by the skeleton pose and approximate surface skinning, true small scale deformations or non-rigid garment motion are captured by fitting the surface to the silhouette. We further propose a novel optimization scheme for skeleton-based pose estimation that exploits the skeleton's tree structure to split the optimization problem into a local one and a lower dimensional global one. We show on various sequences that our approach can capture the 3D motion of animals and humans accurately even in the case of rapid movements and wide apparel like skirts.
本文提出了一种从多视图视频序列中捕捉人类或动物表演的方法。给定一个铰接模板模型和多视图图像序列的轮廓,我们的方法不仅可以恢复骨架的运动,还可以恢复3D表面可能的非刚性时间变形。当骨架姿态和近似表面蒙皮捕获大规模变形或快速运动时,通过将表面拟合到轮廓来捕获真正的小规模变形或非刚性服装运动。我们进一步提出了一种新的基于骨架的姿态估计优化方案,该方案利用骨架的树结构将优化问题分为局部问题和低维全局问题。我们在各种序列上展示,我们的方法可以准确地捕捉动物和人类的3D运动,即使在快速运动和像裙子这样的宽衣服的情况下。
{"title":"Motion capture using joint skeleton tracking and surface estimation","authors":"Juergen Gall, Carsten Stoll, Edilson de Aguiar, C. Theobalt, B. Rosenhahn, H. Seidel","doi":"10.1109/CVPR.2009.5206755","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206755","url":null,"abstract":"This paper proposes a method for capturing the performance of a human or an animal from a multi-view video sequence. Given an articulated template model and silhouettes from a multi-view image sequence, our approach recovers not only the movement of the skeleton, but also the possibly non-rigid temporal deformation of the 3D surface. While large scale deformations or fast movements are captured by the skeleton pose and approximate surface skinning, true small scale deformations or non-rigid garment motion are captured by fitting the surface to the silhouette. We further propose a novel optimization scheme for skeleton-based pose estimation that exploits the skeleton's tree structure to split the optimization problem into a local one and a lower dimensional global one. We show on various sequences that our approach can capture the 3D motion of animals and humans accurately even in the case of rapid movements and wide apparel like skirts.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125439184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 455
期刊
2009 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1