首页 > 最新文献

Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)最新文献

英文 中文
Automatic hierarchical classification using time-based co-occurrences 使用基于时间的共现自动分层分类
C. Stauffer
While a tracking system is unaware of the identity of any object it tracks, the identity remains the same for the entire tracking sequence. Our system leverages this information by using accumulated joint cooccurrences of the representations within the sequence to create a hierarchical binary-tree classifier of the representations. This classifier is useful to classify sequences as well as individual instances. We illustrate the use of this method on two separate representations the tracked object's position, movement, and size; and the tracked object's binary motion silhouettes.
虽然跟踪系统不知道它跟踪的任何对象的身份,但整个跟踪序列的身份保持不变。我们的系统通过使用序列中表示的累积联合并发来利用这些信息,从而创建表示的分层二叉树分类器。这个分类器对于分类序列和单个实例都很有用。我们在跟踪对象的位置、运动和大小两个单独的表示上说明了这种方法的使用;以及被跟踪物体的二进制运动轮廓。
{"title":"Automatic hierarchical classification using time-based co-occurrences","authors":"C. Stauffer","doi":"10.1109/CVPR.1999.784654","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784654","url":null,"abstract":"While a tracking system is unaware of the identity of any object it tracks, the identity remains the same for the entire tracking sequence. Our system leverages this information by using accumulated joint cooccurrences of the representations within the sequence to create a hierarchical binary-tree classifier of the representations. This classifier is useful to classify sequences as well as individual instances. We illustrate the use of this method on two separate representations the tracked object's position, movement, and size; and the tracked object's binary motion silhouettes.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86851785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
3D deformable image matching using multiscale minimization of global energy functions 使用全局能量函数的多尺度最小化的三维可变形图像匹配
O. Musse, F. Heitz, J. Armspach
This paper presents a hierarchical framework to perform deformable matching of three dimensional (3D) images. 3D shape deformations are parameterized at different scales, using a decomposition of the continuous deformation vector field over a sequence of nested subspaces, generated from a single scaling function. The parameterization of the field enables to enforce smoothness and differentiability constraints without performing explicit regularization. A global energy function, depending on the reference image and the transformed one, is minimized via a coarse-to-fine algorithm over this multiscale decomposition. Contrary to standard multigrid approaches, no reduction of image data is applied. The continuous field of deformation is always sampled at the same resolution, ensuring that the same energy function is handled at each scale and that the energy decreases at each step of the minimization.
本文提出了一种用于三维图像可变形匹配的分层框架。三维形状变形在不同的尺度上参数化,使用在一系列嵌套子空间上的连续变形向量场的分解,由单个缩放函数生成。该域的参数化可以在不执行显式正则化的情况下强制执行平滑性和可微性约束。一个全局能量函数,取决于参考图像和转换后的图像,通过一个粗到精的算法在这个多尺度分解上最小化。与标准的多重网格方法相反,该方法没有对图像数据进行约简。连续变形场始终以相同的分辨率进行采样,确保在每个尺度上处理相同的能量函数,并确保最小化的每一步能量都在减少。
{"title":"3D deformable image matching using multiscale minimization of global energy functions","authors":"O. Musse, F. Heitz, J. Armspach","doi":"10.1109/CVPR.1999.784724","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784724","url":null,"abstract":"This paper presents a hierarchical framework to perform deformable matching of three dimensional (3D) images. 3D shape deformations are parameterized at different scales, using a decomposition of the continuous deformation vector field over a sequence of nested subspaces, generated from a single scaling function. The parameterization of the field enables to enforce smoothness and differentiability constraints without performing explicit regularization. A global energy function, depending on the reference image and the transformed one, is minimized via a coarse-to-fine algorithm over this multiscale decomposition. Contrary to standard multigrid approaches, no reduction of image data is applied. The continuous field of deformation is always sampled at the same resolution, ensuring that the same energy function is handled at each scale and that the energy decreases at each step of the minimization.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88010025","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Recognition of strings using nonstationary Markovian models: an application in ZIP code recognition 使用非平稳马尔可夫模型的字符串识别:在邮政编码识别中的应用
D. Bouchaffra, Venu Govindaraju, S. Srihari
This paper presents nonstationary Markovian models and their application to recognition of strings of tokens, such as ZIP codes in the US mailstream. Unlike traditional approaches where digits are simply recognized in isolation, the novelty of our approach lies in the manner in which recognitions scores along with domain specific knowledge about the frequency distribution of various combination of digits are all integrated into one unified model. The domain knowledge is derived from postal directory files. This data feeds into the models as n-grams statistics that are seamlessly integrated with recognition scores of digit images. We present the recognition accuracy (90%) achieved on a set of 20,000 ZIP codes.
本文介绍了非平稳马尔可夫模型及其在标记字符串识别中的应用,如美国邮件流中的邮政编码。与传统方法中数字被简单地孤立识别不同,我们方法的新颖之处在于,识别得分以及关于各种数字组合频率分布的领域特定知识都被集成到一个统一的模型中。领域知识来源于邮政目录文件。这些数据作为n-grams统计数据输入到模型中,这些统计数据与数字图像的识别分数无缝集成。我们展示了在一组20,000个邮政编码上实现的识别精度(90%)。
{"title":"Recognition of strings using nonstationary Markovian models: an application in ZIP code recognition","authors":"D. Bouchaffra, Venu Govindaraju, S. Srihari","doi":"10.1109/CVPR.1999.784626","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784626","url":null,"abstract":"This paper presents nonstationary Markovian models and their application to recognition of strings of tokens, such as ZIP codes in the US mailstream. Unlike traditional approaches where digits are simply recognized in isolation, the novelty of our approach lies in the manner in which recognitions scores along with domain specific knowledge about the frequency distribution of various combination of digits are all integrated into one unified model. The domain knowledge is derived from postal directory files. This data feeds into the models as n-grams statistics that are seamlessly integrated with recognition scores of digit images. We present the recognition accuracy (90%) achieved on a set of 20,000 ZIP codes.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84387926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An integral formulation for differential photometric stereo 微分光度立体的积分公式
James J. Clark, H. Pekau
In this paper we present an integral formulation of the active differential photometric stereo algorithm proposed by Clark (1992) and by Iwahori et al. [1992, 1994). The algorithm presented in this paper does not require measurement of derivatives of image quantities, but requires instead the computation of integrals of image quantities. Thus the algorithm is more robust to sensor noise and light source position errors than the Clark-Iwahori algorithm. We show that the algorithm presented in the paper can be efficiently implemented in practice with a planar distributed light source, and present experimental results demonstrating the efficacy of the algorithm.
在本文中,我们提出了Clark(1992)和Iwahori等人[1992,1994]提出的主动微分光度立体算法的积分公式。本文提出的算法不需要测量图像量的导数,而是需要计算图像量的积分。因此,该算法对传感器噪声和光源位置误差的鲁棒性优于Clark-Iwahori算法。实验结果表明,本文提出的算法在实际应用中可以有效地实现平面分布式光源,并给出了验证算法有效性的实验结果。
{"title":"An integral formulation for differential photometric stereo","authors":"James J. Clark, H. Pekau","doi":"10.1109/CVPR.1999.786927","DOIUrl":"https://doi.org/10.1109/CVPR.1999.786927","url":null,"abstract":"In this paper we present an integral formulation of the active differential photometric stereo algorithm proposed by Clark (1992) and by Iwahori et al. [1992, 1994). The algorithm presented in this paper does not require measurement of derivatives of image quantities, but requires instead the computation of integrals of image quantities. Thus the algorithm is more robust to sensor noise and light source position errors than the Clark-Iwahori algorithm. We show that the algorithm presented in the paper can be efficiently implemented in practice with a planar distributed light source, and present experimental results demonstrating the efficacy of the algorithm.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91045048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Background estimation and removal based on range and color 基于范围和颜色的背景估计和去除
G. Gordon, Trevor Darrell, M. Harville, J. Woodfill
Background estimation and removal based on the joint use of range and color data produces superior results than can be achieved with either data source alone. This is increasingly relevant as inexpensive, real-time, passive range systems become more accessible through novel hardware and increased CPU processing speeds. Range is a powerful signal for segmentation which is largely independent of color and hence not effected by the classic color segmentation problems of shadows and objects with color similar to the background. However range alone is also not sufficient for the good segmentation: depth measurements are rarely available at all pixels in the scene, and foreground objects may be indistinguishable in depth when they are close to the background. Color segmentation is complementary in these cases. Surprisingly, little work has been done to date on joint range and color segmentation. We describe and demonstrate a background estimation method based on a multidimensional (range and color) clustering at each image pixel. Segmentation of the foreground in a given frame is performed via comparison with background statistics in range and normalized color. Important implementation issues such as treatment of shadows and low confidence measurements are discussed in detail.
基于距离和颜色数据联合使用的背景估计和去除比单独使用任何一种数据源都能产生更好的结果。随着新型硬件和CPU处理速度的提高,廉价、实时、无源测距系统变得越来越容易使用,这一点越来越重要。距离是一个强大的分割信号,它在很大程度上独立于颜色,因此不受传统的阴影和与背景颜色相似的物体的颜色分割问题的影响。然而,距离本身也不足以实现良好的分割:在场景中几乎无法对所有像素进行深度测量,当前景物体靠近背景时,它们可能在深度上无法区分。在这些情况下,颜色分割是互补的。令人惊讶的是,迄今为止在联合范围和颜色分割方面做的工作很少。我们描述并演示了一种基于每个图像像素的多维(范围和颜色)聚类的背景估计方法。在给定帧中,前景的分割是通过与背景统计的范围和归一化颜色的比较来完成的。详细讨论了诸如阴影处理和低置信度测量等重要的实现问题。
{"title":"Background estimation and removal based on range and color","authors":"G. Gordon, Trevor Darrell, M. Harville, J. Woodfill","doi":"10.1109/CVPR.1999.784721","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784721","url":null,"abstract":"Background estimation and removal based on the joint use of range and color data produces superior results than can be achieved with either data source alone. This is increasingly relevant as inexpensive, real-time, passive range systems become more accessible through novel hardware and increased CPU processing speeds. Range is a powerful signal for segmentation which is largely independent of color and hence not effected by the classic color segmentation problems of shadows and objects with color similar to the background. However range alone is also not sufficient for the good segmentation: depth measurements are rarely available at all pixels in the scene, and foreground objects may be indistinguishable in depth when they are close to the background. Color segmentation is complementary in these cases. Surprisingly, little work has been done to date on joint range and color segmentation. We describe and demonstrate a background estimation method based on a multidimensional (range and color) clustering at each image pixel. Segmentation of the foreground in a given frame is performed via comparison with background statistics in range and normalized color. Important implementation issues such as treatment of shadows and low confidence measurements are discussed in detail.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91275415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 193
Non-metric calibration of wide-angle lenses and polycameras 广角镜头和多镜头的非公制校准
R. Swaminathan, S. Nayar
Images taken with wide-angle cameras tend to have severe distortions which pull points towards the optical center. This paper proposes a method for recovering the distortion parameters without the use of any calibration objects. The distortions cause straight lines in the scene to appear as curves in the image. Our algorithm seeks to find the distortion parameters that would map the image curves to straight lines. The user selects a small set of points along the image curves. Recovery of the parameters is formulated as the minimization of an objective function which is designed to explicitly account for noise in the selected image points. Experimental results are presented for synthetic data with different noise levels as well as for real images. Once calibrated, the image streams from these cameras can be undistorted in real time using look up tables. We also present an application of this calibration method for wide-angle camera clusters, which we call polycameras. We apply our distortion correction technique to a polycamera with four wide-angle cameras to create a high resolution 360 degree panorama in real-time.
用广角相机拍摄的图像往往有严重的扭曲,这将点拉向光学中心。本文提出了一种不使用任何标定对象即可恢复畸变参数的方法。扭曲导致场景中的直线在图像中显示为曲线。我们的算法旨在找到将图像曲线映射为直线的失真参数。用户沿着图像曲线选择一小组点。参数的恢复被表述为目标函数的最小化,该目标函数被设计为明确地考虑所选图像点中的噪声。给出了不同噪声水平的合成数据和真实图像的实验结果。一旦校准,从这些相机流出的图像流可以使用查找表实时地不失真。我们还介绍了这种校准方法在广角相机集群中的应用,我们称之为多镜头相机。我们将畸变校正技术应用于具有四个广角摄像头的多镜头相机,以实时创建高分辨率360度全景。
{"title":"Non-metric calibration of wide-angle lenses and polycameras","authors":"R. Swaminathan, S. Nayar","doi":"10.1109/CVPR.1999.784714","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784714","url":null,"abstract":"Images taken with wide-angle cameras tend to have severe distortions which pull points towards the optical center. This paper proposes a method for recovering the distortion parameters without the use of any calibration objects. The distortions cause straight lines in the scene to appear as curves in the image. Our algorithm seeks to find the distortion parameters that would map the image curves to straight lines. The user selects a small set of points along the image curves. Recovery of the parameters is formulated as the minimization of an objective function which is designed to explicitly account for noise in the selected image points. Experimental results are presented for synthetic data with different noise levels as well as for real images. Once calibrated, the image streams from these cameras can be undistorted in real time using look up tables. We also present an application of this calibration method for wide-angle camera clusters, which we call polycameras. We apply our distortion correction technique to a polycamera with four wide-angle cameras to create a high resolution 360 degree panorama in real-time.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89980184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 278
Estimating model parameters and boundaries by minimizing a joint, robust objective function 通过最小化联合鲁棒目标函数来估计模型参数和边界
C. Stewart, Kishore Bubna, A. Perera
Many problems in computer vision require estimation of both model parameters and boundaries, which limits the usefulness of standard estimation techniques from statistics. Example problems include surface reconstruction from range data, estimation of parametric motion models, fitting circular or elliptic arcs to edgel data, and many others. This paper introduces a new estimation technique, called the "Domain Bounding M-Estimator", which is a generalization of ordinary M-estimators combining error measures on model parameters and boundaries in a joint, robust objective function. Minimization of the objective function given a rough initialization yields simultaneous estimates of parameters and boundaries. The DBM-Estimator has been applied to estimating line segments, surfaces, and the symmetry transformation between two edgel chains. It is unaffected by outliers and prevents boundary estimates from crossing even small magnitude discontinuities.
计算机视觉中的许多问题都需要对模型参数和边界进行估计,这限制了从统计学角度进行标准估计技术的有效性。示例问题包括从距离数据重建表面,估计参数运动模型,拟合圆弧或椭圆弧到边缘数据,以及许多其他问题。本文介绍了一种新的估计技术,称为“域边界m估计器”,它是普通m估计器的推广,在联合鲁棒目标函数中结合模型参数和边界上的误差度量。给定一个粗略初始化的目标函数的最小化产生参数和边界的同时估计。dbm估计器已被应用于估计线段、曲面和两个边链之间的对称变换。它不受异常值的影响,并防止边界估计跨越甚至小幅度的不连续。
{"title":"Estimating model parameters and boundaries by minimizing a joint, robust objective function","authors":"C. Stewart, Kishore Bubna, A. Perera","doi":"10.1109/CVPR.1999.784710","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784710","url":null,"abstract":"Many problems in computer vision require estimation of both model parameters and boundaries, which limits the usefulness of standard estimation techniques from statistics. Example problems include surface reconstruction from range data, estimation of parametric motion models, fitting circular or elliptic arcs to edgel data, and many others. This paper introduces a new estimation technique, called the \"Domain Bounding M-Estimator\", which is a generalization of ordinary M-estimators combining error measures on model parameters and boundaries in a joint, robust objective function. Minimization of the objective function given a rough initialization yields simultaneous estimates of parameters and boundaries. The DBM-Estimator has been applied to estimating line segments, surfaces, and the symmetry transformation between two edgel chains. It is unaffected by outliers and prevents boundary estimates from crossing even small magnitude discontinuities.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89981043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
The discriminatory power of ordinal measures - towards a new coefficient 序数措施的歧视性力量——朝向一个新的系数
S. Scherer, A. Pinz, P. Werth
Perspective distortion, occlusion and specular reflection are challenging problems in shape-from-stereo. In this paper we review one recently published area-based stereo matching algorithm (Bhat and Nayar, 1998) designed to be robust in these cases. Although the algorithm is an important contribution to stereo-matching, we show that its coefficient has a low discriminatory power, which leads to a significant number of multiple best matches. In order to cope with this drawback we introduce a new normalized ordinal correlation coefficient. Experiments showing the behavior of the proposed coefficient are performed on various datasets including real data with ground truth. The new coefficient reduces the occurrence of multiple best matches to almost zero per cent. It also shows a more robust and equally accurate behavior. These benefits are achieved at almost no additional computational costs.
透视失真、遮挡和镜面反射是立体形状中具有挑战性的问题。在本文中,我们回顾了最近发表的一种基于区域的立体匹配算法(Bhat和Nayar, 1998),该算法在这些情况下具有鲁棒性。尽管该算法对立体匹配做出了重要贡献,但我们表明其系数具有较低的区分能力,这导致了大量的多个最佳匹配。为了克服这一缺点,我们引入了一种新的归一化有序相关系数。实验显示了所提出的系数的行为在各种数据集上进行,包括真实数据与地面真值。新系数将多个最佳匹配的发生率降低到几乎为零。它还显示出更强的鲁棒性和同样准确的行为。这些好处几乎不需要额外的计算成本。
{"title":"The discriminatory power of ordinal measures - towards a new coefficient","authors":"S. Scherer, A. Pinz, P. Werth","doi":"10.1109/CVPR.1999.786920","DOIUrl":"https://doi.org/10.1109/CVPR.1999.786920","url":null,"abstract":"Perspective distortion, occlusion and specular reflection are challenging problems in shape-from-stereo. In this paper we review one recently published area-based stereo matching algorithm (Bhat and Nayar, 1998) designed to be robust in these cases. Although the algorithm is an important contribution to stereo-matching, we show that its coefficient has a low discriminatory power, which leads to a significant number of multiple best matches. In order to cope with this drawback we introduce a new normalized ordinal correlation coefficient. Experiments showing the behavior of the proposed coefficient are performed on various datasets including real data with ground truth. The new coefficient reduces the occurrence of multiple best matches to almost zero per cent. It also shows a more robust and equally accurate behavior. These benefits are achieved at almost no additional computational costs.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84421310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Gesture localization and recognition using probabilistic visual learning 基于概率视觉学习的手势定位与识别
Raouf Hamdan, F. Heitz, L. Thoraval
A generic approach for the extraction and recognition of gesture using raw grey-level images is presented. The probabilistic visual learning approach, a learning method recently proposed by B. Moghaddam and A. Pentland (1997), is used to create a set of compact statistical representations of gesture appearance on low dimensional eigenspaces. The same probabilistic modeling framework is used to extract and track gesture and to perform gesture recognition over long image sequences. Gesture extraction and tracking are based on maximum likelihood gesture detection in the input image. Recognition is performed by using the set of learned probabilistic appearance models as estimates of the emission probabilities of a continuous density hidden Markov model (CDHMM). Although the segmentation and CDHMM-based recognition use raw grey-level images, the method is fast, thanks to the data compression obtained by probabilistic visual learning. The approach is comprehensive and may be applied to other visual motion recognition tasks. It does not require application-tailored extraction of image features, the use of markers or gloves. A real-time implementation of the method on a standard PC-based vision system is under consideration.
提出了一种利用原始灰度图像提取和识别手势的通用方法。概率视觉学习方法是最近由B. Moghaddam和a . Pentland(1997)提出的一种学习方法,用于在低维特征空间上创建手势外观的一组紧凑统计表示。使用相同的概率建模框架来提取和跟踪手势,并对长图像序列进行手势识别。手势提取和跟踪是基于输入图像中的最大似然手势检测。通过使用学习到的概率外观模型集作为连续密度隐马尔可夫模型(CDHMM)发射概率的估计来进行识别。虽然分割和基于cdhmm的识别使用原始灰度图像,但由于采用概率视觉学习获得的数据压缩,该方法速度很快。该方法是全面的,可应用于其他视觉运动识别任务。它不需要应用定制的图像特征提取,也不需要使用标记或手套。目前正在考虑在一个标准的基于pc的视觉系统上实时实现该方法。
{"title":"Gesture localization and recognition using probabilistic visual learning","authors":"Raouf Hamdan, F. Heitz, L. Thoraval","doi":"10.1109/CVPR.1999.784615","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784615","url":null,"abstract":"A generic approach for the extraction and recognition of gesture using raw grey-level images is presented. The probabilistic visual learning approach, a learning method recently proposed by B. Moghaddam and A. Pentland (1997), is used to create a set of compact statistical representations of gesture appearance on low dimensional eigenspaces. The same probabilistic modeling framework is used to extract and track gesture and to perform gesture recognition over long image sequences. Gesture extraction and tracking are based on maximum likelihood gesture detection in the input image. Recognition is performed by using the set of learned probabilistic appearance models as estimates of the emission probabilities of a continuous density hidden Markov model (CDHMM). Although the segmentation and CDHMM-based recognition use raw grey-level images, the method is fast, thanks to the data compression obtained by probabilistic visual learning. The approach is comprehensive and may be applied to other visual motion recognition tasks. It does not require application-tailored extraction of image features, the use of markers or gloves. A real-time implementation of the method on a standard PC-based vision system is under consideration.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83253201","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Real-time periodic motion detection, analysis, and applications 实时周期性运动检测,分析和应用
Ross Cutler, L. Davis
We describe a new technique to detect and analyze periodic motion as seen from both a static and moving camera. By tracking objects of interest, we compute an object's self-similarity as it evolves in time. For periodic motion, the self-similarity measure is also periodic, and we apply time-frequency analysis to detect and characterize the periodic motion. A real-time system has been implemented to track and classify objects using periodicity. Examples of object classification, person counting, and non-stationary periodicity are provided.
我们描述了一种检测和分析周期运动的新技术,从静态和运动摄像机的角度来观察。通过跟踪感兴趣的对象,我们计算对象随时间演变的自相似度。对于周期运动,自相似度量也是周期性的,我们采用时频分析来检测和表征周期运动。实现了一种利用周期性对目标进行跟踪和分类的实时系统。提供了对象分类、人员计数和非平稳周期性的例子。
{"title":"Real-time periodic motion detection, analysis, and applications","authors":"Ross Cutler, L. Davis","doi":"10.1109/CVPR.1999.784652","DOIUrl":"https://doi.org/10.1109/CVPR.1999.784652","url":null,"abstract":"We describe a new technique to detect and analyze periodic motion as seen from both a static and moving camera. By tracking objects of interest, we compute an object's self-similarity as it evolves in time. For periodic motion, the self-similarity measure is also periodic, and we apply time-frequency analysis to detect and characterize the periodic motion. A real-time system has been implemented to track and classify objects using periodicity. Examples of object classification, person counting, and non-stationary periodicity are provided.","PeriodicalId":20644,"journal":{"name":"Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"1999-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82993643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
期刊
Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1