首页 > 最新文献

7th International Conference on Automatic Face and Gesture Recognition (FGR06)最新文献

英文 中文
Face recognition by projection-based 3D normalization and shading subspace orthogonalization 基于投影的三维归一化和阴影子空间正交化人脸识别
Tatsuo Kozakaya, Osamu Yamaguchi
This paper describes a new face recognition method using a projection-based 3D normalization and a shading subspace orthogonalization under variation in facial pose and illumination. The proposed method does not need any reconstruction and reillumination for a personalized 3D model, thus it can avoid these troublesome problems and the recognition process can be done rapidly. The facial size and pose including out of plane rotation can be normalized to a generic 3D model from one still image and the input subspace is generated by perturbed cropped patterns in order to absorb the localization errors. Furthermore, by exploiting the fact that a normalized pattern is fitted to the generic 3D model, illumination robust features are extracted through the shading subspace orthogonalization. Evaluation experiments are performed using several databases and the results show the effectiveness of our method under various facial poses and illuminations
提出了一种基于投影的三维归一化和阴影子空间正交化的人脸识别方法。该方法不需要对个性化的三维模型进行任何重建和重新照明,从而避免了这些麻烦的问题,并且可以快速完成识别过程。将人脸的大小和姿态(包括离面旋转)从一张静止图像归一化为一个通用的三维模型,并通过扰动裁剪模式生成输入子空间,以吸收定位误差。此外,利用归一化模式拟合通用三维模型的特点,通过阴影子空间正交化提取光照鲁棒性特征。在多个数据库中进行了评估实验,结果表明了该方法在各种面部姿态和光照下的有效性
{"title":"Face recognition by projection-based 3D normalization and shading subspace orthogonalization","authors":"Tatsuo Kozakaya, Osamu Yamaguchi","doi":"10.1109/FGR.2006.43","DOIUrl":"https://doi.org/10.1109/FGR.2006.43","url":null,"abstract":"This paper describes a new face recognition method using a projection-based 3D normalization and a shading subspace orthogonalization under variation in facial pose and illumination. The proposed method does not need any reconstruction and reillumination for a personalized 3D model, thus it can avoid these troublesome problems and the recognition process can be done rapidly. The facial size and pose including out of plane rotation can be normalized to a generic 3D model from one still image and the input subspace is generated by perturbed cropped patterns in order to absorb the localization errors. Furthermore, by exploiting the fact that a normalized pattern is fitted to the generic 3D model, illumination robust features are extracted through the shading subspace orthogonalization. Evaluation experiments are performed using several databases and the results show the effectiveness of our method under various facial poses and illuminations","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114426664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Joint spatial and frequency domain motion analysis 关节空间和频域运动分析
N. Ahuja, A. Briassouli
Traditionally, motion estimation and segmentation have been performed mostly in the spatial domain, i.e., using the luminance information in the video sequence. Frequency domain representation offers an alternative, rich source of motion information, which has been used to a very limited extent in the past, and on relatively simple problems such as image registration. We review our work during the last few years on an approach to video motion analysis that combines spatial and Fourier domain information. We review our methods for (1) basic (translation and rotation) motion estimation and segmentation, for multiple moving objects, with constant as well as time varying velocities; and (2) more complicated motions, such as periodic motion, and periodic motion superposed on translation. The joint space analysis leads to more compact and computationally efficient solutions than existing techniques
传统上,运动估计和分割主要是在空间域中进行的,即利用视频序列中的亮度信息。频域表示提供了一种替代的、丰富的运动信息源,它在过去被用于非常有限的程度,以及相对简单的问题,如图像配准。我们回顾了我们的工作,在过去几年的视频运动分析的方法,结合空间和傅里叶域信息。我们回顾了我们的方法:(1)基本(平移和旋转)运动估计和分割,对于多个运动物体,具有恒定和时变的速度;(2)更复杂的运动,如周期运动,和周期运动叠加在平移上。与现有技术相比,关节空间分析可以得到更紧凑、计算效率更高的解决方案
{"title":"Joint spatial and frequency domain motion analysis","authors":"N. Ahuja, A. Briassouli","doi":"10.1109/FGR.2006.68","DOIUrl":"https://doi.org/10.1109/FGR.2006.68","url":null,"abstract":"Traditionally, motion estimation and segmentation have been performed mostly in the spatial domain, i.e., using the luminance information in the video sequence. Frequency domain representation offers an alternative, rich source of motion information, which has been used to a very limited extent in the past, and on relatively simple problems such as image registration. We review our work during the last few years on an approach to video motion analysis that combines spatial and Fourier domain information. We review our methods for (1) basic (translation and rotation) motion estimation and segmentation, for multiple moving objects, with constant as well as time varying velocities; and (2) more complicated motions, such as periodic motion, and periodic motion superposed on translation. The joint space analysis leads to more compact and computationally efficient solutions than existing techniques","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126256636","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Weighted Gabor features in unitary space for face recognition 面向人脸识别的酉空间加权Gabor特征
Yong Gao, Yangsheng Wang, Xinshan Zhu, Xuetao Feng, Xiaoxu Zhou
Gabor filters based features, with their good properties of space-frequency localization and orientation selectivity, seem to be the most effective features for face recognition currently. In this paper, we propose a kind of weighted Gabor complex features which combining Gabor magnitude and phase features in unitary space. Its weights are determined according to recognition rates of magnitude and phase features. Meanwhile, subspace based algorithms, PCA and LDA, are generalized into unitary space, and a rarely used distance measure, unitary space cosine distance, is adopted for unitary subspace based recognition algorithms. Using the generalized subspace algorithms our proposed weighted Gabor complex features (WGCF) produce better recognition result than either Gabor magnitude or Gabor phase features. Experiments on FERET database show good results comparable to the best one reported in literature
基于Gabor滤波器的特征具有良好的空频定位和方向选择性,是目前人脸识别中最有效的特征。在酉空间中,我们提出了一种结合Gabor幅度和相位特征的加权Gabor复特征。根据幅值和相位特征的识别率确定其权重。同时,将基于子空间的PCA和LDA算法推广到幺正空间,并在基于幺正子空间的识别算法中采用了很少使用的距离度量幺正空间余弦距离。利用广义子空间算法,我们提出的加权Gabor复特征(WGCF)比Gabor幅度特征和Gabor相位特征具有更好的识别效果。在FERET数据库上的实验结果与文献报道的最佳结果相当
{"title":"Weighted Gabor features in unitary space for face recognition","authors":"Yong Gao, Yangsheng Wang, Xinshan Zhu, Xuetao Feng, Xiaoxu Zhou","doi":"10.1109/FGR.2006.111","DOIUrl":"https://doi.org/10.1109/FGR.2006.111","url":null,"abstract":"Gabor filters based features, with their good properties of space-frequency localization and orientation selectivity, seem to be the most effective features for face recognition currently. In this paper, we propose a kind of weighted Gabor complex features which combining Gabor magnitude and phase features in unitary space. Its weights are determined according to recognition rates of magnitude and phase features. Meanwhile, subspace based algorithms, PCA and LDA, are generalized into unitary space, and a rarely used distance measure, unitary space cosine distance, is adopted for unitary subspace based recognition algorithms. Using the generalized subspace algorithms our proposed weighted Gabor complex features (WGCF) produce better recognition result than either Gabor magnitude or Gabor phase features. Experiments on FERET database show good results comparable to the best one reported in literature","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130994261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Head and facial action tracking: comparison of two robust approaches 头部和面部动作跟踪:两种鲁棒方法的比较
R. Hérault, F. Davoine, Yves Grandvalet
In this work, we address a method that is able to track simultaneously 3D head movements and facial actions like lip and eyebrow movements in a video sequence. In a baseline framework, an adaptive appearance model is estimated online by the knowledge of a monocular video sequence. This method uses a 3D model of the face and a facial adaptive texture model. Then, we consider and compare two improved models in order to increase robustness to occlusions. First, we use robust statistics in order to downweight the hidden regions or outlier pixels. In a second approach, mixture models provides better integration of occlusions. Experiments demonstrate the benefit of the two robust models. The latter are compared under various occlusions
在这项工作中,我们提出了一种能够同时跟踪3D头部运动和面部动作(如视频序列中的嘴唇和眉毛运动)的方法。在基线框架中,根据单目视频序列的知识在线估计自适应外观模型。该方法使用人脸的三维模型和人脸自适应纹理模型。然后,我们考虑并比较了两种改进的模型,以提高对遮挡的鲁棒性。首先,我们使用鲁棒统计来降低隐藏区域或离群像素的权重。在第二种方法中,混合模型提供了更好的闭塞整合。实验证明了这两种鲁棒模型的有效性。后者在不同的咬合下进行比较
{"title":"Head and facial action tracking: comparison of two robust approaches","authors":"R. Hérault, F. Davoine, Yves Grandvalet","doi":"10.1109/FGR.2006.63","DOIUrl":"https://doi.org/10.1109/FGR.2006.63","url":null,"abstract":"In this work, we address a method that is able to track simultaneously 3D head movements and facial actions like lip and eyebrow movements in a video sequence. In a baseline framework, an adaptive appearance model is estimated online by the knowledge of a monocular video sequence. This method uses a 3D model of the face and a facial adaptive texture model. Then, we consider and compare two improved models in order to increase robustness to occlusions. First, we use robust statistics in order to downweight the hidden regions or outlier pixels. In a second approach, mixture models provides better integration of occlusions. Experiments demonstrate the benefit of the two robust models. The latter are compared under various occlusions","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130957840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Face Alignment with Unified Subspace Optimization of Active Statistical Models 主动统计模型统一子空间优化的人脸对齐
Ming Zhao, Tat-Seng Chua
Active statistical models including active shape models and active appearance models are very powerful for face alignment. They are composed of two parts: the subspace model(s) and the search process. While these two parts are closely correlated, existing efforts treated them separately and had not considered how to optimize them overall. Another problem with the subspace model(s) is that the two kinds of parameters of subspaces (the number of components and the constraints on the components) are also treated separately. So they are not jointly optimized. To tackle these two problems, an unified subspace optimization method is proposed. This method is composed of two unification aspects: (I) unification of the statistical model and the search process: the subspace models are optimized according to the search procedure; (2) unification of the number of components and the constraints: the two kinds of parameters are modelled in an unified way, such that they can be optimized jointly. Experimental results demonstrate that our method can effectively find the optimal subspace model and significantly improve the performance
包括主动形状模型和主动外观模型在内的主动统计模型对人脸对齐非常有效。它们由两部分组成:子空间模型和搜索过程。虽然这两个部分是密切相关的,但现有的工作将它们分开处理,没有考虑如何全面优化它们。子空间模型的另一个问题是子空间的两种参数(组件的数量和组件的约束)也被分开处理。所以它们不是联合优化的。针对这两个问题,提出了一种统一的子空间优化方法。该方法由两个统一方面组成:(1)统计模型与搜索过程的统一:根据搜索过程对子空间模型进行优化;(2)构件数量和约束条件的统一:将两类参数统一建模,可以进行联合优化。实验结果表明,该方法能有效地找到最优子空间模型,显著提高了性能
{"title":"Face Alignment with Unified Subspace Optimization of Active Statistical Models","authors":"Ming Zhao, Tat-Seng Chua","doi":"10.1109/FGR.2006.40","DOIUrl":"https://doi.org/10.1109/FGR.2006.40","url":null,"abstract":"Active statistical models including active shape models and active appearance models are very powerful for face alignment. They are composed of two parts: the subspace model(s) and the search process. While these two parts are closely correlated, existing efforts treated them separately and had not considered how to optimize them overall. Another problem with the subspace model(s) is that the two kinds of parameters of subspaces (the number of components and the constraints on the components) are also treated separately. So they are not jointly optimized. To tackle these two problems, an unified subspace optimization method is proposed. This method is composed of two unification aspects: (I) unification of the statistical model and the search process: the subspace models are optimized according to the search procedure; (2) unification of the number of components and the constraints: the two kinds of parameters are modelled in an unified way, such that they can be optimized jointly. Experimental results demonstrate that our method can effectively find the optimal subspace model and significantly improve the performance","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"128 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129625743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Dance posture recognition using wide-baseline orthogonal stereo cameras 基于宽基线正交立体摄像机的舞蹈姿势识别
Feng Guo, G. Qian
In this paper, a robust 3D dance posture recognition system using two cameras is proposed. A pair of wide-baseline video cameras with approximately orthogonal looking directions is used to reduce pose recognition ambiguities. Silhouettes extracted from these two views are represented using Gaussian mixture models (GMM) and used as features for recognition. Relevance vector machine (RVM) is deployed for robust pose recognition. The proposed system is trained using synthesized silhouettes created using animation software and motion capture data. The experimental results on synthetic and real images illustrate that the proposed approach can recognize 3D postures effectively. In addition, the system is easy to set up without any need of precise camera calibration
本文提出了一种基于双摄像头的三维舞蹈姿态识别系统。采用近似正交的双宽基线摄像机来降低姿态识别的模糊性。从这两个视图中提取的轮廓使用高斯混合模型(GMM)表示,并作为识别的特征。采用相关向量机(RVM)进行鲁棒姿态识别。所提出的系统是使用动画软件和动作捕捉数据创建的合成轮廓进行训练的。在合成图像和真实图像上的实验结果表明,该方法可以有效地识别三维姿态。此外,该系统易于设置,无需精确的相机校准
{"title":"Dance posture recognition using wide-baseline orthogonal stereo cameras","authors":"Feng Guo, G. Qian","doi":"10.1109/FGR.2006.35","DOIUrl":"https://doi.org/10.1109/FGR.2006.35","url":null,"abstract":"In this paper, a robust 3D dance posture recognition system using two cameras is proposed. A pair of wide-baseline video cameras with approximately orthogonal looking directions is used to reduce pose recognition ambiguities. Silhouettes extracted from these two views are represented using Gaussian mixture models (GMM) and used as features for recognition. Relevance vector machine (RVM) is deployed for robust pose recognition. The proposed system is trained using synthesized silhouettes created using animation software and motion capture data. The experimental results on synthetic and real images illustrate that the proposed approach can recognize 3D postures effectively. In addition, the system is easy to set up without any need of precise camera calibration","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115023365","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Incremental kernel SVD for face recognition with image sets 基于图像集的增量核奇异值分解人脸识别
Tat-Jun Chin, K. Schindler, D. Suter
Non-linear subspaces derived using kernel methods have been found to be superior compared to linear subspaces in modeling or classification tasks of several visual phenomena. Such kernel methods include kernel PCA, kernel DA, kernel SVD and kernel QR. Since incremental computation algorithms for these methods do not exist yet, the practicality of these methods on large datasets or online video processing is minimal. We propose an approximate incremental kernel SVD algorithm for computer vision applications that require estimation of non-linear subspaces, specifically face recognition by matching image sets obtained through long-term observations or video recordings. We extend a well-known linear subspace updating algorithm to the nonlinear case by utilizing the kernel trick, and apply a reduced set construction method to produce sparse expressions for the derived subspace basis so as to maintain constant processing speed and memory usage. Experimental results demonstrate the effectiveness of the proposed method
利用核方法导出的非线性子空间在若干视觉现象的建模或分类任务中优于线性子空间。这些核方法包括核PCA、核DA、核SVD和核QR。由于这些方法的增量计算算法还不存在,这些方法在大型数据集或在线视频处理上的实用性很小。我们提出了一种近似增量核SVD算法,用于需要估计非线性子空间的计算机视觉应用,特别是通过匹配通过长期观察或视频记录获得的图像集来识别人脸。我们利用核技巧将一种著名的线性子空间更新算法扩展到非线性情况,并应用简化集构造方法对派生的子空间基产生稀疏表达式,以保持恒定的处理速度和内存使用。实验结果证明了该方法的有效性
{"title":"Incremental kernel SVD for face recognition with image sets","authors":"Tat-Jun Chin, K. Schindler, D. Suter","doi":"10.1109/FGR.2006.67","DOIUrl":"https://doi.org/10.1109/FGR.2006.67","url":null,"abstract":"Non-linear subspaces derived using kernel methods have been found to be superior compared to linear subspaces in modeling or classification tasks of several visual phenomena. Such kernel methods include kernel PCA, kernel DA, kernel SVD and kernel QR. Since incremental computation algorithms for these methods do not exist yet, the practicality of these methods on large datasets or online video processing is minimal. We propose an approximate incremental kernel SVD algorithm for computer vision applications that require estimation of non-linear subspaces, specifically face recognition by matching image sets obtained through long-term observations or video recordings. We extend a well-known linear subspace updating algorithm to the nonlinear case by utilizing the kernel trick, and apply a reduced set construction method to produce sparse expressions for the derived subspace basis so as to maintain constant processing speed and memory usage. Experimental results demonstrate the effectiveness of the proposed method","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125083482","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
A new look at filtering techniques for illumination invariance in automatic face recognition 人脸自动识别中光照不变性滤波技术的新研究
Ognjen Arandjelovic, R. Cipolla
Illumination invariance remains the most researched, yet the most challenging aspect of automatic face recognition. In this paper we propose a novel, general recognition framework for efficient matching of individual face images, sets or sequences. The framework is based on simple image processing filters that compete with unprocessed greyscale input to yield a single matching score between individuals. It is shown how the discrepancy between illumination conditions between novel input and the training data set can be estimated and used to weigh the contribution of two competing representations. We describe an extensive empirical evaluation of the proposed method on 171 individuals and over 1300 video sequences with extreme illumination, pose and head motion variation. On this challenging data set our algorithm consistently demonstrated a dramatic performance improvement over traditional filtering approaches. We demonstrate a reduction of 50-75% in recognition error rates, the best performing method-filter combination correctly recognizing 96% of the individuals
光照不变性是人脸自动识别中研究最多,但也是最具挑战性的一个方面。在本文中,我们提出了一种新的,通用的识别框架,用于有效匹配单个人脸图像,集合或序列。该框架基于简单的图像处理过滤器,这些过滤器与未处理的灰度输入竞争,以产生个体之间的单个匹配分数。它展示了如何估计新输入和训练数据集之间的照明条件之间的差异,并用于权衡两个竞争表示的贡献。我们对171个个体和1300多个具有极端照明、姿势和头部运动变化的视频序列进行了广泛的实证评估。在这个具有挑战性的数据集上,我们的算法始终比传统的过滤方法表现出显著的性能改进。我们证明了识别错误率降低了50-75%,表现最好的方法-滤波器组合正确识别了96%的个体
{"title":"A new look at filtering techniques for illumination invariance in automatic face recognition","authors":"Ognjen Arandjelovic, R. Cipolla","doi":"10.1109/FGR.2006.14","DOIUrl":"https://doi.org/10.1109/FGR.2006.14","url":null,"abstract":"Illumination invariance remains the most researched, yet the most challenging aspect of automatic face recognition. In this paper we propose a novel, general recognition framework for efficient matching of individual face images, sets or sequences. The framework is based on simple image processing filters that compete with unprocessed greyscale input to yield a single matching score between individuals. It is shown how the discrepancy between illumination conditions between novel input and the training data set can be estimated and used to weigh the contribution of two competing representations. We describe an extensive empirical evaluation of the proposed method on 171 individuals and over 1300 video sequences with extreme illumination, pose and head motion variation. On this challenging data set our algorithm consistently demonstrated a dramatic performance improvement over traditional filtering approaches. We demonstrate a reduction of 50-75% in recognition error rates, the best performing method-filter combination correctly recognizing 96% of the individuals","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130228942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Accurate Head Pose Tracking in Low Resolution Video 准确的头部姿态跟踪在低分辨率视频
J. Tu, Thomas S. Huang, Hai Tao
Estimating 3D head poses accurately in low resolution video is a challenging vision task because it is difficult to find continuous one-to-one mapping from person-independent low resolution visual representation to head pose parameters. We propose to track head poses by modeling the shape-free facial textures acquired from the video with subspace learning techniques. In particular, we propose to model the facial appearance variations online by incremental weighted PCA subspace with forgetting mechanism, and we do the tracking in an annealed particle filtering framework. Experiments show that, the tracking accuracy of our approach outperforms past visual face tracking algorithms especially in low resolution videos
在低分辨率视频中准确估计3D头部姿态是一项具有挑战性的视觉任务,因为很难找到从独立于人的低分辨率视觉表示到头部姿态参数的连续一对一映射。我们提出利用子空间学习技术对视频中获取的无形状面部纹理进行建模,从而跟踪头部姿态。特别地,我们提出了基于遗忘机制的增量加权PCA子空间在线建模面部外观变化,并在退火粒子滤波框架中进行跟踪。实验表明,该方法的跟踪精度优于以往的视觉人脸跟踪算法,特别是在低分辨率视频中
{"title":"Accurate Head Pose Tracking in Low Resolution Video","authors":"J. Tu, Thomas S. Huang, Hai Tao","doi":"10.1109/FGR.2006.19","DOIUrl":"https://doi.org/10.1109/FGR.2006.19","url":null,"abstract":"Estimating 3D head poses accurately in low resolution video is a challenging vision task because it is difficult to find continuous one-to-one mapping from person-independent low resolution visual representation to head pose parameters. We propose to track head poses by modeling the shape-free facial textures acquired from the video with subspace learning techniques. In particular, we propose to model the facial appearance variations online by incremental weighted PCA subspace with forgetting mechanism, and we do the tracking in an annealed particle filtering framework. Experiments show that, the tracking accuracy of our approach outperforms past visual face tracking algorithms especially in low resolution videos","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129587071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Fast learning for customizable head pose recognition in robotic wheelchair control 机器人轮椅控制中可定制头部姿态识别的快速学习
C. Bauckhage, Thomas Käster, Andrei M. Rotenstein
In the PLAYBOT project, we aim at assisting disabled children at play. To this end, we are developing a semi autonomous robotic wheelchair. It is equipped with several visual sensors and a robotic manipulator and thus conveniently enhances the innate capabilities of a disabled child. In addition to a touch screen, the child may control the wheelchair using simple head movements. As control based on head posture requires reliable face detection and head pose recognition, we are in need of a robust technique that may effortlessly be tailored to individual users. In this paper, we present a multilinear classification algorithm for fast and reliable face detection. It trains within seconds and thus can easily be customized to the home environment of a disabled child. Subsequent head pose recognition is done using support vector machines. Experimental results show that this two stage approach to head pose-based robotic wheelchair control performs fast and very robust
在PLAYBOT项目中,我们的目标是帮助残疾儿童玩耍。为此,我们正在开发一种半自动机器人轮椅。它配备了几个视觉传感器和一个机器人操纵器,从而方便地提高了残疾儿童的先天能力。除了触摸屏外,孩子还可以通过简单的头部运动来控制轮椅。由于基于头部姿势的控制需要可靠的面部检测和头部姿势识别,我们需要一种可以毫不费力地为个人用户量身定制的强大技术。本文提出了一种快速可靠的多线性分类算法。它可以在几秒钟内训练,因此可以很容易地根据残疾儿童的家庭环境进行定制。随后的头部姿势识别使用支持向量机完成。实验结果表明,这种基于头部姿态的轮椅机器人控制方法具有快速、鲁棒性好等优点
{"title":"Fast learning for customizable head pose recognition in robotic wheelchair control","authors":"C. Bauckhage, Thomas Käster, Andrei M. Rotenstein","doi":"10.1109/FGR.2006.52","DOIUrl":"https://doi.org/10.1109/FGR.2006.52","url":null,"abstract":"In the PLAYBOT project, we aim at assisting disabled children at play. To this end, we are developing a semi autonomous robotic wheelchair. It is equipped with several visual sensors and a robotic manipulator and thus conveniently enhances the innate capabilities of a disabled child. In addition to a touch screen, the child may control the wheelchair using simple head movements. As control based on head posture requires reliable face detection and head pose recognition, we are in need of a robust technique that may effortlessly be tailored to individual users. In this paper, we present a multilinear classification algorithm for fast and reliable face detection. It trains within seconds and thus can easily be customized to the home environment of a disabled child. Subsequent head pose recognition is done using support vector machines. Experimental results show that this two stage approach to head pose-based robotic wheelchair control performs fast and very robust","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115855519","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
7th International Conference on Automatic Face and Gesture Recognition (FGR06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1