首页 > 最新文献

7th International Conference on Automatic Face and Gesture Recognition (FGR06)最新文献

英文 中文
Robust face recognition by multiscale kernel associative memory models based on hierarchical spatial-domain Gabor transforms 基于层次空域Gabor变换的多尺度核关联记忆模型鲁棒人脸识别
Bailing Zhang, C. Leung
Face recognition can be considered as a one-class classification problem and associative memory (AM) based approaches have been proven efficient in previous studies. In this paper, a kernel associative memory (KAM) based face recognition scheme with a multiscale Gabor transform, is proposed, in our method, face images of each person are first decomposed into their multiscale representations by a quasi-complete Gabor transform, which are then modelled by kernel associative memories, The pyramidal multi-scale Gabor wavelet transform not only provides a very efficient implementation of Gabor transform in spatial domain, but also permits a fast reconstruction. In the testing phase, a query face image is also represented by a Gabor multiresolution pyramid and the recalled results from different KAM models corresponding to even Gabor channels are then simply added together to provide a reconstruction. The recognition scheme was thoroughly tested using several benchmark face datasets, including the AR faces, UMIST faces, JAFFE faces and Yale A faces. The experiment results have demonstrated strong robustness in recognizing faces under different conditions, particularly the poses alterations, varying occlusions and expression changes
人脸识别可以被认为是一个单类分类问题,基于联想记忆的方法在以往的研究中已经被证明是有效的。本文提出了一种基于核关联记忆(KAM)的多尺度Gabor变换人脸识别方案,该方法首先通过拟完全Gabor变换将人脸图像分解为其多尺度表示,然后利用核关联记忆对其进行建模,金字塔型多尺度Gabor小波变换不仅在空间域上非常有效地实现Gabor变换,而且可以实现快速重构。在测试阶段,查询人脸图像也由Gabor多分辨率金字塔表示,然后简单地将对应于均匀Gabor通道的不同KAM模型的召回结果加在一起以提供重建。使用AR人脸、UMIST人脸、JAFFE人脸和Yale A人脸等多个基准人脸数据集对该识别方案进行了全面测试。实验结果表明,该方法对不同条件下的人脸识别具有较强的鲁棒性,尤其是对姿态变化、不同遮挡和表情变化的人脸识别
{"title":"Robust face recognition by multiscale kernel associative memory models based on hierarchical spatial-domain Gabor transforms","authors":"Bailing Zhang, C. Leung","doi":"10.1109/FGR.2006.95","DOIUrl":"https://doi.org/10.1109/FGR.2006.95","url":null,"abstract":"Face recognition can be considered as a one-class classification problem and associative memory (AM) based approaches have been proven efficient in previous studies. In this paper, a kernel associative memory (KAM) based face recognition scheme with a multiscale Gabor transform, is proposed, in our method, face images of each person are first decomposed into their multiscale representations by a quasi-complete Gabor transform, which are then modelled by kernel associative memories, The pyramidal multi-scale Gabor wavelet transform not only provides a very efficient implementation of Gabor transform in spatial domain, but also permits a fast reconstruction. In the testing phase, a query face image is also represented by a Gabor multiresolution pyramid and the recalled results from different KAM models corresponding to even Gabor channels are then simply added together to provide a reconstruction. The recognition scheme was thoroughly tested using several benchmark face datasets, including the AR faces, UMIST faces, JAFFE faces and Yale A faces. The experiment results have demonstrated strong robustness in recognizing faces under different conditions, particularly the poses alterations, varying occlusions and expression changes","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130940778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A full-body gesture database for automatic gesture recognition 一个用于自动手势识别的全身手势数据库
Bon-Woo Hwang, Sungmin Kim, Seong-Whan Lee
This paper presents a full-body gesture database which contains 2D video data and 3D motion data of 14 normal gestures, 10 abnormal gestures and 30 command gestures for 20 subjects. We call this database the Korea University Gesture (KUG) database. Using 3D motion cameras and 3 sets of stereo cameras, we captured 3D motion data and 3 pairs of stereo-video data at 3 different directions for normal and abnormal gestures. In case of command gestures, 2 pairs of stereo-video data is obtained by 2 sets of stereo cameras with different focal length in order to effectively capture views of whole body and upper body, simultaneously. In addition to these, the 2D silhouette data is synthesized by separating a subject and background in 2D stereo-video data and saved as binary mask images. In this paper, we describe the gesture capture system, the organization of database, the potential usages of the database and the way of obtaining the KUG database
本文提出了一个包含20名受试者的14个正常手势、10个异常手势和30个命令手势的二维视频数据和三维运动数据的全身手势数据库。我们称这个数据库为高丽大学姿态(KUG)数据库。我们使用3D运动摄像机和3套立体摄像机,在3个不同的方向捕捉正常和异常手势的3D运动数据和3对立体视频数据。在指令手势的情况下,2组不同焦距的立体摄像机获得2对立体视频数据,从而有效地同时捕捉全身和上半身的视角。此外,通过分离二维立体视频数据中的主体和背景来合成二维剪影数据,并将其保存为二值掩模图像。本文介绍了手势捕捉系统、数据库的组织、数据库的潜在用途以及KUG数据库的获取方法
{"title":"A full-body gesture database for automatic gesture recognition","authors":"Bon-Woo Hwang, Sungmin Kim, Seong-Whan Lee","doi":"10.1109/FGR.2006.8","DOIUrl":"https://doi.org/10.1109/FGR.2006.8","url":null,"abstract":"This paper presents a full-body gesture database which contains 2D video data and 3D motion data of 14 normal gestures, 10 abnormal gestures and 30 command gestures for 20 subjects. We call this database the Korea University Gesture (KUG) database. Using 3D motion cameras and 3 sets of stereo cameras, we captured 3D motion data and 3 pairs of stereo-video data at 3 different directions for normal and abnormal gestures. In case of command gestures, 2 pairs of stereo-video data is obtained by 2 sets of stereo cameras with different focal length in order to effectively capture views of whole body and upper body, simultaneously. In addition to these, the 2D silhouette data is synthesized by separating a subject and background in 2D stereo-video data and saved as binary mask images. In this paper, we describe the gesture capture system, the organization of database, the potential usages of the database and the way of obtaining the KUG database","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"2008 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125597368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Facial feature detection and tracking with automatic template selection 面部特征检测和跟踪,具有自动模板选择功能
David Cristinacce, Tim Cootes
We describe an accurate and robust method of locating facial features. The method utilises a set of feature templates in conjunction with a shape constrained search technique. The current feature templates are correlated with the target image to generate a set of response surfaces. The parameters of a statistical shape model are optimised to maximise the sum of responses. Given the new feature locations the feature templates are updated using a nearest neighbour approach to select likely feature templates from the training set. We find that this template selection tracker (TST) method outperforms previous approaches using fixed template feature detectors. It gives results similar to the more complex active appearance model (AAM) algorithm on two publicly available static image sets and outperforms the AAM on a more challenging set of in-car face sequences
我们描述了一种精确和鲁棒的面部特征定位方法。该方法将一组特征模板与形状约束搜索技术相结合。将当前特征模板与目标图像关联,生成一组响应面。统计形状模型的参数被优化以最大化响应的总和。给定新的特征位置,使用最近邻方法从训练集中选择可能的特征模板来更新特征模板。我们发现这种模板选择跟踪(TST)方法优于以前使用固定模板特征检测器的方法。它在两个公开可用的静态图像集上给出了与更复杂的主动外观模型(AAM)算法相似的结果,并且在更具挑战性的车内面部序列集上优于AAM
{"title":"Facial feature detection and tracking with automatic template selection","authors":"David Cristinacce, Tim Cootes","doi":"10.1109/FGR.2006.50","DOIUrl":"https://doi.org/10.1109/FGR.2006.50","url":null,"abstract":"We describe an accurate and robust method of locating facial features. The method utilises a set of feature templates in conjunction with a shape constrained search technique. The current feature templates are correlated with the target image to generate a set of response surfaces. The parameters of a statistical shape model are optimised to maximise the sum of responses. Given the new feature locations the feature templates are updated using a nearest neighbour approach to select likely feature templates from the training set. We find that this template selection tracker (TST) method outperforms previous approaches using fixed template feature detectors. It gives results similar to the more complex active appearance model (AAM) algorithm on two publicly available static image sets and outperforms the AAM on a more challenging set of in-car face sequences","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115431907","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 89
Face Recognition from Unconstrained Images: Progress with Prototypes 来自无约束图像的人脸识别:原型的进展
R. Jenkins, A. Burton, D. White
Artificial face recognition systems typically do not attempt to handle very variable images. By comparison, human perceivers can recognize familiar faces over much more varied conditions. We describe a prototype face representation based on simple image-averaging. We have argued that this forms a good candidate for understanding human face perception. Here we examine the stability of these representations by asking (i) how quickly they converge; and (U) how resistant they are to contamination due to previous misidentifications. We conclude that face averages provide promising representations for use in artificial recognition
人工面部识别系统通常不会尝试处理非常多变的图像。相比之下,人类感知者可以在更多不同的条件下识别熟悉的面孔。我们描述了一个基于简单图像平均的原型人脸表示。我们认为,这形成了理解人类面部感知的一个很好的候选。在这里,我们通过询问(i)它们收敛的速度有多快来检验这些表示的稳定性;(U)由于之前的错误识别,它们对污染的抵抗力如何。我们得出结论,面部平均值为人工识别提供了有前途的表征
{"title":"Face Recognition from Unconstrained Images: Progress with Prototypes","authors":"R. Jenkins, A. Burton, D. White","doi":"10.1109/FGR.2006.45","DOIUrl":"https://doi.org/10.1109/FGR.2006.45","url":null,"abstract":"Artificial face recognition systems typically do not attempt to handle very variable images. By comparison, human perceivers can recognize familiar faces over much more varied conditions. We describe a prototype face representation based on simple image-averaging. We have argued that this forms a good candidate for understanding human face perception. Here we examine the stability of these representations by asking (i) how quickly they converge; and (U) how resistant they are to contamination due to previous misidentifications. We conclude that face averages provide promising representations for use in artificial recognition","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129208528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
Comparison of silhouette shape descriptors for example-based human pose recovery 基于实例的人体姿态恢复中轮廓形状描述符的比较
R. Poppe, M. Poel
Automatically recovering human poses from visual input is useful but challenging due to variations in image space and the high dimensionality of the pose space. In this paper, we assume that a human silhouette can be extracted from monocular visual input. We compare three shape descriptors that are used in the encoding of silhouettes: Fourier descriptors, shape contexts and Hu moments. An example-based approach is taken to recover upper body poses from these descriptors. We perform experiments with deformed silhouettes to test each descriptor's robustness against variations in body dimensions, viewpoint and noise. It is shown that Fourier descriptors and shape context histograms outperform Hu moments for all deformations
从视觉输入中自动恢复人体姿势是有用的,但由于图像空间的变化和姿势空间的高维性,因此具有挑战性。在本文中,我们假设可以从单目视觉输入中提取人体轮廓。我们比较了在轮廓编码中使用的三种形状描述子:傅里叶描述子、形状上下文和胡矩。采用基于实例的方法从这些描述符中恢复上半身姿势。我们对变形的轮廓进行实验,以测试每个描述符对身体尺寸、视点和噪声变化的鲁棒性。结果表明,傅里叶描述子和形状上下文直方图在所有变形中都优于Hu矩
{"title":"Comparison of silhouette shape descriptors for example-based human pose recovery","authors":"R. Poppe, M. Poel","doi":"10.1109/FGR.2006.32","DOIUrl":"https://doi.org/10.1109/FGR.2006.32","url":null,"abstract":"Automatically recovering human poses from visual input is useful but challenging due to variations in image space and the high dimensionality of the pose space. In this paper, we assume that a human silhouette can be extracted from monocular visual input. We compare three shape descriptors that are used in the encoding of silhouettes: Fourier descriptors, shape contexts and Hu moments. An example-based approach is taken to recover upper body poses from these descriptors. We perform experiments with deformed silhouettes to test each descriptor's robustness against variations in body dimensions, viewpoint and noise. It is shown that Fourier descriptors and shape context histograms outperform Hu moments for all deformations","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127890458","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 63
Optimised landmark model matching for face recognition 人脸识别的地标模型匹配优化
R. Senaratne, S. Halgamuge
A new method for face recognition, landmark model matching, is proposed in this paper. It is based on the concepts of elastic bunch graph matching and active shape model, and optimised with particle swarm optimisation. It is a fully automatic algorithm and can be used for face databases where only one image per person is available. A face is represented by a landmark model consisting of nodes labelled with jets and gray-level profiles. A landmark distribution model is created from a few training images. The model similarity between the landmark distribution model and the deformable landmark model that has to be fitted to the face in the image is maximised by particle swarm optimisation, to find the optimal model to represent the face. Improved results were obtained by this method compared with elastic bunch graph matching without optimisation
本文提出了一种新的人脸识别方法——地标模型匹配。该算法基于弹性束图匹配和主动形状模型的概念,并采用粒子群算法进行优化。这是一种全自动算法,可以用于每个人只有一张图像的人脸数据库。人脸由一个地标模型表示,该模型由带有喷流和灰阶轮廓的节点组成。从少量训练图像中创建地标分布模型。通过粒子群优化,将图像中需要拟合人脸的地标分布模型与可变形地标模型之间的模型相似性最大化,找到最优的人脸模型。与未经优化的弹性束图匹配相比,该方法得到了改进的结果
{"title":"Optimised landmark model matching for face recognition","authors":"R. Senaratne, S. Halgamuge","doi":"10.1109/FGR.2006.85","DOIUrl":"https://doi.org/10.1109/FGR.2006.85","url":null,"abstract":"A new method for face recognition, landmark model matching, is proposed in this paper. It is based on the concepts of elastic bunch graph matching and active shape model, and optimised with particle swarm optimisation. It is a fully automatic algorithm and can be used for face databases where only one image per person is available. A face is represented by a landmark model consisting of nodes labelled with jets and gray-level profiles. A landmark distribution model is created from a few training images. The model similarity between the landmark distribution model and the deformable landmark model that has to be fitted to the face in the image is maximised by particle swarm optimisation, to find the optimal model to represent the face. Improved results were obtained by this method compared with elastic bunch graph matching without optimisation","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126408187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Subspace-based age-group classification using facial images under various lighting conditions 基于子空间的不同光照条件下面部图像年龄组分类
K. Ueki, T. Hayashida, Tetsunori Kobayashi
This paper presents a framework of age-group classification using facial images under various lighting conditions. Our method is based on the appearance-based approach that projects images from the original image space into a face-subspace. We propose a two-phased approach (2DLDA+LDA), which is based on 2DPCA and LDA. Our experimental results show that the new 2DLDA+LDA-based approach improves classification accuracy more than the conventional PCA-based and LDA-based approach. Moreover, the effectiveness of eliminating dimensions that do not contain important discriminative information is confirmed. The accuracy rates are 46.3%, 67.8% and 78.1% for age-groups that are in the 5-year, 10-year and 15-year range respectively
本文提出了一种基于不同光照条件下面部图像的年龄组分类框架。我们的方法是基于基于外观的方法,将图像从原始图像空间投影到人脸子空间。我们提出了一种基于2DPCA和LDA的两阶段方法(2DLDA+LDA)。实验结果表明,基于2DLDA+ lda的分类方法比传统的基于pca和lda的分类方法提高了分类精度。此外,还证实了消除不包含重要判别信息的维度的有效性。5岁、10岁和15岁年龄组的准确率分别为46.3%、67.8%和78.1%
{"title":"Subspace-based age-group classification using facial images under various lighting conditions","authors":"K. Ueki, T. Hayashida, Tetsunori Kobayashi","doi":"10.1109/FGR.2006.102","DOIUrl":"https://doi.org/10.1109/FGR.2006.102","url":null,"abstract":"This paper presents a framework of age-group classification using facial images under various lighting conditions. Our method is based on the appearance-based approach that projects images from the original image space into a face-subspace. We propose a two-phased approach (2DLDA+LDA), which is based on 2DPCA and LDA. Our experimental results show that the new 2DLDA+LDA-based approach improves classification accuracy more than the conventional PCA-based and LDA-based approach. Moreover, the effectiveness of eliminating dimensions that do not contain important discriminative information is confirmed. The accuracy rates are 46.3%, 67.8% and 78.1% for age-groups that are in the 5-year, 10-year and 15-year range respectively","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126255139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 106
Regression and classification approaches to eye localization in face images 人脸图像中眼睛定位的回归与分类方法
M. Everingham, Andrew Zisserman
We address the task of accurately localizing the eyes in face images extracted by a face detector, an important problem to be solved because of the negative effect of poor localization on face recognition accuracy. We investigate three approaches to the task: a regression approach aiming to directly minimize errors in the predicted eye positions, a simple Bayesian model of eye and non-eye appearance, and a discriminative eye detector trained using AdaBoost. By using identical training and test data for each method we are able to perform an unbiased comparison. We show that, perhaps surprisingly, the simple Bayesian approach performs best on databases including challenging images, and performance is comparable to more complex state-of-the-art methods
我们解决了人脸检测器提取的人脸图像中眼睛的准确定位问题,这是一个重要的问题,因为定位不良会对人脸识别精度产生负面影响。我们研究了三种方法:旨在直接减少预测眼睛位置误差的回归方法,眼睛和非眼睛外观的简单贝叶斯模型,以及使用AdaBoost训练的判别眼睛检测器。通过对每种方法使用相同的训练和测试数据,我们能够进行无偏比较。我们表明,也许令人惊讶的是,简单的贝叶斯方法在包括具有挑战性的图像的数据库上表现最好,并且性能与更复杂的最先进的方法相当
{"title":"Regression and classification approaches to eye localization in face images","authors":"M. Everingham, Andrew Zisserman","doi":"10.1109/FGR.2006.90","DOIUrl":"https://doi.org/10.1109/FGR.2006.90","url":null,"abstract":"We address the task of accurately localizing the eyes in face images extracted by a face detector, an important problem to be solved because of the negative effect of poor localization on face recognition accuracy. We investigate three approaches to the task: a regression approach aiming to directly minimize errors in the predicted eye positions, a simple Bayesian model of eye and non-eye appearance, and a discriminative eye detector trained using AdaBoost. By using identical training and test data for each method we are able to perform an unbiased comparison. We show that, perhaps surprisingly, the simple Bayesian approach performs best on databases including challenging images, and performance is comparable to more complex state-of-the-art methods","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126028870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 119
3D gait recognition using multiple cameras 使用多个摄像头的3D步态识别
Guoying Zhao, Guoyi Liu, Hua Li, M. Pietikäinen
Gait recognition is used to identify individuals in image sequences by the way they walk. Nearly all of the approaches proposed for gait recognition are 2D methods based on analyzing image sequences captured by a single camera. In this paper, video sequences captured by multiple cameras are used as input, and then a human 3D model is set up. The motion is tracked by applying a local optimization algorithm. The lengths of key segments are extracted as static parameters, and the motion trajectories of lower limbs are used as dynamic features. Finally, linear time normalization is exploited for matching and recognition. The proposed method based on 3D tracking and recognition is robust to the changes of viewpoints. Moreover, better results are achieved for sequences containing difficult surface variations than with 2D methods, which prove the efficiency of our algorithm
步态识别是通过人走路的方式来识别图像序列中的个体。几乎所有提出的步态识别方法都是基于分析单个摄像机捕获的图像序列的二维方法。本文以多台摄像机采集的视频序列作为输入,建立人体三维模型。采用局部优化算法对运动进行跟踪。提取关键段长度作为静态参数,下肢运动轨迹作为动态特征。最后,利用线性时间归一化进行匹配和识别。该方法基于三维跟踪和识别,对视点变化具有鲁棒性。此外,对于包含复杂表面变化的序列,其结果优于二维方法,证明了算法的有效性
{"title":"3D gait recognition using multiple cameras","authors":"Guoying Zhao, Guoyi Liu, Hua Li, M. Pietikäinen","doi":"10.1109/FGR.2006.2","DOIUrl":"https://doi.org/10.1109/FGR.2006.2","url":null,"abstract":"Gait recognition is used to identify individuals in image sequences by the way they walk. Nearly all of the approaches proposed for gait recognition are 2D methods based on analyzing image sequences captured by a single camera. In this paper, video sequences captured by multiple cameras are used as input, and then a human 3D model is set up. The motion is tracked by applying a local optimization algorithm. The lengths of key segments are extracted as static parameters, and the motion trajectories of lower limbs are used as dynamic features. Finally, linear time normalization is exploited for matching and recognition. The proposed method based on 3D tracking and recognition is robust to the changes of viewpoints. Moreover, better results are achieved for sequences containing difficult surface variations than with 2D methods, which prove the efficiency of our algorithm","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127380104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 204
Local versus global segmentation for facial expression recognition 面部表情识别的局部分割与全局分割
J. Whitehill, C. Omlin
We examined the open issue of whether FACS action units (AUs) can be recognized more accurately by classifying local regions around the eyes, brows, and mouth compared to analyzing the face as a whole. Our empirical results showed that, contrary to our intuition, local expression analysis showed no consistent improvement in recognition accuracy. Moreover, global analysis outperformed local analysis on certain AUs of the eye and brow regions. We attributed this unexpected result partly to high correlations between different AUs in the Cohn-Kanade expression database. This underlines the importance of establishing a large, publicly available AU database with singly-occurring AUs to facilitate future research
我们研究了一个悬而未决的问题,即通过对眼睛、眉毛和嘴巴周围的局部区域进行分类,与分析整个面部相比,是否可以更准确地识别FACS动作单元(AUs)。我们的实证结果表明,与我们的直觉相反,局部表达分析在识别精度上没有一致的提高。此外,在眼睛和眉毛区域的某些au上,全局分析优于局部分析。我们将这一意想不到的结果部分归因于Cohn-Kanade表达数据库中不同au之间的高度相关性。这强调了建立一个大型的、公开可用的、包含单个发生的AU的AU数据库的重要性,以促进未来的研究
{"title":"Local versus global segmentation for facial expression recognition","authors":"J. Whitehill, C. Omlin","doi":"10.1109/FGR.2006.74","DOIUrl":"https://doi.org/10.1109/FGR.2006.74","url":null,"abstract":"We examined the open issue of whether FACS action units (AUs) can be recognized more accurately by classifying local regions around the eyes, brows, and mouth compared to analyzing the face as a whole. Our empirical results showed that, contrary to our intuition, local expression analysis showed no consistent improvement in recognition accuracy. Moreover, global analysis outperformed local analysis on certain AUs of the eye and brow regions. We attributed this unexpected result partly to high correlations between different AUs in the Cohn-Kanade expression database. This underlines the importance of establishing a large, publicly available AU database with singly-occurring AUs to facilitate future research","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130462364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
7th International Conference on Automatic Face and Gesture Recognition (FGR06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1