首页 > 最新文献

Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)最新文献

英文 中文
Recognizing lower face action units for facial expression analysis 识别面部下部动作单元进行面部表情分析
Ying-li Tian, T. Kanade, J. Cohn
Most automatic expression analysis systems attempt to recognize a small set of prototypic expressions (e.g., happiness and anger). Such prototypic expressions, however, occur infrequently. Human emotions and intentions are communicated more often by changes in one or two discrete facial features. We develop an automatic system to analyze subtle changes in facial expressions based on both permanent (e.g., mouth, eye, and brow) and transient (e.g., furrows and wrinkles) facial features in a nearly frontal image sequence. Multi-state facial component models are proposed for tracking and modeling different facial features. Based on these multi-state models, and without artificial enhancement, we detect and track the facial features, including mouth, eyes, brow, cheeks, and their related wrinkles and facial furrows. Moreover we recover detailed parametric descriptions of the facial features. With these features as the inputs, 11 individual action units or action unit combinations are recognized by a neural network algorithm. A recognition rate of 96.7% is obtained. The recognition results indicate that our system can identify action units regardless of whether they occur singly or in combinations.
大多数自动表情分析系统都试图识别一小部分原型表情(例如,快乐和愤怒)。然而,这种原型表达很少出现。人类的情感和意图往往是通过一两个离散的面部特征的变化来传达的。我们开发了一个自动系统来分析面部表情的细微变化,该变化基于近正面图像序列中的永久(例如,嘴,眼睛和眉毛)和短暂(例如,皱纹和皱纹)面部特征。针对不同的人脸特征,提出了多状态人脸成分模型进行跟踪和建模。基于这些多状态模型,无需人工增强,我们检测和跟踪面部特征,包括嘴,眼睛,眉毛,脸颊及其相关的皱纹和面部皱纹。此外,我们还恢复了面部特征的详细参数描述。以这些特征作为输入,通过神经网络算法识别11个单独的动作单元或动作单元组合。识别率为96.7%。识别结果表明,我们的系统可以识别动作单元,无论它们是单独发生还是组合发生。
{"title":"Recognizing lower face action units for facial expression analysis","authors":"Ying-li Tian, T. Kanade, J. Cohn","doi":"10.1109/AFGR.2000.840678","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840678","url":null,"abstract":"Most automatic expression analysis systems attempt to recognize a small set of prototypic expressions (e.g., happiness and anger). Such prototypic expressions, however, occur infrequently. Human emotions and intentions are communicated more often by changes in one or two discrete facial features. We develop an automatic system to analyze subtle changes in facial expressions based on both permanent (e.g., mouth, eye, and brow) and transient (e.g., furrows and wrinkles) facial features in a nearly frontal image sequence. Multi-state facial component models are proposed for tracking and modeling different facial features. Based on these multi-state models, and without artificial enhancement, we detect and track the facial features, including mouth, eyes, brow, cheeks, and their related wrinkles and facial furrows. Moreover we recover detailed parametric descriptions of the facial features. With these features as the inputs, 11 individual action units or action unit combinations are recognized by a neural network algorithm. A recognition rate of 96.7% is obtained. The recognition results indicate that our system can identify action units regardless of whether they occur singly or in combinations.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134198987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 115
Learning and synthesizing human body motion and posture 学习和综合人体动作和姿势
Rómer Rosales, S. Sclaroff
A novel approach is presented for estimating human body posture and motion from a video sequence. Human pose is defined as the instantaneous image plane configuration of a single articulated body in terms of the position of a predetermined set of joints. First, statistical segmentation of the human bodies from the background is performed and low-level visual features are found given the segmented body shape. The goal is to be able to map these visual features to body configurations. Given a set of body motion sequences for training, a set of clusters is built in which each has statistically similar configurations. This unsupervised task is done using the expectation maximization algorithm. Then, for each of the clusters, a neural network is trained to build this mapping. Clustering body configurations improves the mapping accuracy. Given new visual features, a mapping from each cluster is performed providing a set of possible poses. From this set, the most likely pose is extracted given the learned probability distribution and the visual feature similarity between hypothesis and input. Performance of the system is characterized using a new set of known body postures, showing promising results.
提出了一种从视频序列中估计人体姿态和运动的新方法。人体姿态被定义为单个关节体根据一组预定关节的位置的瞬时图像平面构型。首先,从背景中对人体进行统计分割,根据分割后的人体形状找到低层次的视觉特征;目标是能够将这些视觉特征映射到身体结构上。给定一组用于训练的身体运动序列,构建一组簇,其中每个簇具有统计上相似的配置。这个无监督任务是使用期望最大化算法完成的。然后,对于每个簇,训练一个神经网络来构建这个映射。聚类体配置提高了映射精度。给定新的视觉特征,从每个集群执行映射,提供一组可能的姿势。从这个集合中,根据学习到的概率分布和假设与输入之间的视觉特征相似性,提取出最可能的姿态。使用一组新的已知身体姿势来表征系统的性能,显示出有希望的结果。
{"title":"Learning and synthesizing human body motion and posture","authors":"Rómer Rosales, S. Sclaroff","doi":"10.1109/AFGR.2000.840681","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840681","url":null,"abstract":"A novel approach is presented for estimating human body posture and motion from a video sequence. Human pose is defined as the instantaneous image plane configuration of a single articulated body in terms of the position of a predetermined set of joints. First, statistical segmentation of the human bodies from the background is performed and low-level visual features are found given the segmented body shape. The goal is to be able to map these visual features to body configurations. Given a set of body motion sequences for training, a set of clusters is built in which each has statistically similar configurations. This unsupervised task is done using the expectation maximization algorithm. Then, for each of the clusters, a neural network is trained to build this mapping. Clustering body configurations improves the mapping accuracy. Given new visual features, a mapping from each cluster is performed providing a set of possible poses. From this set, the most likely pose is extracted given the learned probability distribution and the visual feature similarity between hypothesis and input. Performance of the system is characterized using a new set of known body postures, showing promising results.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131607756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Gender classification with support vector machines 基于支持向量机的性别分类
B. Moghaddam, Ming-Hsuan Yang
Support vector machines (SVM) are investigated for visual gender classification with low-resolution "thumbnail" faces (21-by-12 pixels) processed from 1755 images from the FERET face database. The performance of SVM (3.4% error) is shown to be superior to traditional pattern classifiers (linear, quadratic, Fisher linear discriminant, nearest-neighbor) as well as more modern techniques such as radial basis function (RBF) classifiers and large ensemble-RBF networks. SVM also out-performed human test subjects at the same task: in a perception study with 30 human test subjects, ranging in age from mid-20s to mid-40s, the average error rate was found to be 32% for the "thumbnails" and 6.7% with higher resolution images. The difference in performance between low- and high-resolution tests with SVM was only 1%, demonstrating robustness and relative scale invariance for visual classification.
利用FERET人脸数据库中的1755张低分辨率“缩略图”人脸(21 × 12像素),研究了支持向量机(SVM)的视觉性别分类。支持向量机的性能(误差3.4%)优于传统的模式分类器(线性,二次,Fisher线性判别,最近邻)以及更现代的技术,如径向基函数(RBF)分类器和大型集成-RBF网络。SVM在同样的任务上也优于人类测试对象:在一项由30名年龄在25岁到40岁之间的人类测试对象组成的感知研究中,发现“缩略图”的平均错误率为32%,而更高分辨率图像的平均错误率为6.7%。支持向量机的低分辨率和高分辨率测试之间的性能差异仅为1%,证明了视觉分类的鲁棒性和相对尺度不变性。
{"title":"Gender classification with support vector machines","authors":"B. Moghaddam, Ming-Hsuan Yang","doi":"10.1109/AFGR.2000.840651","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840651","url":null,"abstract":"Support vector machines (SVM) are investigated for visual gender classification with low-resolution \"thumbnail\" faces (21-by-12 pixels) processed from 1755 images from the FERET face database. The performance of SVM (3.4% error) is shown to be superior to traditional pattern classifiers (linear, quadratic, Fisher linear discriminant, nearest-neighbor) as well as more modern techniques such as radial basis function (RBF) classifiers and large ensemble-RBF networks. SVM also out-performed human test subjects at the same task: in a perception study with 30 human test subjects, ranging in age from mid-20s to mid-40s, the average error rate was found to be 32% for the \"thumbnails\" and 6.7% with higher resolution images. The difference in performance between low- and high-resolution tests with SVM was only 1%, demonstrating robustness and relative scale invariance for visual classification.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124286182","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 330
A robust model-based approach for 3D head tracking in video sequences 一种鲁棒的基于模型的视频序列三维头部跟踪方法
M. Malciu, F. Prêteux
We present a generic and robust method for model-based global 3D head pose estimation in monocular and non-calibrated video sequences. The proposed method relies on a 3D/2D matching between 2D image features estimated throughout the sequence and 3D object features of a generic head model. Specifically, it combines motion and texture features in an iterative optimization procedure based on the downhill simplex algorithm. A proper initialization of the pose parameters, based on a block matching procedure, is performed at each frame in order to take into account large amplitude motions. For the same reason, we have developed a nonlinear optical flow-based interpolation algorithm for increasing the frame rate. Experiments demonstrate that this method is stable over extended sequences including large head motions, occlusions, various head postures and lighting variations. The estimation accuracy is related to the head model, as established by using an ellipsoidal model and an ad hoc synthesized model. The proposed method is general enough to be applied to other tracking applications.
我们提出了一种通用的、鲁棒的方法,用于单目和非校准视频序列中基于模型的全局3D头部姿态估计。所提出的方法依赖于整个序列中估计的2D图像特征与通用头部模型的3D对象特征之间的3D/2D匹配。具体来说,它结合了运动和纹理特征的迭代优化过程,基于下坡单纯形算法。基于块匹配过程,在每一帧执行适当的姿态参数初始化,以便考虑大振幅运动。出于同样的原因,我们开发了一种基于非线性光流的插值算法来提高帧率。实验表明,该方法在包括大的头部运动、遮挡、各种头部姿势和光照变化在内的扩展序列中是稳定的。估计精度与头部模型有关,分别采用椭球体模型和自适应综合模型建立了头部模型。该方法具有一定的通用性,可以应用于其他跟踪应用。
{"title":"A robust model-based approach for 3D head tracking in video sequences","authors":"M. Malciu, F. Prêteux","doi":"10.1109/AFGR.2000.840630","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840630","url":null,"abstract":"We present a generic and robust method for model-based global 3D head pose estimation in monocular and non-calibrated video sequences. The proposed method relies on a 3D/2D matching between 2D image features estimated throughout the sequence and 3D object features of a generic head model. Specifically, it combines motion and texture features in an iterative optimization procedure based on the downhill simplex algorithm. A proper initialization of the pose parameters, based on a block matching procedure, is performed at each frame in order to take into account large amplitude motions. For the same reason, we have developed a nonlinear optical flow-based interpolation algorithm for increasing the frame rate. Experiments demonstrate that this method is stable over extended sequences including large head motions, occlusions, various head postures and lighting variations. The estimation accuracy is related to the head model, as established by using an ellipsoidal model and an ad hoc synthesized model. The proposed method is general enough to be applied to other tracking applications.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114695755","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Extraction of parametric human model for posture recognition using genetic algorithm 基于遗传算法的人体姿态识别参数模型提取
Changbo Hu, Q. Yu, Yi Li, Songde Ma
We present in this paper an approach to extracting a human parametric 2D model for the purpose of estimating human posture and recognizing human activity. This task is done in two steps. In the first step, a human silhouette is extracted from a complex background under a fixed camera through a statistical method. By this method, we can reconstruct the background dynamically and obtain the moving silhouette. In the second step, a genetic algorithm is used to match the silhouette of the human body to a model in parametric shape space. In order to reduce the searching dimension, a layer method is proposed to take the advantage of the human model. Additionally we apply a structure-oriented Kalman filter to estimate the motion of body parts. Therefore the initial population and value in the GA can be well constrained. Experiments on real video sequences show that our method can extract the human model robustly and accurately.
本文提出了一种提取人体参数二维模型的方法,用于估计人体姿态和识别人体活动。这个任务分两个步骤完成。第一步,在固定摄像机下,通过统计方法从复杂背景中提取人体轮廓。通过该方法,可以动态地重建背景,得到运动轮廓。第二步,利用遗传算法将人体轮廓与参数化形状空间中的模型进行匹配。为了降低搜索维数,提出了一种利用人体模型的分层方法。此外,我们还应用了面向结构的卡尔曼滤波来估计身体部位的运动。因此,遗传算法中的初始种群和初始值可以得到很好的约束。在真实视频序列上的实验表明,该方法可以鲁棒、准确地提取人体模型。
{"title":"Extraction of parametric human model for posture recognition using genetic algorithm","authors":"Changbo Hu, Q. Yu, Yi Li, Songde Ma","doi":"10.1109/AFGR.2000.840683","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840683","url":null,"abstract":"We present in this paper an approach to extracting a human parametric 2D model for the purpose of estimating human posture and recognizing human activity. This task is done in two steps. In the first step, a human silhouette is extracted from a complex background under a fixed camera through a statistical method. By this method, we can reconstruct the background dynamically and obtain the moving silhouette. In the second step, a genetic algorithm is used to match the silhouette of the human body to a model in parametric shape space. In order to reduce the searching dimension, a layer method is proposed to take the advantage of the human model. Additionally we apply a structure-oriented Kalman filter to estimate the motion of body parts. Therefore the initial population and value in the GA can be well constrained. Experiments on real video sequences show that our method can extract the human model robustly and accurately.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116897548","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 53
Face recognition based on multiple facial features 基于多种面部特征的人脸识别
Rui Liao, S. Li
An automatic face recognition system based on multiple facial features is described. Each facial feature is represented by a Gabor-based complex vector and is localized by an automatic facial feature detection scheme. Two face recognition approaches, named two-layer nearest neighbor (TLNN) and modular nearest feature line (MNFL) respectively, are proposed. Both TLNN and MNFL are based on the multiple facial features detected for each image and their superiority in face recognition is demonstrated.
介绍了一种基于多种面部特征的自动人脸识别系统。每个面部特征由基于gabor的复向量表示,并通过自动面部特征检测方案进行定位。提出了两种人脸识别方法,分别称为两层最近邻(TLNN)和模块化最近邻特征线(MNFL)。TLNN和MNFL都是基于对每张图像检测到的多个面部特征,并证明了它们在人脸识别方面的优越性。
{"title":"Face recognition based on multiple facial features","authors":"Rui Liao, S. Li","doi":"10.1109/AFGR.2000.840641","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840641","url":null,"abstract":"An automatic face recognition system based on multiple facial features is described. Each facial feature is represented by a Gabor-based complex vector and is localized by an automatic facial feature detection scheme. Two face recognition approaches, named two-layer nearest neighbor (TLNN) and modular nearest feature line (MNFL) respectively, are proposed. Both TLNN and MNFL are based on the multiple facial features detected for each image and their superiority in face recognition is demonstrated.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125002757","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Face recognition in a meeting room 会议室的人脸识别
R. Gross, Jie Yang, A. Waibel
We investigate the recognition of human faces in a meeting room. The major challenges of identifying human faces in this environment include low quality of input images, poor illumination, unrestricted head poses and continuously changing facial expressions and occlusion. In order to address these problems we propose a novel algorithm, dynamic space warping (DSW). The basic idea of the algorithm is to combine local features under certain spatial constraints. We compare DSW with the eigenface approach on data collected from various meetings. We have tested both front and profile face images and images with two stages of occlusion. The experimental results indicate that the DSW approach outperforms the eigenface approach in both cases.
我们研究了在会议室中对人脸的识别。在这种环境下识别人脸的主要挑战包括输入图像质量低、光照差、不受限制的头部姿势以及不断变化的面部表情和遮挡。为了解决这些问题,我们提出了一种新的算法——动态空间翘曲(DSW)。该算法的基本思想是在一定的空间约束下对局部特征进行组合。我们比较DSW与特征面方法从各种会议收集的数据。我们测试了正面和侧面的人脸图像以及两个阶段遮挡的图像。实验结果表明,DSW方法在两种情况下都优于特征面方法。
{"title":"Face recognition in a meeting room","authors":"R. Gross, Jie Yang, A. Waibel","doi":"10.1109/AFGR.2000.840649","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840649","url":null,"abstract":"We investigate the recognition of human faces in a meeting room. The major challenges of identifying human faces in this environment include low quality of input images, poor illumination, unrestricted head poses and continuously changing facial expressions and occlusion. In order to address these problems we propose a novel algorithm, dynamic space warping (DSW). The basic idea of the algorithm is to combine local features under certain spatial constraints. We compare DSW with the eigenface approach on data collected from various meetings. We have tested both front and profile face images and images with two stages of occlusion. The experimental results indicate that the DSW approach outperforms the eigenface approach in both cases.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129739961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Focus of attention for face and hand gesture recognition using multiple cameras 使用多个摄像头进行面部和手势识别的注意力焦点
H. Hongo, M. Yasumoto, Y. Niwa, M. Ohya, Kazuhiko Yamamoto
We propose a multi-camera system that can track multiple human faces and hands as well as focus on face and hand gestures for recognition. Our current system consists of four cameras. Two fixed cameras are used as a stereo system to estimate face and hand positions. The stereo camera detects faces and hands by a standard skin color method we propose. The distances of the targets are then estimated. Next to track multiple targets, we estimate the positions and sizes of targets between consecutive frames. The other two cameras perform tracking of such targets as faces and hands. If a target is not the appropriate size for recognition, the tracking cameras acquire its zoomed image. Since our system has two tracking cameras, it can track two targets at the same time. To recognize faces and hand gestures, we propose four directional features by using linear discriminant analysis. Using our system, we experimented on human position estimation, multiple face tracking, and face and hand gesture recognition. These experiments showed that our system could estimate human position with the stereo camera and track multiple targets by using target positions and sizes even if the persons overlapped with each other. In addition, our system could recognize faces and hand gestures by using the four directional features.
我们提出了一种多摄像头系统,可以跟踪多个人脸和手,并专注于人脸和手势进行识别。我们目前的系统由四个摄像头组成。两个固定的摄像机用作立体系统来估计脸和手的位置。立体摄像机通过我们提出的标准肤色方法来检测人脸和手。然后估计目标的距离。其次,我们跟踪多个目标,估计连续帧之间目标的位置和大小。另外两个摄像头负责跟踪面部和手部等目标。如果目标的大小不适合识别,跟踪摄像机将获取其缩放后的图像。由于我们的系统有两个跟踪摄像头,它可以同时跟踪两个目标。为了识别人脸和手势,我们利用线性判别分析提出了四个方向特征。利用该系统,我们对人体位置估计、多人脸跟踪、人脸和手势识别进行了实验。实验结果表明,该系统可以利用立体摄像机估计人体的位置,并利用目标的位置和大小来跟踪多个重叠的目标。此外,我们的系统可以利用这四个方向特征来识别人脸和手势。
{"title":"Focus of attention for face and hand gesture recognition using multiple cameras","authors":"H. Hongo, M. Yasumoto, Y. Niwa, M. Ohya, Kazuhiko Yamamoto","doi":"10.1109/AFGR.2000.840627","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840627","url":null,"abstract":"We propose a multi-camera system that can track multiple human faces and hands as well as focus on face and hand gestures for recognition. Our current system consists of four cameras. Two fixed cameras are used as a stereo system to estimate face and hand positions. The stereo camera detects faces and hands by a standard skin color method we propose. The distances of the targets are then estimated. Next to track multiple targets, we estimate the positions and sizes of targets between consecutive frames. The other two cameras perform tracking of such targets as faces and hands. If a target is not the appropriate size for recognition, the tracking cameras acquire its zoomed image. Since our system has two tracking cameras, it can track two targets at the same time. To recognize faces and hand gestures, we propose four directional features by using linear discriminant analysis. Using our system, we experimented on human position estimation, multiple face tracking, and face and hand gesture recognition. These experiments showed that our system could estimate human position with the stereo camera and track multiple targets by using target positions and sizes even if the persons overlapped with each other. In addition, our system could recognize faces and hand gestures by using the four directional features.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130819165","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 47
View-based active appearance models 基于视图的主动外观模型
Tim Cootes, G. V. Wheeler, K. N. Walker, C. Taylor
We demonstrate that a small number of 2D statistical models are sufficient to capture the shape and appearance of a face from any viewpoint (full profile to front-to-parallel). Each model is linear and can be matched rapidly to new images using the active appearance model algorithm. We show how such a set of models can be used to estimate head pose, to track faces through large angles of head rotation and to synthesize faces from unseen viewpoints.
我们证明,少量的二维统计模型就足以捕捉任何视角(全轮廓到正面到平行)下的脸部形状和外观。每个模型都是线性的,可以使用主动外观模型算法与新图像快速匹配。我们展示了如何利用这样一组模型来估计头部姿态、通过大角度的头部旋转来跟踪人脸,以及如何从未曾见过的视角合成人脸。
{"title":"View-based active appearance models","authors":"Tim Cootes, G. V. Wheeler, K. N. Walker, C. Taylor","doi":"10.1109/AFGR.2000.840639","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840639","url":null,"abstract":"We demonstrate that a small number of 2D statistical models are sufficient to capture the shape and appearance of a face from any viewpoint (full profile to front-to-parallel). Each model is linear and can be matched rapidly to new images using the active appearance model algorithm. We show how such a set of models can be used to estimate head pose, to track faces through large angles of head rotation and to synthesize faces from unseen viewpoints.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126453594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 623
Robust face tracking using color 使用颜色的鲁棒面部跟踪
Karl Schwerdt, J. Crowley
We discuss a new robust tracking technique applied to histograms of intensity-normalized color. This technique supports a video codec based on orthonormal basis coding. Orthonormal basis coding can be very efficient when the images to be coded have been normalized in size and position. However an imprecise tracking procedure can have a negative impact on the efficiency and the quality of reconstruction of this technique, since it may increase the size of the required basis space. The face tracking procedure described in this paper has certain advantages, such as greater stability, higher precision, and less jitter, over conventional tracking techniques using color histograms. In addition to those advantages, the features of the tracked object such as mean and variance are mathematically describable.
讨论了一种应用于灰度归一化颜色直方图的鲁棒跟踪技术。该技术支持基于正交基编码的视频编解码器。当要编码的图像在大小和位置上已经归一化时,标准正交基编码是非常有效的。然而,不精确的跟踪程序可能会对该技术的效率和重建质量产生负面影响,因为它可能会增加所需基空间的大小。与传统的使用颜色直方图的跟踪技术相比,本文所描述的人脸跟踪方法具有稳定性好、精度高、抖动小等优点。除了这些优点,跟踪对象的特征,如均值和方差是数学上可描述的。
{"title":"Robust face tracking using color","authors":"Karl Schwerdt, J. Crowley","doi":"10.1109/AFGR.2000.840617","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840617","url":null,"abstract":"We discuss a new robust tracking technique applied to histograms of intensity-normalized color. This technique supports a video codec based on orthonormal basis coding. Orthonormal basis coding can be very efficient when the images to be coded have been normalized in size and position. However an imprecise tracking procedure can have a negative impact on the efficiency and the quality of reconstruction of this technique, since it may increase the size of the required basis space. The face tracking procedure described in this paper has certain advantages, such as greater stability, higher precision, and less jitter, over conventional tracking techniques using color histograms. In addition to those advantages, the features of the tracked object such as mean and variance are mathematically describable.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126515996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 199
期刊
Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1