首页 > 最新文献

Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)最新文献

英文 中文
Tracking a person with 3-D motion by integrating optical flow and depth 通过整合光流和深度来跟踪具有三维运动的人
R. Okada, Y. Shirai, J. Miura
This paper describes a method of tracking a person with 3D translation and rotation by integrating optical flow and depth. The target region is first extracted based on the probability of each pixel belonging to the target person. The target state (3D position, posture, motion) is estimated based on the shape and the position of the target region in addition to optical flow and depth. Multiple target states are maintained when the image measurements give rise to ambiguities about the target state. Experimental results with real image sequences show the effectiveness of our method.
本文介绍了一种结合光流和深度的三维平移和旋转跟踪人的方法。首先根据每个像素属于目标人的概率提取目标区域。除了光流和深度外,还根据目标区域的形状和位置估计目标状态(3D位置、姿态、运动)。当图像测量引起目标状态的模糊性时,保持多个目标状态。实际图像序列的实验结果表明了该方法的有效性。
{"title":"Tracking a person with 3-D motion by integrating optical flow and depth","authors":"R. Okada, Y. Shirai, J. Miura","doi":"10.1109/AFGR.2000.840656","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840656","url":null,"abstract":"This paper describes a method of tracking a person with 3D translation and rotation by integrating optical flow and depth. The target region is first extracted based on the probability of each pixel belonging to the target person. The target state (3D position, posture, motion) is estimated based on the shape and the position of the target region in addition to optical flow and depth. Multiple target states are maintained when the image measurements give rise to ambiguities about the target state. Experimental results with real image sequences show the effectiveness of our method.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"108 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129877919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
Face recognition algorithms as models of human face processing 人脸识别算法作为人脸处理的模型
A. O’Toole, Y. Cheng, B. Ross, Heather A. Wild, P. Phillips
We evaluated the adequacy of computational algorithms as models of human face processing by looking at how the algorithms and humans process individual faces. By comparing model- and human-generated measures of the similarity between pairs of faces, we were able to assess the accord between several automatic face recognition algorithms and human perceivers. Multidimensional scaling (MDS) was used to create a spatial representation of the subject response patterns. Next, the model response patterns were projected into this space. The results revealed a common bimodal structure for both the subjects and for most of the models. The bimodal subject structure reflected strategy differences in making similarity decisions. For the models, the bimodal structure was related to combined aspects of the representations and the distance metrics used in the implementations.
我们通过观察算法和人类如何处理单个面孔来评估计算算法作为人脸处理模型的充分性。通过比较模型和人类生成的面部对之间的相似性度量,我们能够评估几种自动面部识别算法与人类感知器之间的一致性。多维尺度(MDS)被试反应模式的空间表征。然后,将模型响应模式投影到该空间中。结果显示,受试者和大多数模型都有一个共同的双峰结构。双峰主语结构反映了相似决策的策略差异。对于模型,双峰结构与实现中使用的表示和距离度量的组合方面相关。
{"title":"Face recognition algorithms as models of human face processing","authors":"A. O’Toole, Y. Cheng, B. Ross, Heather A. Wild, P. Phillips","doi":"10.1109/AFGR.2000.840689","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840689","url":null,"abstract":"We evaluated the adequacy of computational algorithms as models of human face processing by looking at how the algorithms and humans process individual faces. By comparing model- and human-generated measures of the similarity between pairs of faces, we were able to assess the accord between several automatic face recognition algorithms and human perceivers. Multidimensional scaling (MDS) was used to create a spatial representation of the subject response patterns. Next, the model response patterns were projected into this space. The results revealed a common bimodal structure for both the subjects and for most of the models. The bimodal subject structure reflected strategy differences in making similarity decisions. For the models, the bimodal structure was related to combined aspects of the representations and the distance metrics used in the implementations.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116992665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Bimodal emotion recognition 双峰情绪识别
L. D. Silva, Pei Chi Ng
This paper describes the use of statistical techniques and hidden Markov models (HMM) in the recognition of emotions. The method aims to classify 6 basic emotions (anger, dislike, fear, happiness, sadness and surprise) from both facial expressions (video) and emotional speech (audio). The emotions of 2 human subjects were recorded and analyzed. The findings show that the audio and video information can be combined using a rule-based system to improve the recognition rate.
本文描述了统计技术和隐马尔可夫模型(HMM)在情绪识别中的应用。该方法旨在从面部表情(视频)和情感语言(音频)中分类出6种基本情绪(愤怒、厌恶、恐惧、快乐、悲伤和惊讶)。记录和分析2名人类受试者的情绪。研究结果表明,基于规则的系统可以将音频和视频信息结合起来,提高识别率。
{"title":"Bimodal emotion recognition","authors":"L. D. Silva, Pei Chi Ng","doi":"10.1109/AFGR.2000.840655","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840655","url":null,"abstract":"This paper describes the use of statistical techniques and hidden Markov models (HMM) in the recognition of emotions. The method aims to classify 6 basic emotions (anger, dislike, fear, happiness, sadness and surprise) from both facial expressions (video) and emotional speech (audio). The emotions of 2 human subjects were recorded and analyzed. The findings show that the audio and video information can be combined using a rule-based system to improve the recognition rate.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121374012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 126
Crane gesture recognition using pseudo 3-D hidden Markov models 基于伪三维隐马尔可夫模型的起重机手势识别
Stefan Müller, S. Eickeler, G. Rigoll
A recognition technique based on novel pseudo 3D hidden Markov models, which can integrate spatial as well as temporal derived features is presented. The approach allows the recognition of dynamic gestures such as waving hands as well as static gestures such as standing in a special pose. Pseudo 3D hidden Markov models (P3DHMM) are an extension of the pseudo 2D case, which has been successfully used for the classification of images and the recognition of faces. In the P3DHMM case the so-called superstates contain P2DHMM and thus whole image sequences can be generated by these models. Our approach has been evaluated on a crane signal database, which consists of 12 different predefined gestures for maneuvering cranes.
提出了一种基于伪三维隐马尔可夫模型的识别方法,该方法可以将空间特征和时间特征相结合。该方法允许识别动态手势,如挥手,以及静态手势,如站在一个特殊的姿势。伪三维隐马尔可夫模型(P3DHMM)是对伪二维隐马尔可夫模型的扩展,已成功地用于图像分类和人脸识别。在P3DHMM的情况下,所谓的超态包含P2DHMM,因此整个图像序列可以由这些模型生成。我们的方法已经在起重机信号数据库上进行了评估,该数据库由12种不同的预定义手势组成,用于操纵起重机。
{"title":"Crane gesture recognition using pseudo 3-D hidden Markov models","authors":"Stefan Müller, S. Eickeler, G. Rigoll","doi":"10.1109/AFGR.2000.840665","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840665","url":null,"abstract":"A recognition technique based on novel pseudo 3D hidden Markov models, which can integrate spatial as well as temporal derived features is presented. The approach allows the recognition of dynamic gestures such as waving hands as well as static gestures such as standing in a special pose. Pseudo 3D hidden Markov models (P3DHMM) are an extension of the pseudo 2D case, which has been successfully used for the classification of images and the recognition of faces. In the P3DHMM case the so-called superstates contain P2DHMM and thus whole image sequences can be generated by these models. Our approach has been evaluated on a crane signal database, which consists of 12 different predefined gestures for maneuvering cranes.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123353035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Estimation of the illuminant colour from human skin colour 根据人的肤色估计光源颜色
M. Störring, H. J. Andersen, E. Granum
Colour is an important and useful feature for object tracking and recognition in computer vision. However, it has the difficulty that the colour of the object changes if the illuminant colour changes. But under known illuminant colour it becomes a robust feature. There are more and more computer vision applications tracking humans, for example in interfaces for human computer interaction or automatic camera men, where skin colour is an often-used feature. Hence, it would be of significant importance to know the illuminant colour in such applications. This paper proposes a novel method to estimate the current illuminant colour from skin colour observations. The method is based on a physical model of reflections, the assumption that illuminant colours are located close to the Planckian locus, and the knowledge about the camera parameters. The method is empirically tested using real images. The average estimation error of the correlated colour temperature is as small as 180 K. Applications are for example in colour-based tracking to adapt to changes in lighting and in visualisation to re-render image colours to their appearance under canonical viewing conditions.
在计算机视觉中,颜色是物体跟踪和识别的重要而有用的特征。但是,如果光源颜色发生变化,物体的颜色也会发生变化。但在已知的光源颜色下,它成为一个强大的特征。越来越多的计算机视觉应用跟踪人类,例如在人机交互界面或自动相机中,肤色是一个经常使用的特征。因此,在这种应用中,了解光源颜色是非常重要的。本文提出了一种从肤色观测中估计当前光源颜色的新方法。该方法基于反射的物理模型,光源颜色位于普朗克轨迹附近的假设,以及关于相机参数的知识。用实际图像对该方法进行了实证检验。相关色温的平均估计误差小至180k。例如,应用于基于颜色的跟踪,以适应照明的变化,并在可视化中重新渲染图像颜色,使其在规范的观看条件下呈现外观。
{"title":"Estimation of the illuminant colour from human skin colour","authors":"M. Störring, H. J. Andersen, E. Granum","doi":"10.1109/AFGR.2000.840613","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840613","url":null,"abstract":"Colour is an important and useful feature for object tracking and recognition in computer vision. However, it has the difficulty that the colour of the object changes if the illuminant colour changes. But under known illuminant colour it becomes a robust feature. There are more and more computer vision applications tracking humans, for example in interfaces for human computer interaction or automatic camera men, where skin colour is an often-used feature. Hence, it would be of significant importance to know the illuminant colour in such applications. This paper proposes a novel method to estimate the current illuminant colour from skin colour observations. The method is based on a physical model of reflections, the assumption that illuminant colours are located close to the Planckian locus, and the knowledge about the camera parameters. The method is empirically tested using real images. The average estimation error of the correlated colour temperature is as small as 180 K. Applications are for example in colour-based tracking to adapt to changes in lighting and in visualisation to re-render image colours to their appearance under canonical viewing conditions.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124125704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 53
Support vector regression and classification based multi-view face detection and recognition 基于支持向量回归和分类的多视图人脸检测与识别
Yongmin Li, S. Gong, H. Liddell
A support vector machine-based multi-view face detection and recognition framework is described. Face detection is carried out by constructing several detectors, each of them in charge of one specific view. The symmetrical property of face images is employed to simplify the complexity of the modelling. The estimation of head pose, which is achieved by using the support vector regression technique, provides crucial information for choosing the appropriate face detector. This helps to improve the accuracy and reduce the computation in multi-view face detection compared to other methods. For video sequences, further computational reduction can be achieved by using a pose change smoothing strategy. When face detectors find a face in frontal view, a support vector machine-based multi-class classifier is activated for face recognition. All the above issues are integrated under a support vector machine framework. Test results on four video sequences are presented, among them the detection rate is above 95%, recognition accuracy is above 90%, average pose estimation error is around 10/spl deg/, and the full detection and recognition speed is up to 4 frames/second on a Pentium II 300 PC.
提出了一种基于支持向量机的多视图人脸检测与识别框架。人脸检测是通过构建多个检测器来实现的,每个检测器负责一个特定的视图。利用人脸图像的对称性,简化了建模的复杂性。利用支持向量回归技术实现头部姿态的估计,为选择合适的人脸检测器提供了重要信息。与其他方法相比,这有助于提高多视图人脸检测的精度和减少计算量。对于视频序列,可以通过使用姿态变化平滑策略进一步减少计算量。当人脸检测器在正面视图中发现人脸时,激活基于支持向量机的多类分类器进行人脸识别。将上述问题集成在支持向量机框架下。给出了在4个视频序列上的测试结果,其中检测率在95%以上,识别精度在90%以上,平均姿态估计误差在10/spl°/左右,在Pentium II 300 PC上的全部检测和识别速度可达4帧/秒。
{"title":"Support vector regression and classification based multi-view face detection and recognition","authors":"Yongmin Li, S. Gong, H. Liddell","doi":"10.1109/AFGR.2000.840650","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840650","url":null,"abstract":"A support vector machine-based multi-view face detection and recognition framework is described. Face detection is carried out by constructing several detectors, each of them in charge of one specific view. The symmetrical property of face images is employed to simplify the complexity of the modelling. The estimation of head pose, which is achieved by using the support vector regression technique, provides crucial information for choosing the appropriate face detector. This helps to improve the accuracy and reduce the computation in multi-view face detection compared to other methods. For video sequences, further computational reduction can be achieved by using a pose change smoothing strategy. When face detectors find a face in frontal view, a support vector machine-based multi-class classifier is activated for face recognition. All the above issues are integrated under a support vector machine framework. Test results on four video sequences are presented, among them the detection rate is above 95%, recognition accuracy is above 90%, average pose estimation error is around 10/spl deg/, and the full detection and recognition speed is up to 4 frames/second on a Pentium II 300 PC.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"259 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131521216","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 276
Comparative evaluation of face sequence matching for content-based video access 基于内容的视频访问人脸序列匹配的比较评价
S. Satoh
The paper presents a comparative evaluation of matching methods of face sequences obtained from actual videos. Face information is quite important in videos, especially in news programs, dramas, and movies. Accurate face sequence matching enables many multimedia applications including content-based face retrieval, automated face annotation, video authoring, etc. However, face sequences in videos are subject to variation in lighting condition, pose, facial expression, etc., which cause difficulty in face matching. In order to cope with this problem, several face sequence matching methods are proposed by extending face still image matching, traditional pattern recognition, and recent pattern recognition techniques. They are expected to be applicable to face sequences extracted from actual videos. The performance of these methods is evaluated as the accuracy of face sequence annotation using the methods. The accuracy is evaluated using a considerable amount of actual drama videos. The evaluation results reveal merits and demerits of these methods, and indicate future research directions of face matching for videos.
本文对实际视频中人脸序列的匹配方法进行了比较评价。面部信息在视频中非常重要,尤其是在新闻节目、电视剧和电影中。准确的人脸序列匹配使许多多媒体应用得以实现,包括基于内容的人脸检索、自动人脸注释、视频创作等。然而,视频中的人脸序列受到光照条件、姿势、面部表情等变化的影响,给人脸匹配带来了困难。为了解决这一问题,通过扩展人脸静止图像匹配、传统模式识别和最新模式识别技术,提出了几种人脸序列匹配方法。它们有望适用于从实际视频中提取的人脸序列。通过对人脸序列标注的准确性来评价这些方法的性能。使用相当数量的实际戏剧视频来评估准确性。评价结果揭示了这些方法的优缺点,并指出了未来视频人脸匹配的研究方向。
{"title":"Comparative evaluation of face sequence matching for content-based video access","authors":"S. Satoh","doi":"10.1109/AFGR.2000.840629","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840629","url":null,"abstract":"The paper presents a comparative evaluation of matching methods of face sequences obtained from actual videos. Face information is quite important in videos, especially in news programs, dramas, and movies. Accurate face sequence matching enables many multimedia applications including content-based face retrieval, automated face annotation, video authoring, etc. However, face sequences in videos are subject to variation in lighting condition, pose, facial expression, etc., which cause difficulty in face matching. In order to cope with this problem, several face sequence matching methods are proposed by extending face still image matching, traditional pattern recognition, and recent pattern recognition techniques. They are expected to be applicable to face sequences extracted from actual videos. The performance of these methods is evaluated as the accuracy of face sequence annotation using the methods. The accuracy is evaluated using a considerable amount of actual drama videos. The evaluation results reveal merits and demerits of these methods, and indicate future research directions of face matching for videos.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"62 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116972306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 89
Learning-based approach to real time tracking and analysis of faces 基于学习的人脸实时跟踪和分析方法
Vinay P. Kumar, T. Poggio
This paper describes a trainable system capable of tracking faces and facial features like eyes and nostrils and estimating basic mouth features such as degrees of openness and smile in real time. In developing this system, we have addressed the twin issues of image representation and algorithms for learning. We have used the invariance properties of image representations based on Haar wavelets to robustly capture various facial features. Similarly, unlike previous approaches this system is entirely trained using examples and does not rely on a priori (hand-crafted) models of facial features based on an optical flow or facial musculature. The system works in several stages that begin with face detection, followed by localization of facial features and estimation of mouth parameters. Each of these stages is formulated as a problem in supervised learning from examples. We apply the new and robust technique of support vector machines (SVM) for classification in the stage of skin segmentation, face detection and eye detection. Estimation of mouth parameters is modeled as a regression from a sparse subset of coefficients (basis functions) of an overcomplete dictionary of Haar wavelets.
本文描述了一个可训练的系统,该系统能够跟踪面部和面部特征,如眼睛和鼻孔,并实时估计基本的嘴部特征,如开放程度和微笑。在开发这个系统的过程中,我们解决了图像表示和学习算法的双重问题。我们利用基于哈尔小波的图像表示的不变性来鲁棒地捕获各种面部特征。同样,与以前的方法不同,该系统完全使用示例进行训练,不依赖于基于光流或面部肌肉组织的先验(手工制作)面部特征模型。该系统分几个阶段工作,从面部检测开始,然后是面部特征的定位和口腔参数的估计。这些阶段中的每一个都被表述为监督学习中的一个问题。在皮肤分割、人脸检测和眼睛检测阶段,我们将新的鲁棒性支持向量机(SVM)技术应用到分类中。口部参数的估计是由Haar小波的过完备字典的系数(基函数)的稀疏子集回归建模的。
{"title":"Learning-based approach to real time tracking and analysis of faces","authors":"Vinay P. Kumar, T. Poggio","doi":"10.1109/AFGR.2000.840618","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840618","url":null,"abstract":"This paper describes a trainable system capable of tracking faces and facial features like eyes and nostrils and estimating basic mouth features such as degrees of openness and smile in real time. In developing this system, we have addressed the twin issues of image representation and algorithms for learning. We have used the invariance properties of image representations based on Haar wavelets to robustly capture various facial features. Similarly, unlike previous approaches this system is entirely trained using examples and does not rely on a priori (hand-crafted) models of facial features based on an optical flow or facial musculature. The system works in several stages that begin with face detection, followed by localization of facial features and estimation of mouth parameters. Each of these stages is formulated as a problem in supervised learning from examples. We apply the new and robust technique of support vector machines (SVM) for classification in the stage of skin segmentation, face detection and eye detection. Estimation of mouth parameters is modeled as a regression from a sparse subset of coefficients (basis functions) of an overcomplete dictionary of Haar wavelets.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124333615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
Real-time multiple face detection using active illumination 使用主动照明的实时多人脸检测
C. Morimoto, M. Flickner
This paper presents a multiple face detector based on a robust pupil detection technique. The pupil detector uses active illumination that exploits the retro-reflectivity property of eyes to facilitate detection. The detection range of this method is appropriate for interactive desktop and kiosk applications. Once the location of the pupil candidates are computed, the candidates are filtered and grouped into pairs that correspond to faces using heuristic rules. To demonstrate the robustness of the face detection technique, a dual-mode face tracker was developed, which is initialized with the most salient detected face. Recursive estimators are used to guarantee the stability of the process and combine the measurements from the multi-face detector and a feature correlation tracker. The estimated position of the face is used to control a pan-tilt servo mechanism in real-time, that moves the camera to keep the tracked face always centered in the image.
提出了一种基于鲁棒瞳孔检测技术的多人脸检测器。瞳孔检测器使用主动照明,利用眼睛的反向反射特性来方便检测。这种方法的检测范围适用于交互式桌面和kiosk应用程序。一旦计算出候选瞳孔的位置,候选瞳孔就会被过滤,并使用启发式规则将候选瞳孔分组成对应于人脸的成对。为了证明人脸检测技术的鲁棒性,开发了一种双模人脸跟踪器,该跟踪器使用检测到的最显著的人脸进行初始化。为了保证过程的稳定性,使用了递归估计器,并将多面检测器的测量结果与特征相关跟踪器相结合。人脸的估计位置用于实时控制平移伺服机构,该机构移动摄像机以使被跟踪的人脸始终处于图像的中心。
{"title":"Real-time multiple face detection using active illumination","authors":"C. Morimoto, M. Flickner","doi":"10.1109/AFGR.2000.840605","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840605","url":null,"abstract":"This paper presents a multiple face detector based on a robust pupil detection technique. The pupil detector uses active illumination that exploits the retro-reflectivity property of eyes to facilitate detection. The detection range of this method is appropriate for interactive desktop and kiosk applications. Once the location of the pupil candidates are computed, the candidates are filtered and grouped into pairs that correspond to faces using heuristic rules. To demonstrate the robustness of the face detection technique, a dual-mode face tracker was developed, which is initialized with the most salient detected face. Recursive estimators are used to guarantee the stability of the process and combine the measurements from the multi-face detector and a feature correlation tracker. The estimated position of the face is used to control a pan-tilt servo mechanism in real-time, that moves the camera to keep the tracked face always centered in the image.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122065988","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
Detection and tracking of facial features in real time using a synergistic approach of spatio-temporal models and generalized Hough-transform techniques 利用时空模型和广义霍夫变换技术的协同方法实时检测和跟踪面部特征
A. Schubert
The proposed algorithm requires the description of the facial features as 3D-polygons (optionally extended by additional intensity information) which are assembled in a 3D-model of the head provided for in separate data files. Detection is achieved by using a special implementation of the generalized Hough transform (GHT) for which the forms are generated by projecting the 3D-model into the image plane. In the initialization phase a comparatively wide range of relative positions and attitudes between head and camera has to be tested for. Aiming for illumination-independence, only information about the sign of the difference between the expected intensities on both sides of the edge of the polygons may be additionally used in the GHT. Once a feature is found, further search for the remaining features can be restricted by the use of the 3D-model. The detection of a minimum number of features starts the tracking phase which is performed by using an extended Kalman filter (EKF) and assuming a first- or second-order dynamical model for the state variables describing the position and the attitude of the head. Synergistic advantages between GHT and EKF can be realized since the EKF and the projection into the image plane yield a rather good prediction of the forms to be detected by the GHT. This reduces considerably the search space in the image and in the parameter space. On the other hand the GHT offers a solution to the matching problem between image and object features. During the tracking phase the GHT can be further enhanced by monitoring the actual intensities along the edges of the polygons, their assignment to the corresponding 3D-object features, and their use for feature selection during the accumulation process. The algorithm runs on a dual Pentium II 333 MHz with a cycle time of 40 ms in real time.
所提出的算法要求将面部特征描述为3d多边形(可选地通过额外的强度信息扩展),这些多边形组装在头部的3d模型中,并在单独的数据文件中提供。检测是通过使用广义霍夫变换(GHT)的特殊实现来实现的,该变换的形式是通过将3d模型投影到图像平面上来生成的。在初始化阶段,必须测试头部和相机之间相对位置和姿态的相对范围。为了与光照无关,只有关于多边形边缘两侧预期强度之间的差异符号的信息可以在GHT中额外使用。一旦发现了一个特征,进一步搜索剩余的特征可能会受到3d模型使用的限制。通过使用扩展卡尔曼滤波(EKF)并假设描述头部位置和姿态的状态变量的一阶或二阶动态模型,检测到最小数量的特征开始跟踪阶段。由于EKF和投影到图像平面上可以很好地预测GHT要检测的形状,因此可以实现GHT和EKF之间的协同优势。这大大减少了图像和参数空间中的搜索空间。另一方面,GHT为图像和物体特征之间的匹配问题提供了一个解决方案。在跟踪阶段,可以通过监测沿多边形边缘的实际强度、它们对相应3d物体特征的分配以及它们在积累过程中用于特征选择来进一步增强GHT。该算法在双Pentium II 333 MHz处理器上实时运行,周期时间为40 ms。
{"title":"Detection and tracking of facial features in real time using a synergistic approach of spatio-temporal models and generalized Hough-transform techniques","authors":"A. Schubert","doi":"10.1109/AFGR.2000.840621","DOIUrl":"https://doi.org/10.1109/AFGR.2000.840621","url":null,"abstract":"The proposed algorithm requires the description of the facial features as 3D-polygons (optionally extended by additional intensity information) which are assembled in a 3D-model of the head provided for in separate data files. Detection is achieved by using a special implementation of the generalized Hough transform (GHT) for which the forms are generated by projecting the 3D-model into the image plane. In the initialization phase a comparatively wide range of relative positions and attitudes between head and camera has to be tested for. Aiming for illumination-independence, only information about the sign of the difference between the expected intensities on both sides of the edge of the polygons may be additionally used in the GHT. Once a feature is found, further search for the remaining features can be restricted by the use of the 3D-model. The detection of a minimum number of features starts the tracking phase which is performed by using an extended Kalman filter (EKF) and assuming a first- or second-order dynamical model for the state variables describing the position and the attitude of the head. Synergistic advantages between GHT and EKF can be realized since the EKF and the projection into the image plane yield a rather good prediction of the forms to be detected by the GHT. This reduces considerably the search space in the image and in the parameter space. On the other hand the GHT offers a solution to the matching problem between image and object features. During the tracking phase the GHT can be further enhanced by monitoring the actual intensities along the edges of the polygons, their assignment to the corresponding 3D-object features, and their use for feature selection during the accumulation process. The algorithm runs on a dual Pentium II 333 MHz with a cycle time of 40 ms in real time.","PeriodicalId":360065,"journal":{"name":"Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2000-03-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129818713","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1