首页 > 最新文献

7th International Conference on Automatic Face and Gesture Recognition (FGR06)最新文献

英文 中文
Robust method for real-time, continuous, 3D detection of obstructed faces in indoors environments 在室内环境中实时、连续、三维检测遮挡人脸的鲁棒方法
S. Grange, C. Baur
We address the need for robust detection of obstructed human features in complex environments, with a focus on intelligent surgical UIs. In our setup, real-time detection is used to find features without the help of local (spatial or temporal) information. Such a detector is used to validate, correct or reject the output of the visual feature tracking, which is locally more robust, but drifts over time. In operating rooms (OR), surgeon faces are typically obstructed by sterile clothing and tools, making statistical and/or feature-based face detection approaches ineffective. We propose a new method for face detection that relies on geometric information from disparity maps, locally refined by color processing. We have applied our method to a surgical mock-up scene, as well as to images gathered during real surgery. Running in a real-time, continuous detection loop, our detector successfully found 99% of target heads (0.1% false positive) in our simulated setup, and 98% of target heads (0.5% false positive) in the surgical theater
我们解决了在复杂环境中对障碍物人体特征进行鲁棒检测的需求,重点是智能外科ui。在我们的设置中,实时检测用于在没有局部(空间或时间)信息的帮助下找到特征。这种检测器用于验证、纠正或拒绝视觉特征跟踪的输出,它在局部更鲁棒,但随着时间的推移而漂移。在手术室(OR),外科医生的脸通常被无菌的衣服和工具挡住,使得统计和/或基于特征的面部检测方法无效。我们提出了一种新的人脸检测方法,该方法依赖于视差图中的几何信息,并通过颜色处理进行局部细化。我们已经将我们的方法应用于一个手术模拟场景,以及在真实手术中收集的图像。在实时、连续的检测循环中,我们的检测器在模拟设置中成功发现了99%的目标头部(0.1%假阳性),在手术室中成功发现了98%的目标头部(0.5%假阳性)
{"title":"Robust method for real-time, continuous, 3D detection of obstructed faces in indoors environments","authors":"S. Grange, C. Baur","doi":"10.1109/FGR.2006.97","DOIUrl":"https://doi.org/10.1109/FGR.2006.97","url":null,"abstract":"We address the need for robust detection of obstructed human features in complex environments, with a focus on intelligent surgical UIs. In our setup, real-time detection is used to find features without the help of local (spatial or temporal) information. Such a detector is used to validate, correct or reject the output of the visual feature tracking, which is locally more robust, but drifts over time. In operating rooms (OR), surgeon faces are typically obstructed by sterile clothing and tools, making statistical and/or feature-based face detection approaches ineffective. We propose a new method for face detection that relies on geometric information from disparity maps, locally refined by color processing. We have applied our method to a surgical mock-up scene, as well as to images gathered during real surgery. Running in a real-time, continuous detection loop, our detector successfully found 99% of target heads (0.1% false positive) in our simulated setup, and 98% of target heads (0.5% false positive) in the surgical theater","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114704966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Accurate face localisation for faces under active near-IR illumination 主动近红外照明下的人脸精确定位
X. Zou, J. Kittler, K. Messer
In this paper we propose a novel approach to accurate face localisation for faces under near-infrared (near-IR) illumination. The circular shape of the bright pupils is a scale and rotation invariant feature which is exploited to quickly detect pupil candidates. As the first step of face localisation, a rule-based pupil detector is employed to find candidate pupil edges from the edge map. Candidate eye centres for each eye are selected from the neighborhood of corresponding pupil regions and sorted based on the similarity to eye templates. Two support vector machine (SVM) classifiers based on eye appearance are employed to validate those candidates for each eye individually. Finally candidates are further validated in pair by an SVM classifier based on global face appearance. In the experiment on a near-IR face database with 40 subjects and 48 images per subject, 96.5% images are accurately localised using the proposed approach
本文提出了一种近红外光照下人脸精确定位的新方法。明亮瞳孔的圆形是一种尺度和旋转不变性特征,利用这种特征可以快速检测候选瞳孔。作为人脸定位的第一步,使用基于规则的瞳孔检测器从边缘图中寻找候选瞳孔边缘。从对应瞳孔区域的邻域中选择候选眼中心,并根据与眼模板的相似性对候选眼中心进行排序。采用两种基于眼睛外观的支持向量机(SVM)分类器分别对每只眼睛的候选对象进行验证。最后,使用基于全局人脸外观的SVM分类器进一步对候选人脸进行配对验证。在一个近红外人脸数据库中,40个受试者,每个受试者48张图像,使用该方法精确定位了96.5%的图像
{"title":"Accurate face localisation for faces under active near-IR illumination","authors":"X. Zou, J. Kittler, K. Messer","doi":"10.1109/FGR.2006.18","DOIUrl":"https://doi.org/10.1109/FGR.2006.18","url":null,"abstract":"In this paper we propose a novel approach to accurate face localisation for faces under near-infrared (near-IR) illumination. The circular shape of the bright pupils is a scale and rotation invariant feature which is exploited to quickly detect pupil candidates. As the first step of face localisation, a rule-based pupil detector is employed to find candidate pupil edges from the edge map. Candidate eye centres for each eye are selected from the neighborhood of corresponding pupil regions and sorted based on the similarity to eye templates. Two support vector machine (SVM) classifiers based on eye appearance are employed to validate those candidates for each eye individually. Finally candidates are further validated in pair by an SVM classifier based on global face appearance. In the experiment on a near-IR face database with 40 subjects and 48 images per subject, 96.5% images are accurately localised using the proposed approach","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121722737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A landmark paper in face recognition 人脸识别领域具有里程碑意义的论文
G. M. Beumer, Q. Tao, A. Bazen, R. Veldhuis
Good registration (alignment to a reference) is essential for accurate face recognition. The effects of the number of landmarks on the mean localization error and the recognition performance are studied. Two landmarking methods are explored and compared for that purpose: (1) the most likely-landmark locator (MLLL), based on maximizing the likelihood ratio, and (2) Viola-Jones detection. Both use the locations of facial features (eyes, nose, mouth, etc) as landmarks. Further, a landmark-correction method (BILBO) based on projection into a subspace is introduced. The MLLL has been trained for locating 17 landmarks and the Viola-Jones method for 5. The mean localization errors and effects on the verification performance have been measured. It was found that on the eyes, the Viola-Jones detector is about 1% of the interocular distance more accurate than the MLLL-BILBO combination. On the nose and mouth, the MLLL-BILBO combination is about 0.5% of the inter-ocular distance more accurate than the Viola-Jones detector. Using more landmarks will result in lower equal-error rates, even when the landmarking is not so accurate. If the same landmarks are used, the most accurate landmarking method gives the best verification performance
良好的配准(与参考点对齐)对于准确的人脸识别至关重要。研究了标记数对平均定位误差和识别性能的影响。为此,我们探索并比较了两种地标定位方法:(1)基于最大似然比的最可能地标定位器(MLLL)和(2)Viola-Jones检测。两者都使用面部特征(眼睛、鼻子、嘴巴等)的位置作为地标。在此基础上,提出了一种基于子空间投影的地标校正方法(BILBO)。MLLL已经被训练定位17个地标,Viola-Jones方法被训练定位5个地标。测量了平均定位误差及其对验证性能的影响。研究发现,在眼睛上,Viola-Jones探测器比mll - bilbo组合的眼间距离精度提高了约1%。在鼻子和嘴巴上,MLLL-BILBO组合比Viola-Jones检测器的眼间距离精确约0.5%。使用更多的地标将导致更低的等错误率,即使地标不是那么准确。如果使用相同的地标,最精确的地标方法可以提供最佳的验证性能
{"title":"A landmark paper in face recognition","authors":"G. M. Beumer, Q. Tao, A. Bazen, R. Veldhuis","doi":"10.1109/FGR.2006.10","DOIUrl":"https://doi.org/10.1109/FGR.2006.10","url":null,"abstract":"Good registration (alignment to a reference) is essential for accurate face recognition. The effects of the number of landmarks on the mean localization error and the recognition performance are studied. Two landmarking methods are explored and compared for that purpose: (1) the most likely-landmark locator (MLLL), based on maximizing the likelihood ratio, and (2) Viola-Jones detection. Both use the locations of facial features (eyes, nose, mouth, etc) as landmarks. Further, a landmark-correction method (BILBO) based on projection into a subspace is introduced. The MLLL has been trained for locating 17 landmarks and the Viola-Jones method for 5. The mean localization errors and effects on the verification performance have been measured. It was found that on the eyes, the Viola-Jones detector is about 1% of the interocular distance more accurate than the MLLL-BILBO combination. On the nose and mouth, the MLLL-BILBO combination is about 0.5% of the inter-ocular distance more accurate than the Viola-Jones detector. Using more landmarks will result in lower equal-error rates, even when the landmarking is not so accurate. If the same landmarks are used, the most accurate landmarking method gives the best verification performance","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134354965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
The isometric self-organizing map for 3D hand pose estimation 三维手部姿态估计的等距自组织图
Haiying Guan, R. Feris, M. Turk
We propose an isometric self-organizing map (ISO-SOM) method for nonlinear dimensionality reduction, which integrates a self-organizing map model and an ISOMAP dimension reduction algorithm, organizing the high dimension data in a low dimension lattice structure. We apply the proposed method to the problem of appearance-based 3D hand posture estimation. As a learning stage, we use a realistic 3D hand model to generate data encoding the mapping between the hand pose space and the image feature space. The intrinsic dimension of such nonlinear mapping is learned by ISOSOM, which clusters the data into a lattice map. We perform 3D hand posture estimation on this map, showing that the ISOSOM algorithm performs better than traditional image retrieval algorithms for pose estimation. We also show that a 2.5D feature representation based on depth edges is clearly superior to intensity edge features commonly used in previous methods
本文提出了一种等距自组织映射(ISO-SOM)的非线性降维方法,该方法将自组织映射模型与ISOMAP降维算法相结合,将高维数据组织在低维点阵结构中。我们将提出的方法应用于基于外观的三维手部姿态估计问题。作为学习阶段,我们使用一个真实的三维手部模型来生成编码手部姿态空间与图像特征空间之间映射的数据。这种非线性映射的内在维数由ISOSOM学习,它将数据聚类到一个点阵图中。我们在这张地图上进行了三维手部姿态估计,结果表明ISOSOM算法比传统的图像检索算法在姿态估计方面表现更好。我们还表明,基于深度边缘的2.5D特征表示明显优于先前方法中常用的强度边缘特征
{"title":"The isometric self-organizing map for 3D hand pose estimation","authors":"Haiying Guan, R. Feris, M. Turk","doi":"10.1109/FGR.2006.103","DOIUrl":"https://doi.org/10.1109/FGR.2006.103","url":null,"abstract":"We propose an isometric self-organizing map (ISO-SOM) method for nonlinear dimensionality reduction, which integrates a self-organizing map model and an ISOMAP dimension reduction algorithm, organizing the high dimension data in a low dimension lattice structure. We apply the proposed method to the problem of appearance-based 3D hand posture estimation. As a learning stage, we use a realistic 3D hand model to generate data encoding the mapping between the hand pose space and the image feature space. The intrinsic dimension of such nonlinear mapping is learned by ISOSOM, which clusters the data into a lattice map. We perform 3D hand posture estimation on this map, showing that the ISOSOM algorithm performs better than traditional image retrieval algorithms for pose estimation. We also show that a 2.5D feature representation based on depth edges is clearly superior to intensity edge features commonly used in previous methods","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128894052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
Face Recognition from a Tabula Rasa Perspective 从白板角度的人脸识别
M. C. Santana, O. Déniz-Suárez, J. Lorenzo-Navarro, M. Hernández-Tejera
In this paper a system for face recognition from a tabula rasa (i.e. blank slate) perspective is described. A priori, the system has the only ability to detect automatically faces and represent them in a space of reduced dimension. Later, the system is exposed to over 400 different identities, observing its recognition performance evolution. The preliminary results achieved indicate on the one side that the system is able to reject most of unknown individuals after an initialization stage. On the other side the ability to recognize known individuals (or revisitors) is still far from being reliable. However, the observation of the recognition evolution results for individuals frequently met suggests that the more meetings are held, the lower recognition error is achieved
本文描述了一种基于白板视角的人脸识别系统。先验地,系统只有自动检测人脸并在降维空间中表示它们的能力。随后,将该系统暴露在400多个不同的身份下,观察其识别性能的演变。所取得的初步结果一方面表明,系统在初始化阶段后能够拒绝大多数未知个体。另一方面,识别已知个体(或回访者)的能力还远远不够可靠。然而,对频繁会面个体的识别演化结果观察表明,会面次数越多,识别误差越小
{"title":"Face Recognition from a Tabula Rasa Perspective","authors":"M. C. Santana, O. Déniz-Suárez, J. Lorenzo-Navarro, M. Hernández-Tejera","doi":"10.1109/FGR.2006.44","DOIUrl":"https://doi.org/10.1109/FGR.2006.44","url":null,"abstract":"In this paper a system for face recognition from a tabula rasa (i.e. blank slate) perspective is described. A priori, the system has the only ability to detect automatically faces and represent them in a space of reduced dimension. Later, the system is exposed to over 400 different identities, observing its recognition performance evolution. The preliminary results achieved indicate on the one side that the system is able to reject most of unknown individuals after an initialization stage. On the other side the ability to recognize known individuals (or revisitors) is still far from being reliable. However, the observation of the recognition evolution results for individuals frequently met suggests that the more meetings are held, the lower recognition error is achieved","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"83 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124118013","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Automatic gesture recognition for intelligent human-robot interaction 用于智能人机交互的自动手势识别
Seong-Whan Lee
An intelligent robot requires natural interaction with humans. Visual interpretation of gestures can be useful in accomplishing natural human-robot interaction (HRl). Previous HRI researches were focused on issues such as hand gesture, sign language, and command gesture recognition. However, automatic recognition of whole body gestures is required in order to operate HRI naturally. This can be a challenging problem because describing and modeling meaningful gesture patterns from whole body gestures are complex tasks. This paper presents a new method for spotting and recognizing whole body key gestures at the same time on a mobile robot. Our method is simultaneously used with other HRI approaches such as speech recognition, face recognition, and so forth. In this regard, both of execution speed and recognition performance should be considered. For efficient and natural operation, we used several approaches at each step of gesture recognition; learning and extraction of articulated joint information, representing gesture as a sequence of clusters, spotting and recognizing a gesture with HMM. In addition, we constructed a large gesture database, with which we verified our method. As a result, our method is successfully included and operated in a mobile robot
智能机器人需要与人类自然互动。手势的视觉解释在完成自然人机交互(HRl)中是有用的。以往的HRI研究主要集中在手势、手语和命令手势识别等问题上。然而,为了自然地操作HRI,需要对全身手势进行自动识别。这可能是一个具有挑战性的问题,因为从全身手势描述和建模有意义的手势模式是一项复杂的任务。提出了一种同时识别移动机器人全身按键手势的新方法。我们的方法与其他HRI方法(如语音识别、人脸识别等)同时使用。在这方面,需要同时考虑执行速度和识别性能。为了高效、自然地进行操作,我们在手势识别的每一步都使用了几种方法;学习和提取关节信息,用聚类序列表示手势,用HMM识别手势。此外,我们构建了一个大型手势数据库,并用该数据库验证了我们的方法。结果表明,该方法已成功地应用于移动机器人中
{"title":"Automatic gesture recognition for intelligent human-robot interaction","authors":"Seong-Whan Lee","doi":"10.1109/FGR.2006.25","DOIUrl":"https://doi.org/10.1109/FGR.2006.25","url":null,"abstract":"An intelligent robot requires natural interaction with humans. Visual interpretation of gestures can be useful in accomplishing natural human-robot interaction (HRl). Previous HRI researches were focused on issues such as hand gesture, sign language, and command gesture recognition. However, automatic recognition of whole body gestures is required in order to operate HRI naturally. This can be a challenging problem because describing and modeling meaningful gesture patterns from whole body gestures are complex tasks. This paper presents a new method for spotting and recognizing whole body key gestures at the same time on a mobile robot. Our method is simultaneously used with other HRI approaches such as speech recognition, face recognition, and so forth. In this regard, both of execution speed and recognition performance should be considered. For efficient and natural operation, we used several approaches at each step of gesture recognition; learning and extraction of articulated joint information, representing gesture as a sequence of clusters, spotting and recognizing a gesture with HMM. In addition, we constructed a large gesture database, with which we verified our method. As a result, our method is successfully included and operated in a mobile robot","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126040802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
Facial features extraction in color images using enhanced active shape model 基于增强主动形状模型的彩色图像人脸特征提取
M. Mahoor, M. Abdel-Mottaleb
In this paper, we present an improved active shape model (ASM) for facial feature extraction. The original ASM method developed by Cootes et al. highly relies on the initialization and the representation of the local structure of the facial features in the image. We use color information to improve the ASM approach for facial feature extraction. The color information is used to localize the centers of the mouth and the eyes to assist the initialization step. Moreover, we model the local structure of the feature points in the RGB color space. Besides, we use 2D affine transformation to align facial features that are perturbed by head pose variations. In fact, the 2D affine transformation compensates for the effects of both head pose variations and the projection of 3D data to 2D. Experiments on a face database of 50 subjects show that our approach outperforms the standard ASM and is successful in facial feature extraction
本文提出了一种改进的主动形状模型(ASM)用于人脸特征提取。Cootes等人开发的原始ASM方法高度依赖于图像中面部特征局部结构的初始化和表示。我们利用颜色信息来改进ASM的人脸特征提取方法。颜色信息用于定位嘴巴和眼睛的中心,以辅助初始化步骤。此外,我们在RGB色彩空间中对特征点的局部结构进行建模。此外,我们使用二维仿射变换来对齐受头部姿态变化干扰的面部特征。事实上,二维仿射变换补偿了头部姿势变化和3D数据到2D的投影的影响。在50个受试者的人脸数据库上进行的实验表明,该方法优于标准的ASM,能够成功地提取人脸特征
{"title":"Facial features extraction in color images using enhanced active shape model","authors":"M. Mahoor, M. Abdel-Mottaleb","doi":"10.1109/FGR.2006.51","DOIUrl":"https://doi.org/10.1109/FGR.2006.51","url":null,"abstract":"In this paper, we present an improved active shape model (ASM) for facial feature extraction. The original ASM method developed by Cootes et al. highly relies on the initialization and the representation of the local structure of the facial features in the image. We use color information to improve the ASM approach for facial feature extraction. The color information is used to localize the centers of the mouth and the eyes to assist the initialization step. Moreover, we model the local structure of the feature points in the RGB color space. Besides, we use 2D affine transformation to align facial features that are perturbed by head pose variations. In fact, the 2D affine transformation compensates for the effects of both head pose variations and the projection of 3D data to 2D. Experiments on a face database of 50 subjects show that our approach outperforms the standard ASM and is successful in facial feature extraction","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116083109","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 66
Human action recognition using multi-view image sequences 基于多视图图像序列的人体动作识别
Mohiudding Ahmad, Seong-Whan Lee
Recognizing human action from image sequences is an active area of research in computer vision. In this paper, we present a novel method for human action recognition from image sequences in different viewing angles that uses the Cartesian component of optical flow velocity and human body shape feature vector information. We use principal component analysis to reduce the higher dimensional shape feature space into low dimensional shape feature space. We represent each action using a set of multidimensional discrete hidden Markov model and model each action for any viewing direction. We performed experiments of the proposed method by using KU gesture database. Experimental results based on this database of different actions show that our method is robust
从图像序列中识别人类行为是计算机视觉研究的一个活跃领域。本文提出了一种利用光流速度的笛卡尔分量和人体形状特征向量信息对不同视角图像序列进行人体动作识别的新方法。利用主成分分析将高维形状特征空间降为低维形状特征空间。我们使用一组多维离散隐马尔可夫模型来表示每个动作,并对每个动作进行任意观察方向的建模。我们利用KU手势数据库对该方法进行了实验。实验结果表明,该方法具有较好的鲁棒性
{"title":"Human action recognition using multi-view image sequences","authors":"Mohiudding Ahmad, Seong-Whan Lee","doi":"10.1109/FGR.2006.65","DOIUrl":"https://doi.org/10.1109/FGR.2006.65","url":null,"abstract":"Recognizing human action from image sequences is an active area of research in computer vision. In this paper, we present a novel method for human action recognition from image sequences in different viewing angles that uses the Cartesian component of optical flow velocity and human body shape feature vector information. We use principal component analysis to reduce the higher dimensional shape feature space into low dimensional shape feature space. We represent each action using a set of multidimensional discrete hidden Markov model and model each action for any viewing direction. We performed experiments of the proposed method by using KU gesture database. Experimental results based on this database of different actions show that our method is robust","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132574206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
Automatic feature extraction for multiview 3D face recognition 多视角三维人脸识别的自动特征提取
Xiaoguang Lu, Anil K. Jain
Current 2D face recognition systems encounter difficulties in recognizing faces with large pose variations. Utilizing the pose-invariant features of 3D face data has the potential to handle multiview face matching. A feature extractor based on the directional maximum is proposed to estimate the nose tip location and the pose angle simultaneously. A nose profile model represented by subspaces is used to select the best candidates for the nose tip. Assisted by a statistical feature location model, a multimodal scheme is presented to extract eye and mouth corners. Using the automatic feature extractor, a fully automatic 3D face recognition system is developed. The system is evaluated on two databases, the MSU database (300 multiview test scans from 100 subjects) and the UND database (953 near frontal scans from 277 subjects). The automatic system provides recognition accuracy that is comparable to the accuracy of a system with manually labeled feature points
目前的二维人脸识别系统在识别姿态变化较大的人脸时遇到了困难。利用三维人脸数据的位姿不变特征,有可能处理多视图人脸匹配。提出了一种基于方向最大值的特征提取器,用于同时估计鼻尖位置和姿态角。采用子空间表示的鼻廓模型来选择鼻尖的最佳候选者。在统计特征定位模型的辅助下,提出了一种多模态的眼、嘴角提取方案。利用自动特征提取器,开发了一个全自动三维人脸识别系统。该系统在两个数据库上进行评估,MSU数据库(来自100名受试者的300次多视图测试扫描)和UND数据库(来自277名受试者的953次近正面扫描)。自动系统提供的识别精度与手动标记特征点的系统的精度相当
{"title":"Automatic feature extraction for multiview 3D face recognition","authors":"Xiaoguang Lu, Anil K. Jain","doi":"10.1109/FGR.2006.23","DOIUrl":"https://doi.org/10.1109/FGR.2006.23","url":null,"abstract":"Current 2D face recognition systems encounter difficulties in recognizing faces with large pose variations. Utilizing the pose-invariant features of 3D face data has the potential to handle multiview face matching. A feature extractor based on the directional maximum is proposed to estimate the nose tip location and the pose angle simultaneously. A nose profile model represented by subspaces is used to select the best candidates for the nose tip. Assisted by a statistical feature location model, a multimodal scheme is presented to extract eye and mouth corners. Using the automatic feature extractor, a fully automatic 3D face recognition system is developed. The system is evaluated on two databases, the MSU database (300 multiview test scans from 100 subjects) and the UND database (953 near frontal scans from 277 subjects). The automatic system provides recognition accuracy that is comparable to the accuracy of a system with manually labeled feature points","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"253 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132759517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 142
A layered deformable model for gait analysis 一种用于步态分析的分层变形模型
Haiping Lu, K. Plataniotis, A. Venetsanopoulos
In this paper, a layered deformable model (LDM) is proposed for human body pose recovery in gait analysis. This model is inspired by the manually labeled silhouettes in (Z. Liu, et al., July 2004) and it is designed to closely match them. For fronto-parallel gait, the introduced LDM model defines the body part widths and lengths, the position and the joint angles of human body using 22 parameters. The model consists of four layers and allows for limb deformation. With this model, our objective is to recover its parameters (and thus the human body pose) from automatically extracted silhouettes. LDM recovery algorithm is first developed for manual silhouettes, in order to generate ground truth sequences for comparison and useful statistics regarding the LDM parameters. It is then extended for automatically extracted silhouettes. The proposed methodologies have been tested on 10005 frames from 285 gait sequences captured under various conditions and an average error rate of 7% is achieved for the lower limb joint angles of all the frames, showing great potential for model-based gait recognition
提出了一种用于步态分析中人体姿态恢复的分层变形模型(LDM)。该模型的灵感来自于(Z. Liu, et al., July 2004)中手工标记的轮廓,并被设计为与它们紧密匹配。对于前平行步态,引入的LDM模型使用22个参数定义了人体部位的宽度和长度、位置和关节角度。该模型由四层组成,并允许肢体变形。有了这个模型,我们的目标是从自动提取的轮廓中恢复其参数(从而恢复人体姿势)。LDM恢复算法首先是针对人工轮廓而开发的,目的是生成地面真值序列,用于LDM参数的比较和有用的统计。然后将其扩展为自动提取的轮廓。在各种条件下对285个步态序列中的10005帧进行了测试,所有帧的下肢关节角度的平均错误率为7%,显示出基于模型的步态识别的巨大潜力
{"title":"A layered deformable model for gait analysis","authors":"Haiping Lu, K. Plataniotis, A. Venetsanopoulos","doi":"10.1109/FGR.2006.11","DOIUrl":"https://doi.org/10.1109/FGR.2006.11","url":null,"abstract":"In this paper, a layered deformable model (LDM) is proposed for human body pose recovery in gait analysis. This model is inspired by the manually labeled silhouettes in (Z. Liu, et al., July 2004) and it is designed to closely match them. For fronto-parallel gait, the introduced LDM model defines the body part widths and lengths, the position and the joint angles of human body using 22 parameters. The model consists of four layers and allows for limb deformation. With this model, our objective is to recover its parameters (and thus the human body pose) from automatically extracted silhouettes. LDM recovery algorithm is first developed for manual silhouettes, in order to generate ground truth sequences for comparison and useful statistics regarding the LDM parameters. It is then extended for automatically extracted silhouettes. The proposed methodologies have been tested on 10005 frames from 285 gait sequences captured under various conditions and an average error rate of 7% is achieved for the lower limb joint angles of all the frames, showing great potential for model-based gait recognition","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"335 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133215871","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
期刊
7th International Conference on Automatic Face and Gesture Recognition (FGR06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1