首页 > 最新文献

7th International Conference on Automatic Face and Gesture Recognition (FGR06)最新文献

英文 中文
Adding holistic dimensions to a facial composite system 为面部合成系统增加整体尺寸
C. Frowd, V. Bruce, A. McIntyre, P. Hancock
Facial composites are typically constructed by witnesses to crime by describing a suspect's face and then selecting facial features from a kit of parts. Unfortunately, when produced in this way, composites are very poorly identified. In contrast, there is mounting evidence that other, more recognition-based approaches can produce a much better likeness of a suspect. With the EvoFIT system, for example, witnesses are presented with sets of complete faces and a composite is `evolved' through a process of selection and breeding. The current work serves to augment EvoFIT by developing a set of psychologically useful `knobs' that allow faces to be manipulated along dimensions such as facial weight, masculinity, and age. These holistic dimensions were implemented by increasing the size and variability of the underlying face model and obtaining perceptual ratings so that the space could be suitably vectorised. Two evaluations suggested that the new dimensions were operating appropriately
面部合成材料通常是由犯罪证人通过描述嫌疑人的面部,然后从一组部件中选择面部特征来构建的。不幸的是,当以这种方式生产时,复合材料的识别非常差。相比之下,越来越多的证据表明,其他更基于识别的方法可以产生更像嫌疑人的图像。例如,在EvoFIT系统中,目击者可以看到一组完整的面孔,并通过选择和繁殖过程“进化”出一种合成物。目前的工作是通过开发一套心理上有用的“旋钮”来增强EvoFIT,这些“旋钮”可以根据面部体重、男子气概和年龄等维度来操纵面部。这些整体维度是通过增加底层人脸模型的大小和可变性来实现的,并获得感知评级,以便对空间进行适当的矢量化。两项评价表明,新的层面运作良好
{"title":"Adding holistic dimensions to a facial composite system","authors":"C. Frowd, V. Bruce, A. McIntyre, P. Hancock","doi":"10.1109/FGR.2006.20","DOIUrl":"https://doi.org/10.1109/FGR.2006.20","url":null,"abstract":"Facial composites are typically constructed by witnesses to crime by describing a suspect's face and then selecting facial features from a kit of parts. Unfortunately, when produced in this way, composites are very poorly identified. In contrast, there is mounting evidence that other, more recognition-based approaches can produce a much better likeness of a suspect. With the EvoFIT system, for example, witnesses are presented with sets of complete faces and a composite is `evolved' through a process of selection and breeding. The current work serves to augment EvoFIT by developing a set of psychologically useful `knobs' that allow faces to be manipulated along dimensions such as facial weight, masculinity, and age. These holistic dimensions were implemented by increasing the size and variability of the underlying face model and obtaining perceptual ratings so that the space could be suitably vectorised. Two evaluations suggested that the new dimensions were operating appropriately","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132568729","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
A new look at filtering techniques for illumination invariance in automatic face recognition 人脸自动识别中光照不变性滤波技术的新研究
Ognjen Arandjelovic, R. Cipolla
Illumination invariance remains the most researched, yet the most challenging aspect of automatic face recognition. In this paper we propose a novel, general recognition framework for efficient matching of individual face images, sets or sequences. The framework is based on simple image processing filters that compete with unprocessed greyscale input to yield a single matching score between individuals. It is shown how the discrepancy between illumination conditions between novel input and the training data set can be estimated and used to weigh the contribution of two competing representations. We describe an extensive empirical evaluation of the proposed method on 171 individuals and over 1300 video sequences with extreme illumination, pose and head motion variation. On this challenging data set our algorithm consistently demonstrated a dramatic performance improvement over traditional filtering approaches. We demonstrate a reduction of 50-75% in recognition error rates, the best performing method-filter combination correctly recognizing 96% of the individuals
光照不变性是人脸自动识别中研究最多,但也是最具挑战性的一个方面。在本文中,我们提出了一种新的,通用的识别框架,用于有效匹配单个人脸图像,集合或序列。该框架基于简单的图像处理过滤器,这些过滤器与未处理的灰度输入竞争,以产生个体之间的单个匹配分数。它展示了如何估计新输入和训练数据集之间的照明条件之间的差异,并用于权衡两个竞争表示的贡献。我们对171个个体和1300多个具有极端照明、姿势和头部运动变化的视频序列进行了广泛的实证评估。在这个具有挑战性的数据集上,我们的算法始终比传统的过滤方法表现出显著的性能改进。我们证明了识别错误率降低了50-75%,表现最好的方法-滤波器组合正确识别了96%的个体
{"title":"A new look at filtering techniques for illumination invariance in automatic face recognition","authors":"Ognjen Arandjelovic, R. Cipolla","doi":"10.1109/FGR.2006.14","DOIUrl":"https://doi.org/10.1109/FGR.2006.14","url":null,"abstract":"Illumination invariance remains the most researched, yet the most challenging aspect of automatic face recognition. In this paper we propose a novel, general recognition framework for efficient matching of individual face images, sets or sequences. The framework is based on simple image processing filters that compete with unprocessed greyscale input to yield a single matching score between individuals. It is shown how the discrepancy between illumination conditions between novel input and the training data set can be estimated and used to weigh the contribution of two competing representations. We describe an extensive empirical evaluation of the proposed method on 171 individuals and over 1300 video sequences with extreme illumination, pose and head motion variation. On this challenging data set our algorithm consistently demonstrated a dramatic performance improvement over traditional filtering approaches. We demonstrate a reduction of 50-75% in recognition error rates, the best performing method-filter combination correctly recognizing 96% of the individuals","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130228942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
A realtime shrug detector 一个实时耸肩探测器
Huazhong Ning, T. Han, Yuxiao Hu, ZhenQiu Zhang, Yun Fu, Thomas S. Huang
A realtime system for shrug detection is discussed in this paper. The system is automatically initialized by a face detector based on Ada-boost [P. Viola and M. Jones, May 2004]. After frontal face is localized by the face detector, shoulder position is detected by fitting a parabola to the nearby horizontal edges using weighted Hough transform [K. Sugawara, 1997]. Since shrug is an action which is defined not only by the distance between face and shoulder but also the relative temporal-spatial changing between them, we propose a parameterizing scheme using two different parabolas, named as "stable parabola" (SP) and "transient parabola" (TP) to characterize the action shrug. Stable parabola represents the mean shoulder position over a long time duration, while transient parabola represents the mean shoulder position of a very short time duration. By using this scheme (only 6 dimensions), we avoid the high dimensional representation of the temporal process-shrug, and therefore make the realtime implementation possible. The shrug detector is then trained in the parameter space using Fisher discriminant analysis (FDA). The experiments show that the proposed shrug detector is able to not only detect the shrug action correctly and efficiently (in realtime), but also tolerate the large in-class variation caused by different subject, different action speed, illumination, partial occlusion, and background clutter. So the proposed realtime shrug detector is promising in video analysis under an uncontrolled environment
本文讨论了一种实时耸肩检测系统。系统由基于Ada-boost的人脸检测器自动初始化[P]。Viola and M. Jones, 2004年5月]。人脸检测器对正面人脸进行定位后,利用加权霍夫变换[K]对附近水平边缘拟合抛物线来检测肩部位置。Sugawara, 1997]。由于耸耸肩是一个动作,它不仅是由脸和肩膀之间的距离定义的,而且是它们之间的相对时空变化,我们提出了一个参数化方案,使用两种不同的抛物线,称为“稳定抛物线”(SP)和“瞬态抛物线”(TP)来表征动作耸耸肩。稳定抛物线表示长时间内肩膀的平均位置,而瞬态抛物线表示很短时间内肩膀的平均位置。通过使用这种方案(只有6个维度),我们避免了时态过程耸耸肩的高维表示,因此使实时实现成为可能。然后使用Fisher判别分析(FDA)在参数空间中训练耸肩检测器。实验表明,所提出的耸肩检测器不仅能够正确有效地检测耸肩动作(实时),而且能够容忍不同主体、不同动作速度、光照、局部遮挡和背景杂波等引起的类内较大变化。因此,所提出的实时耸肩检测器在非受控环境下的视频分析中具有广阔的应用前景
{"title":"A realtime shrug detector","authors":"Huazhong Ning, T. Han, Yuxiao Hu, ZhenQiu Zhang, Yun Fu, Thomas S. Huang","doi":"10.1109/FGR.2006.15","DOIUrl":"https://doi.org/10.1109/FGR.2006.15","url":null,"abstract":"A realtime system for shrug detection is discussed in this paper. The system is automatically initialized by a face detector based on Ada-boost [P. Viola and M. Jones, May 2004]. After frontal face is localized by the face detector, shoulder position is detected by fitting a parabola to the nearby horizontal edges using weighted Hough transform [K. Sugawara, 1997]. Since shrug is an action which is defined not only by the distance between face and shoulder but also the relative temporal-spatial changing between them, we propose a parameterizing scheme using two different parabolas, named as \"stable parabola\" (SP) and \"transient parabola\" (TP) to characterize the action shrug. Stable parabola represents the mean shoulder position over a long time duration, while transient parabola represents the mean shoulder position of a very short time duration. By using this scheme (only 6 dimensions), we avoid the high dimensional representation of the temporal process-shrug, and therefore make the realtime implementation possible. The shrug detector is then trained in the parameter space using Fisher discriminant analysis (FDA). The experiments show that the proposed shrug detector is able to not only detect the shrug action correctly and efficiently (in realtime), but also tolerate the large in-class variation caused by different subject, different action speed, illumination, partial occlusion, and background clutter. So the proposed realtime shrug detector is promising in video analysis under an uncontrolled environment","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"27 5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132116544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Learning sparse features in granular space for multi-view face detection 学习颗粒空间中的稀疏特征用于多视图人脸检测
Chang Huang, H. Ai, Yuan Li, S. Lao
In this paper, a novel sparse feature set is introduced into the Adaboost learning framework for multi-view face detection (MVFD), and a learning algorithm based on heuristic search is developed to select sparse features in granular space. Compared with Haar-like features, sparse features are more generic and powerful to characterize multi-view face pattern that is more diverse and asymmetric than frontal face pattern. In order to cut down search space to a manageable size, we propose a multi-scaled search algorithm that is about 6 times faster than brute-force search. With this method, a MVFD system is implemented that covers face pose changes over +/-45deg rotation in plane (RIP) and +/-90deg rotation off plane (ROP). Experiments over well-know test set are reported to show its high performance in both accuracy and speed
本文在Adaboost多视图人脸检测(MVFD)学习框架中引入了一种新的稀疏特征集,并开发了一种基于启发式搜索的学习算法来选择颗粒空间中的稀疏特征。与Haar-like特征相比,稀疏特征在多视图人脸特征中更为通用和强大,比正面人脸特征更具多样性和非对称性。为了将搜索空间缩小到可管理的大小,我们提出了一种比暴力搜索快6倍的多尺度搜索算法。通过这种方法,实现了一个MVFD系统,该系统可以覆盖平面内+/-45度旋转(RIP)和平面外+/-90度旋转(ROP)时面部姿态的变化。在知名的测试集上进行的实验表明,该方法在精度和速度上都有很高的性能
{"title":"Learning sparse features in granular space for multi-view face detection","authors":"Chang Huang, H. Ai, Yuan Li, S. Lao","doi":"10.1109/FGR.2006.70","DOIUrl":"https://doi.org/10.1109/FGR.2006.70","url":null,"abstract":"In this paper, a novel sparse feature set is introduced into the Adaboost learning framework for multi-view face detection (MVFD), and a learning algorithm based on heuristic search is developed to select sparse features in granular space. Compared with Haar-like features, sparse features are more generic and powerful to characterize multi-view face pattern that is more diverse and asymmetric than frontal face pattern. In order to cut down search space to a manageable size, we propose a multi-scaled search algorithm that is about 6 times faster than brute-force search. With this method, a MVFD system is implemented that covers face pose changes over +/-45deg rotation in plane (RIP) and +/-90deg rotation off plane (ROP). Experiments over well-know test set are reported to show its high performance in both accuracy and speed","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133867121","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
Cascaded classification of gender and facial expression using active appearance models 使用主动外观模型的性别和面部表情级联分类
Yunus Saatci, C. Town
This paper presents an approach to recognising the gender and expression of face images by means of active appearance models (AAM). Features extracted by a trained AAM are used to construct support vector machine (SVM) classifiers for 4 elementary emotional states (happy, angry, sad, neutral). These classifiers are arranged into a cascade structure in order to optimise overall recognition performance. Furthermore, it is shown how performance can be further improved by first classifying the gender of the face images using an SVM trained in a similar manner. Both gender-specific expression classification and expression-specific gender classification cascades are considered, with the former yielding better recognition performance. We conclude that there are gender-specific differences in the appearance of facial expressions that can be exploited for automated recognition, and that cascades are an efficient and effective way of performing multi-class recognition of facial expressions
本文提出了一种基于主动外观模型(AAM)的人脸图像性别和表情识别方法。使用训练好的AAM提取的特征构建支持向量机(SVM)分类器,用于4种基本情绪状态(快乐、愤怒、悲伤、中性)。这些分类器排列成级联结构,以优化整体识别性能。此外,还展示了如何通过使用以类似方式训练的支持向量机首先对人脸图像的性别进行分类来进一步提高性能。基于性别的表达分类和基于表达的性别分类级联都被考虑在内,前者具有更好的识别性能。我们得出的结论是,面部表情的外观存在性别差异,可以用于自动识别,并且级联是执行多类别面部表情识别的高效和有效的方法
{"title":"Cascaded classification of gender and facial expression using active appearance models","authors":"Yunus Saatci, C. Town","doi":"10.1109/FGR.2006.29","DOIUrl":"https://doi.org/10.1109/FGR.2006.29","url":null,"abstract":"This paper presents an approach to recognising the gender and expression of face images by means of active appearance models (AAM). Features extracted by a trained AAM are used to construct support vector machine (SVM) classifiers for 4 elementary emotional states (happy, angry, sad, neutral). These classifiers are arranged into a cascade structure in order to optimise overall recognition performance. Furthermore, it is shown how performance can be further improved by first classifying the gender of the face images using an SVM trained in a similar manner. Both gender-specific expression classification and expression-specific gender classification cascades are considered, with the former yielding better recognition performance. We conclude that there are gender-specific differences in the appearance of facial expressions that can be exploited for automated recognition, and that cascades are an efficient and effective way of performing multi-class recognition of facial expressions","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129516394","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 159
Video-based face recognition evaluation in the CHIL project - Run 1 基于视频的人脸识别评估CHIL项目-运行1
H. K. Ekenel, Aristodemos Pnevmatikakis
This paper describes the video-based face recognition evaluation performed under the CHIL project and the systems that participated to it, along with the obtained first year results. The evaluation methodology comprises a specially built database of videos and an evaluation protocol. Two complete automatic face detection and recognition systems from two academic institutions participated to the evaluation. For comparison purposes, a baseline system is also developed using well-known methods for face detection and recognition
本文描述了在CHIL项目下进行的基于视频的人脸识别评估以及参与该项目的系统,以及获得的第一年结果。评价方法包括专门建立的录像数据库和评价规程。来自两个学术机构的两个完整的自动人脸检测和识别系统参与了评估。为了进行比较,我们还使用众所周知的人脸检测和识别方法开发了一个基线系统
{"title":"Video-based face recognition evaluation in the CHIL project - Run 1","authors":"H. K. Ekenel, Aristodemos Pnevmatikakis","doi":"10.1109/FGR.2006.110","DOIUrl":"https://doi.org/10.1109/FGR.2006.110","url":null,"abstract":"This paper describes the video-based face recognition evaluation performed under the CHIL project and the systems that participated to it, along with the obtained first year results. The evaluation methodology comprises a specially built database of videos and an evaluation protocol. Two complete automatic face detection and recognition systems from two academic institutions participated to the evaluation. For comparison purposes, a baseline system is also developed using well-known methods for face detection and recognition","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127098320","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Robust distance measures for face-recognition supporting revocable biometric tokens 支持可撤销生物特征令牌的人脸识别鲁棒距离度量
T. Boult
This paper explores a form of robust distance measures for biometrics and presents experiments showing that, when applied per "class" they can dramatically improve the accuracy of face recognition. We "robustify'' many distance measures included in the CSU face-recognition toolkit, and apply them to PCA, LDA and EBGM. The resulting performance puts each of these algorithms, for the FERET datasets tested, on par with commercial face recognition results. Unlike passwords, biometric signatures cannot be changed or revoked. This paper shows how the robust distance measures introduce can be used for secure robust revocable biometrics. The technique produces what we call Biotopestrade, which provide public-key cryptographic security, supports matching in encoded form, cannot be linked across different databases and are revocable. Biotopes support a robust distance measure computed on the encoded form that is proven not to decrease, and that may potentially increase, accurately. The approach is demonstrated, to improve performance beyond the already impressive gains from the robust distance measure
本文探讨了一种用于生物识别的鲁棒距离测量形式,并提出了实验表明,当应用于每个“类”时,它们可以显着提高人脸识别的准确性。我们“鲁棒化”了CSU人脸识别工具包中包含的许多距离度量,并将它们应用于PCA, LDA和EBGM。结果表明,对于测试的FERET数据集,这些算法的性能与商业人脸识别结果相当。与密码不同,生物特征签名不能更改或撤销。本文展示了引入的鲁棒距离度量如何用于安全的鲁棒可撤销生物识别。这种技术产生了我们所说的Biotopestrade,它提供了公开密钥加密安全性,支持编码形式的匹配,不能跨不同数据库链接,并且是可撤销的。生物群落支持一种基于编码形式计算的可靠的距离测量,这种距离测量被证明不会减少,而且可能会准确地增加。该方法已被证明,以提高性能超出已经令人印象深刻的增益从鲁棒距离测量
{"title":"Robust distance measures for face-recognition supporting revocable biometric tokens","authors":"T. Boult","doi":"10.1109/FGR.2006.94","DOIUrl":"https://doi.org/10.1109/FGR.2006.94","url":null,"abstract":"This paper explores a form of robust distance measures for biometrics and presents experiments showing that, when applied per \"class\" they can dramatically improve the accuracy of face recognition. We \"robustify'' many distance measures included in the CSU face-recognition toolkit, and apply them to PCA, LDA and EBGM. The resulting performance puts each of these algorithms, for the FERET datasets tested, on par with commercial face recognition results. Unlike passwords, biometric signatures cannot be changed or revoked. This paper shows how the robust distance measures introduce can be used for secure robust revocable biometrics. The technique produces what we call Biotopestrade, which provide public-key cryptographic security, supports matching in encoded form, cannot be linked across different databases and are revocable. Biotopes support a robust distance measure computed on the encoded form that is proven not to decrease, and that may potentially increase, accurately. The approach is demonstrated, to improve performance beyond the already impressive gains from the robust distance measure","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132861995","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 111
Tracking using dynamic programming for appearance-based sign language recognition 基于外观的手语识别的动态规划跟踪
P. Dreuw, Thomas Deselaers, David Rybach, Daniel Keysers, H. Ney
We present a novel tracking algorithm that uses dynamic programming to determine the path of target objects and that is able to track an arbitrary number of different objects. The traceback method used to track the targets avoids taking possibly wrong local decisions and thus reconstructs the best tracking paths using the whole observation sequence. The tracking method can be compared to the nonlinear time alignment in automatic speech recognition (ASR) and it can analogously be integrated into a hidden Markov model based recognition process. We show how the method can be applied to the tracking of hands and the face for automatic sign language recognition
我们提出了一种新的跟踪算法,该算法使用动态规划来确定目标物体的路径,并且能够跟踪任意数量的不同物体。采用回溯法对目标进行跟踪,避免了可能出现的局部决策错误,从而利用整个观测序列重构出最佳跟踪路径。该跟踪方法可与自动语音识别(ASR)中的非线性时间对准进行比较,并可类比地集成到基于隐马尔可夫模型的识别过程中。我们展示了如何将该方法应用于手部和面部的跟踪,以实现自动手语识别
{"title":"Tracking using dynamic programming for appearance-based sign language recognition","authors":"P. Dreuw, Thomas Deselaers, David Rybach, Daniel Keysers, H. Ney","doi":"10.1109/FGR.2006.107","DOIUrl":"https://doi.org/10.1109/FGR.2006.107","url":null,"abstract":"We present a novel tracking algorithm that uses dynamic programming to determine the path of target objects and that is able to track an arbitrary number of different objects. The traceback method used to track the targets avoids taking possibly wrong local decisions and thus reconstructs the best tracking paths using the whole observation sequence. The tracking method can be compared to the nonlinear time alignment in automatic speech recognition (ASR) and it can analogously be integrated into a hidden Markov model based recognition process. We show how the method can be applied to the tracking of hands and the face for automatic sign language recognition","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114163299","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 68
Making recognisable faces 做出可识别的面孔
D. Chatting
When delivering visual content on multiple devices and services, faces can often become unrecognisable. This paper draws together research from across the cognitive psychology literature to argue that faces should be treated as a special case when rendering content. Where available, we suggest methods by which recognition can be imp roved within the constraints of the device and service. Firstly, we review the psychology literature to discuss recognition performance when manipulating the face's scale, colour palette, orientation and motion. Secondly, we consider how characteristics of the individual faces can aide or hinder recognition and how caricature may be applied, especially within crowds, to improve it. Thirdly, we show how context can make even the most abstract faces recognisable. Fourthly, we highlight the challenges of making a good portrait, beyond the criteria of simply being recognisable. Finally, we begin to describe a framework for automatically rendering faces 'smartly', such that they will be most recognisable given the device and service of which they are a part
当在多个设备和服务上传递视觉内容时,人脸通常会变得无法识别。本文汇集了来自认知心理学文献的研究,认为在呈现内容时应将面部视为特殊情况。在可行的情况下,我们建议在设备和服务的限制下改进识别的方法。首先,我们回顾了心理学文献,讨论了人脸尺度、调色板、方向和运动对识别性能的影响。其次,我们考虑个人面部特征如何帮助或阻碍识别,以及如何应用漫画,特别是在人群中,以改善识别。第三,我们展示了上下文如何使即使是最抽象的面孔也能被识别。第四,我们强调了制作一幅好肖像的挑战,超越了简单的可识别性标准。最后,我们开始描述一个“智能”自动渲染人脸的框架,这样他们将是最可识别的,因为他们是其中的一部分的设备和服务
{"title":"Making recognisable faces","authors":"D. Chatting","doi":"10.1109/FGR.2006.76","DOIUrl":"https://doi.org/10.1109/FGR.2006.76","url":null,"abstract":"When delivering visual content on multiple devices and services, faces can often become unrecognisable. This paper draws together research from across the cognitive psychology literature to argue that faces should be treated as a special case when rendering content. Where available, we suggest methods by which recognition can be imp roved within the constraints of the device and service. Firstly, we review the psychology literature to discuss recognition performance when manipulating the face's scale, colour palette, orientation and motion. Secondly, we consider how characteristics of the individual faces can aide or hinder recognition and how caricature may be applied, especially within crowds, to improve it. Thirdly, we show how context can make even the most abstract faces recognisable. Fourthly, we highlight the challenges of making a good portrait, beyond the criteria of simply being recognisable. Finally, we begin to describe a framework for automatically rendering faces 'smartly', such that they will be most recognisable given the device and service of which they are a part","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123533793","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Vision Based Interface for Local Collaborative Music Synthesis 基于视觉的本地协同音乐合成界面
João Carreira, P. Peixoto
The computer is an ubiquitous element of modern society, nonetheless, human computer interaction is still rather inflexible. Particularly in local collaborative environments, like office meetings, the property that the mouse and keyboard exhibit of being a gateway for the individual to act upon a workspace, makes local computer mediated collaboration uncomfortable, as users have to time-share their actions upon that workspace. We present in this paper a novel, very fast, vision based interface that allows multiple users to interact simultaneously with a single computer by performing hand gestures, which are filmed by a static video camera. This interface attempts to continuously recognize predefined postures and movements using a view-dependent method. We also present A.C.O, an application which receives input from the vision based interface and allows users around a table to collaborate playing synthesized music instruments by moving their hands
计算机是现代社会中无处不在的元素,然而,人机交互仍然相当不灵活。特别是在本地协作环境中,比如办公室会议,鼠标和键盘作为个人对工作空间进行操作的门户的特性,使得本地计算机中介的协作不舒服,因为用户必须在该工作空间上分时共享他们的操作。我们在这篇论文中提出了一个新颖的,非常快速的,基于视觉的界面,允许多个用户通过执行手势同时与一台计算机进行交互,这些手势由静态摄像机拍摄。该接口尝试使用依赖于视图的方法连续识别预定义的姿势和动作。我们还介绍了A.C.O,这是一个应用程序,它可以从基于视觉的界面接收输入,并允许围坐在桌子周围的用户通过移动他们的手来协作演奏合成乐器
{"title":"A Vision Based Interface for Local Collaborative Music Synthesis","authors":"João Carreira, P. Peixoto","doi":"10.1109/FGR.2006.16","DOIUrl":"https://doi.org/10.1109/FGR.2006.16","url":null,"abstract":"The computer is an ubiquitous element of modern society, nonetheless, human computer interaction is still rather inflexible. Particularly in local collaborative environments, like office meetings, the property that the mouse and keyboard exhibit of being a gateway for the individual to act upon a workspace, makes local computer mediated collaboration uncomfortable, as users have to time-share their actions upon that workspace. We present in this paper a novel, very fast, vision based interface that allows multiple users to interact simultaneously with a single computer by performing hand gestures, which are filmed by a static video camera. This interface attempts to continuously recognize predefined postures and movements using a view-dependent method. We also present A.C.O, an application which receives input from the vision based interface and allows users around a table to collaborate playing synthesized music instruments by moving their hands","PeriodicalId":109260,"journal":{"name":"7th International Conference on Automatic Face and Gesture Recognition (FGR06)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-04-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125678126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
7th International Conference on Automatic Face and Gesture Recognition (FGR06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1