首页 > 最新文献

2015 International Conference on Biometrics (ICB)最新文献

英文 中文
Quaternion-based Local Binary Patterns for illumination invariant face recognition 基于四元数的局部二值模式光照不变人脸识别
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139093
Dayron Rizo-Rodriguez, D. Ziou
Face recognition under varying illumination conditions has benefited from the encapsulation of face features into a quaternion representation. In particular, a quaternion-based representation encoding four LBP descriptors (one-pixel to four-pixel radius) has delivered interesting recognition scores when using just one training image per subject. However, each coefficient of such a representation only encapsulates the eight sampling values required for computing a classic LBP code. In this paper, we propose a quaternion representation which encodes additional LBP codes at each radius. Consequently, a wider pixel descriptor is obtained because further sampling values are considered in the region surrounding the reference pixel. Illumination invariant face verification and identification experiments are conducted by using only one training face image. The representation proposed improves the recognition rates reported by the representation only encapsulating classic LBP codes.
不同光照条件下的人脸识别得益于将人脸特征封装成四元数表示。特别是,编码四个LBP描述符(1像素到4像素半径)的基于四元数的表示在每个主题只使用一个训练图像时提供了有趣的识别分数。然而,这种表示的每个系数只封装了计算经典LBP代码所需的8个采样值。在本文中,我们提出了一种四元数表示,它在每个半径上编码额外的LBP码。因此,由于在参考像素周围的区域中考虑了进一步的采样值,因此获得了更宽的像素描述符。仅使用一张训练人脸图像进行光照不变人脸验证与识别实验。所提出的表示提高了仅封装经典LBP代码的表示报告的识别率。
{"title":"Quaternion-based Local Binary Patterns for illumination invariant face recognition","authors":"Dayron Rizo-Rodriguez, D. Ziou","doi":"10.1109/ICB.2015.7139093","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139093","url":null,"abstract":"Face recognition under varying illumination conditions has benefited from the encapsulation of face features into a quaternion representation. In particular, a quaternion-based representation encoding four LBP descriptors (one-pixel to four-pixel radius) has delivered interesting recognition scores when using just one training image per subject. However, each coefficient of such a representation only encapsulates the eight sampling values required for computing a classic LBP code. In this paper, we propose a quaternion representation which encodes additional LBP codes at each radius. Consequently, a wider pixel descriptor is obtained because further sampling values are considered in the region surrounding the reference pixel. Illumination invariant face verification and identification experiments are conducted by using only one training face image. The representation proposed improves the recognition rates reported by the representation only encapsulating classic LBP codes.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130403258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Towards Bloom filter-based indexing of iris biometric data 基于Bloom滤波的虹膜生物特征数据索引研究
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139105
C. Rathgeb, Harald Baier, C. Busch, Frank Breitinger
Conventional biometric identification systems require exhaustive 1 : N comparisons in order to identify biometric probes, i.e. comparison time frequently dominates the overall computational workload. Biometric database indexing represents a challenging task since biometric data is fuzzy and does not exhibit any natural sorting order. In this paper we present a preliminary study on the feasibility of applying Bloom filters for the purpose of iris biometric database indexing. It is shown, that by constructing a binary tree data structure of Bloom filters extracted from binary iris biometric templates (iris-codes) the search space can be reduced to O(logN). In experiments, which are carried out on a database of N = 256 classes, biometric performance (accuracy) is maintained for different conventional identification systems. Further, perspectives on how to employ the proposed scheme on large-scale databases are given.
传统的生物特征识别系统需要详尽的1:N比较来识别生物特征探针,即比较时间经常占总计算工作量的主导地位。生物特征数据库索引是一项具有挑战性的任务,因为生物特征数据是模糊的,并且没有表现出任何自然的排序顺序。本文对应用布隆滤波器进行虹膜生物特征数据库索引的可行性进行了初步研究。研究表明,通过构建从二进制虹膜生物特征模板(虹膜编码)中提取的Bloom滤波器的二叉树数据结构,可以将搜索空间缩减到O(logN)。在一个N = 256类的数据库上进行的实验中,不同的传统识别系统保持了生物识别性能(准确性)。此外,本文还对如何在大型数据库中应用所提出的方案进行了展望。
{"title":"Towards Bloom filter-based indexing of iris biometric data","authors":"C. Rathgeb, Harald Baier, C. Busch, Frank Breitinger","doi":"10.1109/ICB.2015.7139105","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139105","url":null,"abstract":"Conventional biometric identification systems require exhaustive 1 : N comparisons in order to identify biometric probes, i.e. comparison time frequently dominates the overall computational workload. Biometric database indexing represents a challenging task since biometric data is fuzzy and does not exhibit any natural sorting order. In this paper we present a preliminary study on the feasibility of applying Bloom filters for the purpose of iris biometric database indexing. It is shown, that by constructing a binary tree data structure of Bloom filters extracted from binary iris biometric templates (iris-codes) the search space can be reduced to O(logN). In experiments, which are carried out on a database of N = 256 classes, biometric performance (accuracy) is maintained for different conventional identification systems. Further, perspectives on how to employ the proposed scheme on large-scale databases are given.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130692631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
VT-KFER: A Kinect-based RGBD+time dataset for spontaneous and non-spontaneous facial expression recognition 基于动作的RGBD+时间数据集用于自发和非自发面部表情识别
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139081
S. Aly, Andrea Trubanova, A. L. Abbott, S. White, A. Youssef
Human facial expressions have been extensively studied using 2D static images or 2D video sequences. The main limitations of 2D-based analysis are problems associated with large variations in pose and illumination. Therefore, an alternative is to utilize depth information, captured from 3D sensors, which is both pose and illumination invariant. The Kinect sensor is an inexpensive, portable, and fast way to capture the depth information. However, only a few researchers have utilized the Kinect sensor for the automatic recognition of facial expressions. This is partly due to the lack of a Kinect-based publicly available RGBD facial expression recognition (FER) dataset that contains the relevant facial expressions and their associated semantic labels. This paper addresses this problem by presenting the first publicly available RGBD+time facial expression recognition dataset using the Kinect 1.0 sensor in both scripted (acted) and unscripted (spontaneous) scenarios. Our fully annotated dataset includes seven expressions (happiness, sadness, surprise, disgust, fear, anger, and neutral) for 32 subjects (males and females) aged from 10 to 30 and with different skin tones. Both human and machine evaluation were conducted. Each scripted expression was ranked quantitatively by two research assistants in the Psychology department. Baseline machine evaluation resulted in average recognition accuracy levels of 60% and 58.3% for 6 expressions and 7 expressions recognition, respectively, when features from 2D and 3D data were combined.
人类的面部表情已经被广泛研究使用二维静态图像或二维视频序列。基于2d的分析的主要限制是与姿态和照明的大变化相关的问题。因此,另一种方法是利用从3D传感器捕获的深度信息,这是姿势和照明不变的。Kinect传感器是一种廉价、便携、快速获取深度信息的方法。然而,利用Kinect传感器自动识别面部表情的研究人员很少。这部分是由于缺乏一个基于kinect的公开可用的RGBD面部表情识别(FER)数据集,该数据集包含相关的面部表情及其相关的语义标签。本文提出了第一个公开可用的RGBD+时间面部表情识别数据集,该数据集使用Kinect 1.0传感器在脚本(表演)和非脚本(自发)场景中使用。我们的完整注释数据集包括32名年龄在10到30岁之间、肤色不同的受试者(男性和女性)的七种表情(快乐、悲伤、惊讶、厌恶、恐惧、愤怒和中性)。进行了人和机器评估。每个脚本表达由心理学系的两名研究助理进行定量排序。当结合2D和3D数据的特征时,基线机器评估的6个表情和7个表情识别的平均准确率分别为60%和58.3%。
{"title":"VT-KFER: A Kinect-based RGBD+time dataset for spontaneous and non-spontaneous facial expression recognition","authors":"S. Aly, Andrea Trubanova, A. L. Abbott, S. White, A. Youssef","doi":"10.1109/ICB.2015.7139081","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139081","url":null,"abstract":"Human facial expressions have been extensively studied using 2D static images or 2D video sequences. The main limitations of 2D-based analysis are problems associated with large variations in pose and illumination. Therefore, an alternative is to utilize depth information, captured from 3D sensors, which is both pose and illumination invariant. The Kinect sensor is an inexpensive, portable, and fast way to capture the depth information. However, only a few researchers have utilized the Kinect sensor for the automatic recognition of facial expressions. This is partly due to the lack of a Kinect-based publicly available RGBD facial expression recognition (FER) dataset that contains the relevant facial expressions and their associated semantic labels. This paper addresses this problem by presenting the first publicly available RGBD+time facial expression recognition dataset using the Kinect 1.0 sensor in both scripted (acted) and unscripted (spontaneous) scenarios. Our fully annotated dataset includes seven expressions (happiness, sadness, surprise, disgust, fear, anger, and neutral) for 32 subjects (males and females) aged from 10 to 30 and with different skin tones. Both human and machine evaluation were conducted. Each scripted expression was ranked quantitatively by two research assistants in the Psychology department. Baseline machine evaluation resulted in average recognition accuracy levels of 60% and 58.3% for 6 expressions and 7 expressions recognition, respectively, when features from 2D and 3D data were combined.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133611462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Composite sketch recognition via deep network - a transfer learning approach 基于深度网络的复合草图识别-一种迁移学习方法
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139092
Paritosh Mittal, Mayank Vatsa, Richa Singh
Sketch recognition is one of the integral components used by law enforcement agencies in solving crime. In recent past, software generated composite sketches are being preferred as they are more consistent and faster to construct than hand drawn sketches. Matching these composite sketches to face photographs is a complex task because the composite sketches are drawn based on the witness description and lack minute details which are present in photographs. This paper presents a novel algorithm for matching composite sketches with photographs using transfer learning with deep learning representation. In the proposed algorithm, first the deep learning architecture based facial representation is learned using large face database of photos and then the representation is updated using small problem-specific training database. Experiments are performed on the extended PRIP database and it is observed that the proposed algorithm outperforms recently proposed approach and a commercial face recognition system.
素描识别是执法机关破案工作中不可或缺的组成部分之一。在最近的过去,软件生成的合成草图是首选,因为它们比手绘草图更一致,更快地构建。将这些合成草图与人脸照片相匹配是一项复杂的任务,因为合成草图是根据证人的描述绘制的,缺乏照片中存在的微小细节。本文提出了一种基于深度学习表示的迁移学习的复合草图与照片匹配算法。在该算法中,首先使用大型照片人脸数据库学习基于深度学习架构的面部表征,然后使用小型特定问题训练数据库更新表征。在扩展的PRIP数据库上进行了实验,观察到所提出的算法优于最近提出的方法和商业人脸识别系统。
{"title":"Composite sketch recognition via deep network - a transfer learning approach","authors":"Paritosh Mittal, Mayank Vatsa, Richa Singh","doi":"10.1109/ICB.2015.7139092","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139092","url":null,"abstract":"Sketch recognition is one of the integral components used by law enforcement agencies in solving crime. In recent past, software generated composite sketches are being preferred as they are more consistent and faster to construct than hand drawn sketches. Matching these composite sketches to face photographs is a complex task because the composite sketches are drawn based on the witness description and lack minute details which are present in photographs. This paper presents a novel algorithm for matching composite sketches with photographs using transfer learning with deep learning representation. In the proposed algorithm, first the deep learning architecture based facial representation is learned using large face database of photos and then the representation is updated using small problem-specific training database. Experiments are performed on the extended PRIP database and it is observed that the proposed algorithm outperforms recently proposed approach and a commercial face recognition system.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114492410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
A longitudinal study of automatic face recognition 自动人脸识别的纵向研究
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139087
L. Best-Rowden, Anil K. Jain
With the deployment of automatic face recognition systems for many large-scale applications, it is crucial that we gain a thorough understanding of how facial aging affects the recognition performance, particularly across a large population. Because aging is a complex process involving genetic and environmental factors, some faces “age well” while the appearance of others changes drastically over time. This heterogeneity (inter-subject variability) suggests the need for a subject-specific aging analysis. In this paper, we conduct such an analysis using a longitudinal database of 147,784 operational mug shots of 18,007 repeat criminal offenders, where each subject has at least five face images acquired over a minimum of five years. By fitting multilevel statistical models to genuine similarity scores from two commercial-off-the-shelf (COTS) matchers, we quantify (i) the population average rate of change in genuine scores with respect to the elapsed time between two face images, and (ii) how closely the subject-specific rates of change follow the population average. Longitudinal analysis of the scores from the more accurate COTS matcher shows that despite decreasing genuine scores over time, the average subject can still be correctly verified at a false accept rate (FAR) of 0.01% across all 16 years of elapsed time in our database. We also investigate (i) the effects of several other covariates (gender, race, face quality), and (ii) the probability of true acceptance over time.
随着自动人脸识别系统在许多大规模应用中的部署,我们必须彻底了解面部老化如何影响识别性能,特别是在大量人口中。因为衰老是一个涉及遗传和环境因素的复杂过程,有些人的脸“很老”,而另一些人的脸则随着时间的推移发生了巨大的变化。这种异质性(受试者间的可变性)表明需要对受试者进行特定的衰老分析。在本文中,我们使用纵向数据库进行了这样的分析,该数据库包含18007名惯犯的147,784张操作面部照片,其中每个受试者至少有五张在至少五年内获得的面部图像。通过将多层统计模型拟合到两个商用现货(COTS)匹配器的真实相似分数,我们量化了(i)相对于两张人脸图像之间经过的时间,真实分数的总体平均变化率,以及(ii)受试者特定变化率与总体平均变化率的密切程度。对更精确的COTS匹配器得分的纵向分析表明,尽管真实得分随着时间的推移而下降,但在我们数据库中所有16年的运行时间中,平均受试者仍然可以以0.01%的错误接受率(FAR)正确验证。我们还研究了(i)其他几个协变量(性别、种族、面部质量)的影响,以及(ii)随着时间的推移,真实接受的概率。
{"title":"A longitudinal study of automatic face recognition","authors":"L. Best-Rowden, Anil K. Jain","doi":"10.1109/ICB.2015.7139087","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139087","url":null,"abstract":"With the deployment of automatic face recognition systems for many large-scale applications, it is crucial that we gain a thorough understanding of how facial aging affects the recognition performance, particularly across a large population. Because aging is a complex process involving genetic and environmental factors, some faces “age well” while the appearance of others changes drastically over time. This heterogeneity (inter-subject variability) suggests the need for a subject-specific aging analysis. In this paper, we conduct such an analysis using a longitudinal database of 147,784 operational mug shots of 18,007 repeat criminal offenders, where each subject has at least five face images acquired over a minimum of five years. By fitting multilevel statistical models to genuine similarity scores from two commercial-off-the-shelf (COTS) matchers, we quantify (i) the population average rate of change in genuine scores with respect to the elapsed time between two face images, and (ii) how closely the subject-specific rates of change follow the population average. Longitudinal analysis of the scores from the more accurate COTS matcher shows that despite decreasing genuine scores over time, the average subject can still be correctly verified at a false accept rate (FAR) of 0.01% across all 16 years of elapsed time in our database. We also investigate (i) the effects of several other covariates (gender, race, face quality), and (ii) the probability of true acceptance over time.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"531 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133432349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 80
Touch-interaction behavior for continuous user authentication on smartphones 在智能手机上进行连续用户认证的触摸交互行为
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139046
Chao Shen, Yong Zhang, Zhongmin Cai, Tianwen Yu, X. Guan
The increasing use of smartphones to access sensitive and privacy data has given rise to the need of secure authentication technique. Existing authentication mechanisms on smartphones can only provide one-time verification, but the verified users are still vulnerable to the session hijacking. In this article, we propose a transparent and continuous authentication system based on users' touch interaction data. We consider four common types of touch operations, and extract behavioral features to characterize users' touch behavior. Distance-measurement technique is applied to mitigate the behavioral variability. Then a multiple decision procedure is developed to perform continuous authentication, in which one-class classification algorithms are applied for classification. The efficacy of our approach is validated through a series of experiments in four real-world application scenarios. Our experimental results show that the proposed approach could verify a user in an accurate and timely manner, and suffice to be an enhancement for traditional authentication mechanisms in smartphones.
越来越多地使用智能手机来访问敏感和隐私数据,这引起了对安全认证技术的需求。智能手机上现有的身份验证机制只能提供一次性验证,但验证后的用户仍然容易受到会话劫持的攻击。在本文中,我们提出了一种基于用户触摸交互数据的透明连续认证系统。我们考虑了四种常见的触摸操作类型,并提取了行为特征来描述用户的触摸行为。采用距离测量技术来减轻行为的可变性。在此基础上,提出了一种多决策过程进行连续认证,并采用一类分类算法进行分类。通过四种实际应用场景的一系列实验验证了该方法的有效性。我们的实验结果表明,该方法可以准确及时地验证用户,足以增强智能手机中的传统身份验证机制。
{"title":"Touch-interaction behavior for continuous user authentication on smartphones","authors":"Chao Shen, Yong Zhang, Zhongmin Cai, Tianwen Yu, X. Guan","doi":"10.1109/ICB.2015.7139046","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139046","url":null,"abstract":"The increasing use of smartphones to access sensitive and privacy data has given rise to the need of secure authentication technique. Existing authentication mechanisms on smartphones can only provide one-time verification, but the verified users are still vulnerable to the session hijacking. In this article, we propose a transparent and continuous authentication system based on users' touch interaction data. We consider four common types of touch operations, and extract behavioral features to characterize users' touch behavior. Distance-measurement technique is applied to mitigate the behavioral variability. Then a multiple decision procedure is developed to perform continuous authentication, in which one-class classification algorithms are applied for classification. The efficacy of our approach is validated through a series of experiments in four real-world application scenarios. Our experimental results show that the proposed approach could verify a user in an accurate and timely manner, and suffice to be an enhancement for traditional authentication mechanisms in smartphones.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"335 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123779202","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Harnessing social context for improved face recognition 利用社会环境改善面部识别
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139085
Romil Bhardwaj, Gaurav Goswami, Richa Singh, Mayank Vatsa
Face recognition is traditionally based on features extracted from face images which capture various intrinsic characteristics of faces to distinguish between individuals. However, humans do not perform face recognition in isolation and instead utilize a wide variety of contextual cues as well in order to perform accurate recognition. Social context or co-occurrence of individuals is one such cue that humans utilize to reinforce face recognition output. A social graph can adequately model social-relationships between different individuals and this can be utilized to augment traditional face recognition methods. In this research, we propose a novel method to generate a social-graph based on a collection of group photographs and learn the social context information. We also propose a novel algorithm to combine results from a commercial face recognition system and social context information to perform face identification. Experimental results on two publicly available datasets show that social context information can improve face recognition and help bridge the gap between humans and machines in face recognition.
人脸识别传统上是基于从人脸图像中提取的特征来捕捉人脸的各种内在特征来区分个体。然而,人类并不是孤立地进行面部识别,而是利用各种各样的上下文线索来进行准确的识别。社会背景或共同出现的个人是一个这样的线索,人类利用以加强人脸识别输出。社交图可以充分地模拟不同个体之间的社会关系,这可以用来增强传统的人脸识别方法。在本研究中,我们提出了一种基于集体照片集合生成社交图并学习社交背景信息的新方法。我们还提出了一种新的算法,将商业人脸识别系统的结果与社会背景信息结合起来进行人脸识别。在两个公开数据集上的实验结果表明,社会背景信息可以提高人脸识别能力,并有助于弥合人与机器在人脸识别方面的差距。
{"title":"Harnessing social context for improved face recognition","authors":"Romil Bhardwaj, Gaurav Goswami, Richa Singh, Mayank Vatsa","doi":"10.1109/ICB.2015.7139085","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139085","url":null,"abstract":"Face recognition is traditionally based on features extracted from face images which capture various intrinsic characteristics of faces to distinguish between individuals. However, humans do not perform face recognition in isolation and instead utilize a wide variety of contextual cues as well in order to perform accurate recognition. Social context or co-occurrence of individuals is one such cue that humans utilize to reinforce face recognition output. A social graph can adequately model social-relationships between different individuals and this can be utilized to augment traditional face recognition methods. In this research, we propose a novel method to generate a social-graph based on a collection of group photographs and learn the social context information. We also propose a novel algorithm to combine results from a commercial face recognition system and social context information to perform face identification. Experimental results on two publicly available datasets show that social context information can improve face recognition and help bridge the gap between humans and machines in face recognition.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"5 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130095918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
I know that voice: Identifying the voice actor behind the voice 我知道那个声音:识别声音背后的配音演员
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139074
L. Uzan, Lior Wolf
Intentional voice modifications by electronic or nonelectronic means challenge automatic speaker recognition systems. Previous work focused on detecting the act of disguise or identifying everyday speakers disguising their voices. Here, we propose a benchmark for the study of voice disguise, by studying the voice variability of professional voice actors. A dataset of 114 actors playing 647 characters is created. It contains 19 hours of captured speech, divided into 29,733 utterances tagged by character and actor names, which is then further sampled. Text-independent speaker identification of the actors based on a novel benchmark training on a subset of the characters they play, while testing on new unseen characters, shows an EER of 17.1%, HTER of 15.9%, and rank-1 recognition rate of 63.5% per utterance when training a Convolutional Neural Network on spectrograms generated from the utterances. An I-Vector based system was trained and tested on the same data, resulting in 39.7% EER, 39.4% HTER, and rank-1 recognition rate of 13.6%.
通过电子或非电子手段故意修改语音对自动说话人识别系统提出了挑战。以前的工作集中在检测伪装行为或识别日常说话者伪装他们的声音。在这里,我们通过研究专业配音演员的声音变异性,提出了一个研究声音伪装的基准。创建了114名演员扮演647个角色的数据集。它包含19个小时的捕获语音,分为29,733个由角色和演员名字标记的话语,然后进一步采样。在对演员所扮演的角色子集进行新的基准训练的基础上,对演员进行了文本独立的说话人识别,同时对新的未见过的角色进行了测试,结果表明,在从话语生成的频谱图上训练卷积神经网络时,每个话语的EER为17.1%,HTER为15.9%,rank-1识别率为63.5%。基于I-Vector的系统在相同的数据上进行训练和测试,获得了39.7%的EER、39.4%的HTER和13.6%的rank-1识别率。
{"title":"I know that voice: Identifying the voice actor behind the voice","authors":"L. Uzan, Lior Wolf","doi":"10.1109/ICB.2015.7139074","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139074","url":null,"abstract":"Intentional voice modifications by electronic or nonelectronic means challenge automatic speaker recognition systems. Previous work focused on detecting the act of disguise or identifying everyday speakers disguising their voices. Here, we propose a benchmark for the study of voice disguise, by studying the voice variability of professional voice actors. A dataset of 114 actors playing 647 characters is created. It contains 19 hours of captured speech, divided into 29,733 utterances tagged by character and actor names, which is then further sampled. Text-independent speaker identification of the actors based on a novel benchmark training on a subset of the characters they play, while testing on new unseen characters, shows an EER of 17.1%, HTER of 15.9%, and rank-1 recognition rate of 63.5% per utterance when training a Convolutional Neural Network on spectrograms generated from the utterances. An I-Vector based system was trained and tested on the same data, resulting in 39.7% EER, 39.4% HTER, and rank-1 recognition rate of 13.6%.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129754565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Contactless biometric hand geometry recognition using a low-cost 3D camera 使用低成本3D相机的非接触式生物特征手部几何识别
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139109
Jan Svoboda, M. Bronstein, M. Drahanský
In the past decade, the interest in using 3D data for biometric person authentication has increased significantly, propelled by the availability of affordable 3D sensors. The adoption of 3D features has been especially successful in face recognition applications, leading to several commercial 3D face recognition products. In other biometric modalities such as hand recognition, several studies have shown the potential advantage of using 3D geometric information, however, no commercial-grade systems are currently available. In this paper, we present a contactless 3D hand recognition system based on the novel Intel RealSense camera, the first mass-produced embeddable 3D sensor. The small form factor and low cost make this sensor especially appealing for commercial biometric applications, however, they come at the price of lower resolution compared to more expensive 3D scanners used in previous research. We analyze the robustness of several existing 2D and 3D features that can be extracted from the images captured by the RealSense camera and study the use of metric learning for their fusion.
在过去的十年中,由于价格合理的3D传感器的可用性,使用3D数据进行生物识别身份验证的兴趣大大增加。采用3D功能在人脸识别应用中尤其成功,导致了一些商业化的3D人脸识别产品。在其他生物识别模式中,如手识别,一些研究已经显示了使用3D几何信息的潜在优势,然而,目前还没有商业级系统可用。在本文中,我们提出了一种基于新型英特尔RealSense相机的非接触式3D手部识别系统,这是第一个批量生产的嵌入式3D传感器。小尺寸和低成本使得这种传感器对商业生物识别应用特别有吸引力,然而,与之前研究中使用的更昂贵的3D扫描仪相比,它们的分辨率较低。我们分析了几种现有的2D和3D特征的鲁棒性,这些特征可以从RealSense相机捕获的图像中提取,并研究了使用度量学习进行融合的方法。
{"title":"Contactless biometric hand geometry recognition using a low-cost 3D camera","authors":"Jan Svoboda, M. Bronstein, M. Drahanský","doi":"10.1109/ICB.2015.7139109","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139109","url":null,"abstract":"In the past decade, the interest in using 3D data for biometric person authentication has increased significantly, propelled by the availability of affordable 3D sensors. The adoption of 3D features has been especially successful in face recognition applications, leading to several commercial 3D face recognition products. In other biometric modalities such as hand recognition, several studies have shown the potential advantage of using 3D geometric information, however, no commercial-grade systems are currently available. In this paper, we present a contactless 3D hand recognition system based on the novel Intel RealSense camera, the first mass-produced embeddable 3D sensor. The small form factor and low cost make this sensor especially appealing for commercial biometric applications, however, they come at the price of lower resolution compared to more expensive 3D scanners used in previous research. We analyze the robustness of several existing 2D and 3D features that can be extracted from the images captured by the RealSense camera and study the use of metric learning for their fusion.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"509 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122758073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
A transfer learning approach to cross-database facial expression recognition 跨数据库面部表情识别的迁移学习方法
Pub Date : 2015-05-19 DOI: 10.1109/ICB.2015.7139098
Ronghang Zhu, Tingting Zhang, Qijun Zhao, Zhihong Wu
Despite the high facial expression recognition accuracy reported on individual databases, cross-database facial expression recognition is still a challenging problem. This is essentially a problem of generalizing a facial expression recognizer trained with data of certain subjects under certain conditions to different subjects and/or different conditions. Such generalization capability is crucial in real-world applications. However, little attention has been focused on this problem in the literature. Transfer learning, a domain adaptation approach, provides effective techniques for transferring knowledge from source (training) data to target (testing) data when they are characterized by different properties. This paper makes the first attempt to apply transferring learning to cross-database facial expression recognition. It proposes a transfer learning based cross-database facial expression recognition approach, in which two training stages are involved: One for learning knowledge from source data, and the other for adapting the learned knowledge to target data. This approach has been implemented based on Gabor features extracted from facial images, regression tree classifiers, the AdaBoosting algorithm, and support vector machines. Evaluation experiments have been done on the JAFFE, FEED, and extended Cohn-Kanade databases. The results demonstrate that using the proposed transferring learning approach the cross-database facial expression recognition accuracy can be improved by more than 20%.
尽管在单个数据库上的面部表情识别准确率很高,但跨数据库的面部表情识别仍然是一个具有挑战性的问题。这本质上是一个用特定对象在特定条件下的数据训练的面部表情识别器推广到不同对象和/或不同条件的问题。这种泛化能力在实际应用中是至关重要的。然而,在文献中很少关注这一问题。迁移学习是一种领域自适应方法,它提供了将知识从源(训练)数据转移到具有不同特性的目标(测试)数据的有效技术。本文首次尝试将迁移学习应用于跨数据库面部表情识别。提出了一种基于迁移学习的跨数据库面部表情识别方法,该方法包括两个训练阶段:一个是从源数据中学习知识,另一个是将学习到的知识适应目标数据。该方法基于从面部图像中提取的Gabor特征、回归树分类器、AdaBoosting算法和支持向量机实现。在JAFFE、FEED和扩展的Cohn-Kanade数据库上进行了评价实验。结果表明,采用迁移学习方法进行跨数据库面部表情识别的准确率提高了20%以上。
{"title":"A transfer learning approach to cross-database facial expression recognition","authors":"Ronghang Zhu, Tingting Zhang, Qijun Zhao, Zhihong Wu","doi":"10.1109/ICB.2015.7139098","DOIUrl":"https://doi.org/10.1109/ICB.2015.7139098","url":null,"abstract":"Despite the high facial expression recognition accuracy reported on individual databases, cross-database facial expression recognition is still a challenging problem. This is essentially a problem of generalizing a facial expression recognizer trained with data of certain subjects under certain conditions to different subjects and/or different conditions. Such generalization capability is crucial in real-world applications. However, little attention has been focused on this problem in the literature. Transfer learning, a domain adaptation approach, provides effective techniques for transferring knowledge from source (training) data to target (testing) data when they are characterized by different properties. This paper makes the first attempt to apply transferring learning to cross-database facial expression recognition. It proposes a transfer learning based cross-database facial expression recognition approach, in which two training stages are involved: One for learning knowledge from source data, and the other for adapting the learned knowledge to target data. This approach has been implemented based on Gabor features extracted from facial images, regression tree classifiers, the AdaBoosting algorithm, and support vector machines. Evaluation experiments have been done on the JAFFE, FEED, and extended Cohn-Kanade databases. The results demonstrate that using the proposed transferring learning approach the cross-database facial expression recognition accuracy can be improved by more than 20%.","PeriodicalId":237372,"journal":{"name":"2015 International Conference on Biometrics (ICB)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123212384","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 18
期刊
2015 International Conference on Biometrics (ICB)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1