首页 > 最新文献

2017 IEEE International Joint Conference on Biometrics (IJCB)最新文献

英文 中文
Computing an image Phylogeny Tree from photometrically modified iris images 从光度法修改的虹膜图像计算图像系统发育树
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272749
Sudipta Banerjee, A. Ross
Iris recognition entails the use of iris images to recognize an individual. In some cases, the iris image acquired from an individual can be modified by subjecting it to successive photometric transformations such as brightening, gamma correction, median filtering and Gaussian smoothing, resulting in a family of transformed images. Automatically inferring the relationship between the set of transformed images is important in the context of digital image forensics. In this regard, we develop a method to generate an Image Phylogeny Tree (IPT) from a set of such transformed images. Our strategy entails modeling an arbitrary photometric transformation as a linear or non-linear function and utilizing the parameters of the model to quantify the relationship between pairs of images. The estimated parameters are then used to generate the IPT. Modest, yet promising, results are obtained in terms of parameter estimation and IPT generation.
虹膜识别需要使用虹膜图像来识别个体。在某些情况下,从个体获得的虹膜图像可以通过对其进行连续的光度变换(如增亮、伽马校正、中值滤波和高斯平滑)来进行修改,从而产生一系列变换后的图像。在数字图像取证的背景下,自动推断变换图像之间的关系是很重要的。在这方面,我们开发了一种从一组这样的转换图像生成图像系统发育树(IPT)的方法。我们的策略需要将任意光度变换建模为线性或非线性函数,并利用模型的参数来量化图像对之间的关系。然后使用估计的参数来生成IPT。在参数估计和IPT生成方面获得了适度但有希望的结果。
{"title":"Computing an image Phylogeny Tree from photometrically modified iris images","authors":"Sudipta Banerjee, A. Ross","doi":"10.1109/BTAS.2017.8272749","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272749","url":null,"abstract":"Iris recognition entails the use of iris images to recognize an individual. In some cases, the iris image acquired from an individual can be modified by subjecting it to successive photometric transformations such as brightening, gamma correction, median filtering and Gaussian smoothing, resulting in a family of transformed images. Automatically inferring the relationship between the set of transformed images is important in the context of digital image forensics. In this regard, we develop a method to generate an Image Phylogeny Tree (IPT) from a set of such transformed images. Our strategy entails modeling an arbitrary photometric transformation as a linear or non-linear function and utilizing the parameters of the model to quantify the relationship between pairs of images. The estimated parameters are then used to generate the IPT. Modest, yet promising, results are obtained in terms of parameter estimation and IPT generation.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130798426","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
DNA2FACE: An approach to correlating 3D facial structure and DNA DNA2FACE:一种将三维面部结构与DNA相关联的方法
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272746
Nisha Srinivas, Ryan Tokola, A. Mikkilineni, I. Nookaew, M. Leuze, Chris Boehnen
In this paper we introduce the concept of correlating genetic variations in an individual's specific genetic code (DNA) and facial morphology. This is the first step in the research effort to estimate facial appearance from DNA samples, which is gaining momentum within intelligence, law enforcement and national security communities. The dataset for the study consisting of genetic data and 3D facial scans (phenotype) data was obtained through the FaceBase Consortium. The proposed approach has three main steps: phenotype feature extraction from 3D face images, genotype feature extraction from a DNA sample, and genome-wide association analysis to determine genetic variations that contribute to facial structure and appearance. Results indicate that there exist significant correlations between genetic information and facial structure. We have identified 30 single nucleotide polymorphisms (SNPs), i.e. genetic variations, that significantly contribute to facial structure and appearance. We conclude with a preliminary attempt at facial reconstruction from the genetic data and emphasize on the complexity of the problem and the challenges encountered.
在本文中,我们介绍了个体特定遗传密码(DNA)与面部形态相关的遗传变异的概念。这是从DNA样本中估计面部特征的研究工作的第一步,在情报、执法和国家安全领域正获得越来越多的动力。该研究的数据集由遗传数据和3D面部扫描(表型)数据组成,通过FaceBase联盟获得。提出的方法有三个主要步骤:从3D面部图像中提取表型特征,从DNA样本中提取基因型特征,以及全基因组关联分析,以确定影响面部结构和外观的遗传变异。结果表明,遗传信息与面部结构存在显著相关性。我们已经确定了30个单核苷酸多态性(SNPs),即遗传变异,对面部结构和外观有重要影响。我们总结了从遗传数据中进行面部重建的初步尝试,并强调了问题的复杂性和遇到的挑战。
{"title":"DNA2FACE: An approach to correlating 3D facial structure and DNA","authors":"Nisha Srinivas, Ryan Tokola, A. Mikkilineni, I. Nookaew, M. Leuze, Chris Boehnen","doi":"10.1109/BTAS.2017.8272746","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272746","url":null,"abstract":"In this paper we introduce the concept of correlating genetic variations in an individual's specific genetic code (DNA) and facial morphology. This is the first step in the research effort to estimate facial appearance from DNA samples, which is gaining momentum within intelligence, law enforcement and national security communities. The dataset for the study consisting of genetic data and 3D facial scans (phenotype) data was obtained through the FaceBase Consortium. The proposed approach has three main steps: phenotype feature extraction from 3D face images, genotype feature extraction from a DNA sample, and genome-wide association analysis to determine genetic variations that contribute to facial structure and appearance. Results indicate that there exist significant correlations between genetic information and facial structure. We have identified 30 single nucleotide polymorphisms (SNPs), i.e. genetic variations, that significantly contribute to facial structure and appearance. We conclude with a preliminary attempt at facial reconstruction from the genetic data and emphasize on the complexity of the problem and the challenges encountered.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122832353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Multi-view 3D face reconstruction with deep recurrent neural networks 基于深度递归神经网络的多视图三维人脸重建
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272733
Pengfei Dou, I. Kakadiaris
Image-based 3D face reconstruction has great potential in different areas, such as facial recognition, facial analysis, and facial animation. Due to the variations in image quality, single-image-based 3D face reconstruction might not be sufficient to accurately reconstruct a 3D face. To overcome this limitation, multi-view 3D face reconstruction uses multiple images of the same subject and aggregates complementary information for better accuracy. Though theoretically appealing, there are multiple challenges in practice. Among these challenges, the most significant is that it is difficult to establish coherent and accurate correspondence among a set of images, especially when these images are captured in different conditions. In this paper, we propose a method, Deep Recurrent 3D FAce Reconstruction (DRFAR), to solve the task ofmulti-view 3D face reconstruction using a subspace representation of the 3D facial shape and a deep recurrent neural network that consists of both a deep con-volutional neural network (DCNN) and a recurrent neural network (RNN). The DCNN disentangles the facial identity and the facial expression components for each single image independently, while the RNN fuses identity-related features from the DCNN and aggregates the identity specific contextual information, or the identity signal, from the whole set of images to predict the facial identity parameter, which is robust to variations in image quality and is consistent over the whole set of images. Through extensive experiments, we evaluate our proposed method and demonstrate its superiority over existing methods.
基于图像的三维人脸重建在人脸识别、人脸分析和人脸动画等不同领域具有巨大的潜力。由于图像质量的差异,基于单图像的3D人脸重建可能不足以准确地重建3D人脸。为了克服这一限制,多视图3D人脸重建使用同一主题的多幅图像并聚合互补信息以提高准确性。虽然理论上很有吸引力,但在实践中存在多重挑战。在这些挑战中,最重要的是很难在一组图像之间建立连贯和准确的对应关系,特别是当这些图像在不同条件下捕获时。本文提出了一种深度递归3D人脸重建(Deep Recurrent 3D FAce Reconstruction, DRFAR)方法,该方法利用三维人脸形状的子空间表示和由深度卷积神经网络(DCNN)和递归神经网络(RNN)组成的深度递归神经网络来解决多视图3D人脸重建任务。DCNN可以独立分离每张图像的面部身份和面部表情成分,而RNN则融合来自DCNN的身份相关特征,并聚合来自整组图像的身份特定上下文信息或身份信号来预测面部身份参数,该参数对图像质量的变化具有鲁棒性,并且在整组图像上保持一致。通过大量的实验,我们评估了我们提出的方法,并证明了它比现有方法的优越性。
{"title":"Multi-view 3D face reconstruction with deep recurrent neural networks","authors":"Pengfei Dou, I. Kakadiaris","doi":"10.1109/BTAS.2017.8272733","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272733","url":null,"abstract":"Image-based 3D face reconstruction has great potential in different areas, such as facial recognition, facial analysis, and facial animation. Due to the variations in image quality, single-image-based 3D face reconstruction might not be sufficient to accurately reconstruct a 3D face. To overcome this limitation, multi-view 3D face reconstruction uses multiple images of the same subject and aggregates complementary information for better accuracy. Though theoretically appealing, there are multiple challenges in practice. Among these challenges, the most significant is that it is difficult to establish coherent and accurate correspondence among a set of images, especially when these images are captured in different conditions. In this paper, we propose a method, Deep Recurrent 3D FAce Reconstruction (DRFAR), to solve the task ofmulti-view 3D face reconstruction using a subspace representation of the 3D facial shape and a deep recurrent neural network that consists of both a deep con-volutional neural network (DCNN) and a recurrent neural network (RNN). The DCNN disentangles the facial identity and the facial expression components for each single image independently, while the RNN fuses identity-related features from the DCNN and aggregates the identity specific contextual information, or the identity signal, from the whole set of images to predict the facial identity parameter, which is robust to variations in image quality and is consistent over the whole set of images. Through extensive experiments, we evaluate our proposed method and demonstrate its superiority over existing methods.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126606987","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
Accuracy evaluation of handwritten signature verification: Rethinking the random-skilled forgeries dichotomy 手写签名验证的准确性评估:重新思考随机-熟练伪造二分法
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272711
Javier Galbally, M. Gomez-Barrero, A. Ross
Traditionally, the accuracy of signature verification systems has been evaluated following a protocol that considers two independent impostor scenarios: random forgeries and skilled forgeries. Although such an approach is not necessarily incorrect, it can lead to a misinterpretation of the results of the assessment process. Furthermore, such a full separation between both types of impostors may be unrealistic in many operational real-world applications. The current article discusses the soundness of the random-skilled impostor dichotomy and proposes complementary approaches to report the accuracy of signature verification systems, discussing their advantages and limitations.
传统上,签名验证系统的准确性是根据一种协议来评估的,该协议考虑了两种独立的冒充者场景:随机伪造和熟练伪造。虽然这种方法不一定是不正确的,但它可能导致对评估过程结果的误解。此外,在许多实际操作的应用程序中,完全区分这两种类型的冒名顶替者可能是不现实的。本文讨论了随机技术冒充者二分法的合理性,并提出了报告签名验证系统准确性的补充方法,讨论了它们的优点和局限性。
{"title":"Accuracy evaluation of handwritten signature verification: Rethinking the random-skilled forgeries dichotomy","authors":"Javier Galbally, M. Gomez-Barrero, A. Ross","doi":"10.1109/BTAS.2017.8272711","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272711","url":null,"abstract":"Traditionally, the accuracy of signature verification systems has been evaluated following a protocol that considers two independent impostor scenarios: random forgeries and skilled forgeries. Although such an approach is not necessarily incorrect, it can lead to a misinterpretation of the results of the assessment process. Furthermore, such a full separation between both types of impostors may be unrealistic in many operational real-world applications. The current article discusses the soundness of the random-skilled impostor dichotomy and proposes complementary approaches to report the accuracy of signature verification systems, discussing their advantages and limitations.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"316 7","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"113958766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Deep convolutional dynamic texture learning with adaptive channel-discriminability for 3D mask face anti-spoofing 基于自适应通道判别的深度卷积动态纹理学习3D蒙版人脸防欺骗
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272765
Rui Shao, X. Lan, P. Yuen
3D mask spoofing attack has been one of the main challenges in face recognition. A real face displays a different motion behaviour compared to a 3D mask spoof attempt, which is reflected by different facial dynamic textures. However, the different dynamic information usually exists in the subtle texture level, which cannot be fully differentiated by traditional hand-crafted texture-based methods. In this paper, we propose a novel method for 3D mask face anti-spoofing, namely deep convolutional dynamic texture learning, which learns robust dynamic texture information from fine-grained deep convolutional features. Moreover, channel-discriminability constraint is adaptively incorporated to weight the discriminability of feature channels in the learning process. Experiments on both public datasets validate that the proposed method achieves promising results under intra and cross dataset scenario.
三维面具欺骗攻击一直是人脸识别中的主要挑战之一。与3D面具欺骗相比,真实的面部表现出不同的运动行为,这反映在不同的面部动态纹理上。然而,不同的动态信息通常存在于细微纹理层面,传统的基于手工纹理的方法无法完全区分这些动态信息。本文提出了一种新的3D掩模人脸防欺骗方法,即深度卷积动态纹理学习,该方法从细粒度深度卷积特征中学习鲁棒动态纹理信息。此外,在学习过程中,自适应地引入信道可判别性约束,对特征信道的可判别性进行加权。在两个公共数据集上的实验验证了该方法在数据集内和跨数据集场景下都取得了很好的效果。
{"title":"Deep convolutional dynamic texture learning with adaptive channel-discriminability for 3D mask face anti-spoofing","authors":"Rui Shao, X. Lan, P. Yuen","doi":"10.1109/BTAS.2017.8272765","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272765","url":null,"abstract":"3D mask spoofing attack has been one of the main challenges in face recognition. A real face displays a different motion behaviour compared to a 3D mask spoof attempt, which is reflected by different facial dynamic textures. However, the different dynamic information usually exists in the subtle texture level, which cannot be fully differentiated by traditional hand-crafted texture-based methods. In this paper, we propose a novel method for 3D mask face anti-spoofing, namely deep convolutional dynamic texture learning, which learns robust dynamic texture information from fine-grained deep convolutional features. Moreover, channel-discriminability constraint is adaptively incorporated to weight the discriminability of feature channels in the learning process. Experiments on both public datasets validate that the proposed method achieves promising results under intra and cross dataset scenario.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"82 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126283746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 67
On the guessability of binary biometric templates: A practical guessing entropy based approach 二元生物特征模板的可猜测性:一种实用的基于猜测熵的方法
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272719
Guangcan Mai, M. Lim, P. Yuen
A security index for biometric systems is essential because biometrics have been widely adopted as a secure authentication component in critical systems. Most of bio-metric systems secured by template protection schemes are based on binary templates. To adopt popular template protection schemes such as fuzzy commitment and fuzzy extractor that can be applied on binary templates only, non-binary templates (e.g., real-valued, point-set based) need to be converted to binary. However, existing security measurements for binary template based biometric systems either cannot reflect the actual attack difficulties or are too computationally expensive to be practical. This paper presents an acceleration of the guessing entropy which reflects the expected number of guessing trials in attacking the binary template based biometric systems. The acceleration benefits from computation reuse and pruning. Experimental results on two datasets show that the acceleration has more than 6x, 20x, and 200x speed up without losing the estimation accuracy in different system settings.
由于生物识别技术已被广泛采用为关键系统的安全认证组件,因此生物识别系统的安全索引是必不可少的。大多数采用模板保护方案的生物识别系统都是基于二进制模板的。为了采用目前流行的仅适用于二进制模板的模糊承诺、模糊提取等模板保护方案,需要将非二进制模板(如实值、基于点集的模板)转换为二进制模板。然而,现有的基于二进制模板的生物识别系统的安全措施要么不能反映实际的攻击困难,要么计算成本太高而不实用。本文提出了一种加速猜测熵的方法,它反映了攻击基于二元模板的生物识别系统的预期猜测次数。这种加速得益于计算重用和修剪。在两个数据集上的实验结果表明,在不同的系统设置下,在不损失估计精度的情况下,该算法的加速速度分别超过6倍、20倍和200倍。
{"title":"On the guessability of binary biometric templates: A practical guessing entropy based approach","authors":"Guangcan Mai, M. Lim, P. Yuen","doi":"10.1109/BTAS.2017.8272719","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272719","url":null,"abstract":"A security index for biometric systems is essential because biometrics have been widely adopted as a secure authentication component in critical systems. Most of bio-metric systems secured by template protection schemes are based on binary templates. To adopt popular template protection schemes such as fuzzy commitment and fuzzy extractor that can be applied on binary templates only, non-binary templates (e.g., real-valued, point-set based) need to be converted to binary. However, existing security measurements for binary template based biometric systems either cannot reflect the actual attack difficulties or are too computationally expensive to be practical. This paper presents an acceleration of the guessing entropy which reflects the expected number of guessing trials in attacking the binary template based biometric systems. The acceleration benefits from computation reuse and pruning. Experimental results on two datasets show that the acceleration has more than 6x, 20x, and 200x speed up without losing the estimation accuracy in different system settings.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126350922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Conditional random fields incorporate convolutional neural networks for human eye sclera semantic segmentation
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272768
Russel Mesbah, B. McCane, S. Mills
Sclera segmentation as an ocular biometric has been of an interest in a variety of security and medical applications. The current approaches mostly rely on handcrafted features which make the generalisation of the learnt hypothesis challenging encountering images taken from various angles, and in different visible light spectrums. Convolutional Neural Networks (CNNs) are capable of extracting the corresponding features automatically. Despite the fact that CNNs showed a remarkable performance in a variety of image semantic segmentations, the output can be noisy and less accurate particularly in object boundaries. To address this issue, we have used Conditional Random Fields (CRFs) to regulate the CNN outputs. The results of applying this technique to sclera segmentation dataset (SSERBC 2017) are comparable with the state of the art solutions.
巩膜分割作为一种眼部生物识别技术已经在各种安全和医学应用中引起了人们的兴趣。目前的方法主要依赖于手工制作的特征,这使得从不同角度和不同可见光光谱拍摄的图像难以概括所学假设。卷积神经网络(cnn)能够自动提取相应的特征。尽管cnn在各种图像语义分割中表现出色,但输出可能会有噪声,特别是在物体边界上的准确性较低。为了解决这个问题,我们使用条件随机场(CRFs)来调节CNN输出。将该技术应用于巩膜分割数据集(SSERBC 2017)的结果与最先进的解决方案相当。
{"title":"Conditional random fields incorporate convolutional neural networks for human eye sclera semantic segmentation","authors":"Russel Mesbah, B. McCane, S. Mills","doi":"10.1109/BTAS.2017.8272768","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272768","url":null,"abstract":"Sclera segmentation as an ocular biometric has been of an interest in a variety of security and medical applications. The current approaches mostly rely on handcrafted features which make the generalisation of the learnt hypothesis challenging encountering images taken from various angles, and in different visible light spectrums. Convolutional Neural Networks (CNNs) are capable of extracting the corresponding features automatically. Despite the fact that CNNs showed a remarkable performance in a variety of image semantic segmentations, the output can be noisy and less accurate particularly in object boundaries. To address this issue, we have used Conditional Random Fields (CRFs) to regulate the CNN outputs. The results of applying this technique to sclera segmentation dataset (SSERBC 2017) are comparable with the state of the art solutions.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"480 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123054720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Deep learning with time-frequency representation for pulse estimation from facial videos 基于时频表示的深度学习人脸视频脉冲估计
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272721
G. Hsu, Arulmurugan Ambikapathi, Ming Chen
Accurate pulse estimation is of pivotal importance in acquiring the critical physical conditions of human subjects under test, and facial video based pulse estimation approaches recently gained attention owing to their simplicity. In this work, we have endeavored to develop a novel deep learning approach as the core part for pulse (heart rate) estimation by using a common RGB camera. Our approach consists of four steps. We first begin by detecting the face and its landmarks, and thereby locate the required facial ROI. In Step 2, we extract the sample mean sequences of the R, G, and B channels from the facial ROI, and explore three processing schemes for noise removal and signal enhancement. In Step 3, the Short-Time Fourier Transform (STFT) is employed to build the 2D Time-Frequency Representations (TFRs) of the sequences. The 2D TFR enables the formulation of the pulse estimation as an image-based classification problem, which can be solved in Step 4 by a deep Con-volutional Neural Network (CNN). Our approach is one of the pioneering works for attempting real-time pulse estimation using a deep learning framework. We have developed a pulse database, called the Pulse from Face (PFF), and used it to train the CNN. The PFF database will be made publicly available to advance related research. When compared to state-of-the-art pulse estimation approaches on the standard MAHNOB-HCI database, the proposed approach has exhibited superior performance.
准确的脉冲估计对于获取被测人体的临界身体状态至关重要,基于人脸视频的脉冲估计方法因其简单而受到关注。在这项工作中,我们努力开发一种新的深度学习方法,作为使用普通RGB相机进行脉搏(心率)估计的核心部分。我们的方法包括四个步骤。我们首先从检测人脸及其地标开始,从而定位所需的人脸ROI。在步骤2中,我们从面部ROI中提取R、G和B通道的样本均值序列,并探索了三种去噪和信号增强的处理方案。在步骤3中,使用短时傅里叶变换(STFT)来构建序列的二维时频表示(TFRs)。2D TFR可以将脉冲估计表述为基于图像的分类问题,该问题可以在步骤4中通过深度卷积神经网络(CNN)解决。我们的方法是尝试使用深度学习框架进行实时脉冲估计的开创性工作之一。我们开发了一个脉冲数据库,称为面部脉冲(PFF),并使用它来训练CNN。PFF数据库将向公众开放,以推进有关研究。当与标准MAHNOB-HCI数据库上的最先进的脉冲估计方法进行比较时,所提出的方法表现出优越的性能。
{"title":"Deep learning with time-frequency representation for pulse estimation from facial videos","authors":"G. Hsu, Arulmurugan Ambikapathi, Ming Chen","doi":"10.1109/BTAS.2017.8272721","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272721","url":null,"abstract":"Accurate pulse estimation is of pivotal importance in acquiring the critical physical conditions of human subjects under test, and facial video based pulse estimation approaches recently gained attention owing to their simplicity. In this work, we have endeavored to develop a novel deep learning approach as the core part for pulse (heart rate) estimation by using a common RGB camera. Our approach consists of four steps. We first begin by detecting the face and its landmarks, and thereby locate the required facial ROI. In Step 2, we extract the sample mean sequences of the R, G, and B channels from the facial ROI, and explore three processing schemes for noise removal and signal enhancement. In Step 3, the Short-Time Fourier Transform (STFT) is employed to build the 2D Time-Frequency Representations (TFRs) of the sequences. The 2D TFR enables the formulation of the pulse estimation as an image-based classification problem, which can be solved in Step 4 by a deep Con-volutional Neural Network (CNN). Our approach is one of the pioneering works for attempting real-time pulse estimation using a deep learning framework. We have developed a pulse database, called the Pulse from Face (PFF), and used it to train the CNN. The PFF database will be made publicly available to advance related research. When compared to state-of-the-art pulse estimation approaches on the standard MAHNOB-HCI database, the proposed approach has exhibited superior performance.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"604 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129215996","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Face anti-spoofing using patch and depth-based CNNs 使用补丁和基于深度的cnn进行人脸防欺骗
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272713
Yousef Atoum, Yaojie Liu, Amin Jourabloo, Xiaoming Liu
The face image is the most accessible biometric modality which is used for highly accurate face recognition systems, while it is vulnerable to many different types of presentation attacks. Face anti-spoofing is a very critical step before feeding the face image to biometric systems. In this paper, we propose a novel two-stream CNN-based approach for face anti-spoofing, by extracting the local features and holistic depth maps from the face images. The local features facilitate CNN to discriminate the spoof patches independent of the spatial face areas. On the other hand, holistic depth map examine whether the input image has a face-like depth. Extensive experiments are conducted on the challenging databases (CASIA-FASD, MSU-USSA, and Replay Attack), with comparison to the state of the art.
人脸图像是用于高精度人脸识别系统的最容易获得的生物识别模式,但它容易受到许多不同类型的表示攻击。人脸防欺骗是将人脸图像输入生物识别系统之前非常关键的一步。在本文中,我们提出了一种新的基于cnn的双流人脸防欺骗方法,通过从人脸图像中提取局部特征和整体深度图。局部特征使CNN能够独立于空间人脸区域区分恶搞斑块。另一方面,整体深度图检查输入图像是否具有类似人脸的深度。在具有挑战性的数据库(CASIA-FASD, MSU-USSA和Replay Attack)上进行了广泛的实验,并与最新技术进行了比较。
{"title":"Face anti-spoofing using patch and depth-based CNNs","authors":"Yousef Atoum, Yaojie Liu, Amin Jourabloo, Xiaoming Liu","doi":"10.1109/BTAS.2017.8272713","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272713","url":null,"abstract":"The face image is the most accessible biometric modality which is used for highly accurate face recognition systems, while it is vulnerable to many different types of presentation attacks. Face anti-spoofing is a very critical step before feeding the face image to biometric systems. In this paper, we propose a novel two-stream CNN-based approach for face anti-spoofing, by extracting the local features and holistic depth maps from the face images. The local features facilitate CNN to discriminate the spoof patches independent of the spatial face areas. On the other hand, holistic depth map examine whether the input image has a face-like depth. Extensive experiments are conducted on the challenging databases (CASIA-FASD, MSU-USSA, and Replay Attack), with comparison to the state of the art.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121188977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 325
Learning optimised representations for view-invariant gait recognition 视觉不变步态识别的学习优化表征
Pub Date : 2017-10-01 DOI: 10.1109/BTAS.2017.8272769
Ning Jia, Victor Sanchez, Chang-Tsun Li
Gait recognition can be performed without subject cooperation under harsh conditions, thus it is an important tool in forensic gait analysis, security control, and other commercial applications. One critical issue that prevents gait recognition systems from being widely accepted is the performance drop when the camera viewpoint varies between the registered templates and the query data. In this paper, we explore the potential of combining feature optimisers and representations learned by convolutional neural networks (CNN) to achieve efficient view-invariant gait recognition. The experimental results indicate that CNN learns highly discriminative representations across moderate view variations, and these representations can be further improved using view-invariant feature selectors, achieving a high matching accuracy across views.
步态识别在恶劣条件下无需受试者配合即可完成,是法医步态分析、安全控制等商业应用的重要工具。阻碍步态识别系统被广泛接受的一个关键问题是,当摄像机视点在注册模板和查询数据之间变化时,性能会下降。在本文中,我们探索了将特征优化器和卷积神经网络(CNN)学习的表征相结合的潜力,以实现高效的视觉不变步态识别。实验结果表明,CNN在适度的视图变化中学习了高度判别的表征,并且这些表征可以使用视图不变的特征选择器进一步改进,从而在视图之间实现较高的匹配精度。
{"title":"Learning optimised representations for view-invariant gait recognition","authors":"Ning Jia, Victor Sanchez, Chang-Tsun Li","doi":"10.1109/BTAS.2017.8272769","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272769","url":null,"abstract":"Gait recognition can be performed without subject cooperation under harsh conditions, thus it is an important tool in forensic gait analysis, security control, and other commercial applications. One critical issue that prevents gait recognition systems from being widely accepted is the performance drop when the camera viewpoint varies between the registered templates and the query data. In this paper, we explore the potential of combining feature optimisers and representations learned by convolutional neural networks (CNN) to achieve efficient view-invariant gait recognition. The experimental results indicate that CNN learns highly discriminative representations across moderate view variations, and these representations can be further improved using view-invariant feature selectors, achieving a high matching accuracy across views.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116031842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2017 IEEE International Joint Conference on Biometrics (IJCB)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1