首页 > 最新文献

2017 IEEE International Joint Conference on Biometrics (IJCB)最新文献

英文 中文
Periocular recognition in cross-spectral scenario 交叉光谱场景下的眼周识别
Pub Date : 2017-06-29 DOI: 10.1109/BTAS.2017.8272757
S. S. Behera, Mahesh Gour, Vivek Kanhangad, N. Puhan
Periocular recognition has been an active area of research in the past few years. In spite of the advancements made in this area, the cross-spectral matching of visible (VIS) and near-infrared (NIR) periocular images remains a challenge. In this paper, we propose a method based on illumination normalization of VIS and NIR periocular images. Specifically, the approach involves normalizing the images using the difference of Gaussian (DoG) filtering, followed by the computation of a descriptor that captures structural details in the illumination normalized images using histogram of oriented gradients (HOG). Finally, the feature vectors corresponding to the query and the enrolled image are compared using the cosine similarity metric to generate a matching score. Performance of our algorithm has been evaluated on three publicly available benchmark databases of cross-spectral periocular images. Our approach yields significant improvement in performance over the existing approach.
眼周识别是近年来研究的一个活跃领域。尽管在这方面取得了一些进展,但可见光(VIS)和近红外(NIR)近眼图像的交叉光谱匹配仍然是一个挑战。本文提出了一种基于照度归一化的VIS和NIR眼周图像检测方法。具体来说,该方法包括使用高斯差分(DoG)滤波对图像进行归一化,然后使用定向梯度直方图(HOG)计算捕获照明归一化图像中的结构细节的描述符。最后,使用余弦相似度度量比较查询和注册图像对应的特征向量以生成匹配分数。我们的算法的性能已经在三个公开可用的交叉光谱眼周图像基准数据库上进行了评估。我们的方法在性能上比现有的方法有了显著的提高。
{"title":"Periocular recognition in cross-spectral scenario","authors":"S. S. Behera, Mahesh Gour, Vivek Kanhangad, N. Puhan","doi":"10.1109/BTAS.2017.8272757","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272757","url":null,"abstract":"Periocular recognition has been an active area of research in the past few years. In spite of the advancements made in this area, the cross-spectral matching of visible (VIS) and near-infrared (NIR) periocular images remains a challenge. In this paper, we propose a method based on illumination normalization of VIS and NIR periocular images. Specifically, the approach involves normalizing the images using the difference of Gaussian (DoG) filtering, followed by the computation of a descriptor that captures structural details in the illumination normalized images using histogram of oriented gradients (HOG). Finally, the feature vectors corresponding to the query and the enrolled image are compared using the cosine similarity metric to generate a matching score. Performance of our algorithm has been evaluated on three publicly available benchmark databases of cross-spectral periocular images. Our approach yields significant improvement in performance over the existing approach.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-06-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126316805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
You are how you walk: Uncooperative MoCap gait identification for video surveillance with incomplete and noisy data 你就是你走路的样子:不合作的动作捕捉步态识别,用于视频监控不完整和嘈杂的数据
Pub Date : 2017-05-10 DOI: 10.1109/BTAS.2017.8272700
Michal Balazia, Petr Sojka
This work offers a design of a video surveillance system based on a soft biometric — gait identification from MoCap data. The main focus is on two substantial issues of the video surveillance scenario: (1) the walkers do not cooperate in providing learning data to establish their identities and (2) the data are often noisy or incomplete. We show that only a few examples of human gait cycles are required to learn a projection of raw MoCap data onto a low-dimensional subspace where the identities are well separable. Latent features learned by Maximum Margin Criterion (MMC) method discriminate better than any collection of geometric features. The MMC method is also highly robust to noisy data and works properly even with only a fraction of joints tracked. The overall workflow of the design is directly applicable for a day-to-day operation based on the available MoCap technology and algorithms for gait analysis. In the concept we introduce, a walker's identity is represented by a cluster of gait data collected at their incidents within the surveillance system: They are how they walk.
本文提出了一种基于动作捕捉数据的软生物识别-步态识别的视频监控系统设计。主要关注视频监控场景的两个实质性问题:(1)步行者不合作提供学习数据以确定其身份;(2)数据通常有噪声或不完整。我们表明,只需要几个人类步态周期的例子,就可以学习将原始动作捕捉数据投影到低维子空间中,其中身份可以很好地分离。利用最大边界准则(MMC)学习到的潜在特征比任何几何特征集合都更好。MMC方法对噪声数据具有很强的鲁棒性,即使只跟踪一小部分关节也能正常工作。基于可用的动作捕捉技术和步态分析算法,该设计的整体工作流程可直接适用于日常操作。在我们介绍的概念中,步行者的身份由监控系统中收集的一组步态数据来表示:他们是如何走路的。
{"title":"You are how you walk: Uncooperative MoCap gait identification for video surveillance with incomplete and noisy data","authors":"Michal Balazia, Petr Sojka","doi":"10.1109/BTAS.2017.8272700","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272700","url":null,"abstract":"This work offers a design of a video surveillance system based on a soft biometric — gait identification from MoCap data. The main focus is on two substantial issues of the video surveillance scenario: (1) the walkers do not cooperate in providing learning data to establish their identities and (2) the data are often noisy or incomplete. We show that only a few examples of human gait cycles are required to learn a projection of raw MoCap data onto a low-dimensional subspace where the identities are well separable. Latent features learned by Maximum Margin Criterion (MMC) method discriminate better than any collection of geometric features. The MMC method is also highly robust to noisy data and works properly even with only a fraction of joints tracked. The overall workflow of the design is directly applicable for a day-to-day operation based on the available MoCap technology and algorithms for gait analysis. In the concept we introduce, a walker's identity is represented by a cluster of gait data collected at their incidents within the surveillance system: They are how they walk.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121205156","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Deep person re-identification with improved embedding and efficient training 通过改进的嵌入和有效的培训进行深度人员再识别
Pub Date : 2017-05-09 DOI: 10.1109/BTAS.2017.8272706
Haibo Jin, Xiaobo Wang, Shengcai Liao, S. Li
Person re-identification task has been greatly boosted by deep convolutional neural networks (CNNs) in recent years. The core of which is to enlarge the inter-class distinction as well as reduce the intra-class variance. However, to achieve this, existing deep models prefer to adopt image pairs or triplets to form verification loss, which is inefficient and unstable since the number of training pairs or triplets grows rapidly as the number of training data grows. Moreover, their performance is limited since they ignore the fact that different dimension of embedding may play different importance. In this paper, we propose to employ identification loss with center loss to train a deep model for person re-identification. The training process is efficient since it does not require image pairs or triplets for training while the inter-class distinction and intra-class variance are well handled. To boost the performance, a new feature reweighting (FRW) layer is designed to explicitly emphasize the importance of each embedding dimension, thus leading to an improved embedding. Experiments 1 on several benchmark datasets have shown the superiority of our method over the state-of-the-art alternatives on both accuracy and speed.
近年来,深度卷积神经网络(cnn)极大地促进了人的再识别任务。其核心是扩大阶级间差异,缩小阶级内差异。然而,为了实现这一点,现有的深度模型更倾向于采用图像对或三元组来形成验证损失,由于随着训练数据数量的增加,训练对或三元组的数量会迅速增加,这种方法效率低下且不稳定。此外,由于忽略了嵌入的不同维度可能具有不同的重要性这一事实,它们的表现受到了限制。在本文中,我们提出使用识别损失和中心损失来训练一个人再识别的深度模型。训练过程不需要图像对或三元组进行训练,同时很好地处理了类间差异和类内方差。为了提高性能,设计了一个新的特征重加权(FRW)层,明确强调每个嵌入维度的重要性,从而改进了嵌入。在几个基准数据集上的实验1表明,我们的方法在准确性和速度上都优于最先进的替代方法。
{"title":"Deep person re-identification with improved embedding and efficient training","authors":"Haibo Jin, Xiaobo Wang, Shengcai Liao, S. Li","doi":"10.1109/BTAS.2017.8272706","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272706","url":null,"abstract":"Person re-identification task has been greatly boosted by deep convolutional neural networks (CNNs) in recent years. The core of which is to enlarge the inter-class distinction as well as reduce the intra-class variance. However, to achieve this, existing deep models prefer to adopt image pairs or triplets to form verification loss, which is inefficient and unstable since the number of training pairs or triplets grows rapidly as the number of training data grows. Moreover, their performance is limited since they ignore the fact that different dimension of embedding may play different importance. In this paper, we propose to employ identification loss with center loss to train a deep model for person re-identification. The training process is efficient since it does not require image pairs or triplets for training while the inter-class distinction and intra-class variance are well handled. To boost the performance, a new feature reweighting (FRW) layer is designed to explicitly emphasize the importance of each embedding dimension, thus leading to an improved embedding. Experiments 1 on several benchmark datasets have shown the superiority of our method over the state-of-the-art alternatives on both accuracy and speed.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131063068","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Generative convolutional networks for latent fingerprint reconstruction 基于生成卷积网络的潜在指纹重建
Pub Date : 2017-05-04 DOI: 10.1109/BTAS.2017.8272727
Jan Svoboda, Federico Monti, M. Bronstein
Performance of fingerprint recognition depends heavily on the extraction of minutiae points. Enhancement of the fingerprint ridge pattern is thus an essential pre-processing step that noticeably reduces false positive and negative detection rates. A particularly challenging setting is when the fingerprint images are corrupted or partially missing. In this work, we apply generative convolutional networks to denoise visible minutiae and predict the missing parts of the ridge pattern. The proposed enhancement approach is tested as a pre-processing step in combination with several standard feature extraction methods such as MINDTCT, followed by biometric comparison using MCC and BO-ZORTH3. We evaluate our method on several publicly available latent fingerprint datasets captured using different sensors.
指纹识别的性能在很大程度上取决于特征点的提取。因此,指纹脊纹的增强是一个必要的预处理步骤,可以显著降低假阳性和假阴性的检测率。一个特别具有挑战性的设置是当指纹图像损坏或部分丢失时。在这项工作中,我们应用生成卷积网络去噪可见细节并预测脊图案的缺失部分。结合MINDTCT等几种标准特征提取方法,对所提出的增强方法作为预处理步骤进行了测试,然后使用MCC和BO-ZORTH3进行生物特征比较。我们在使用不同传感器捕获的几个公开可用的潜在指纹数据集上评估了我们的方法。
{"title":"Generative convolutional networks for latent fingerprint reconstruction","authors":"Jan Svoboda, Federico Monti, M. Bronstein","doi":"10.1109/BTAS.2017.8272727","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272727","url":null,"abstract":"Performance of fingerprint recognition depends heavily on the extraction of minutiae points. Enhancement of the fingerprint ridge pattern is thus an essential pre-processing step that noticeably reduces false positive and negative detection rates. A particularly challenging setting is when the fingerprint images are corrupted or partially missing. In this work, we apply generative convolutional networks to denoise visible minutiae and predict the missing parts of the ridge pattern. The proposed enhancement approach is tested as a pre-processing step in combination with several standard feature extraction methods such as MINDTCT, followed by biometric comparison using MCC and BO-ZORTH3. We evaluate our method on several publicly available latent fingerprint datasets captured using different sensors.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123099944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
SREFI: Synthesis of realistic example face images SREFI:真实示例人脸图像的合成
Pub Date : 2017-04-21 DOI: 10.1109/BTAS.2017.8272680
Sandipan Banerjee, John S. Bernhard, W. Scheirer, K. Bowyer, P. Flynn
In this paper, we propose a novel face synthesis approach that can generate an arbitrarily large number of synthetic images of both real and synthetic identities. Thus a face image dataset can be expanded in terms of the number of identities represented and the number of images per identity using this approach, without the identity-labeling and privacy complications that come from downloading images from the web. To measure the visual fidelity and uniqueness of the synthetic face images and identities, we conducted face matching experiments with both human participants and a CNN pre-trained on a dataset of 2.6M real face images. To evaluate the stability of these synthetic faces, we trained a CNN model with an augmented dataset containing close to 200,000 synthetic faces. We used a snapshot of this trained CNN to recognize extremely challenging frontal (real) face images. Experiments showed training with the augmented faces boosted the face recognition performance of the CNN.
在本文中,我们提出了一种新的人脸合成方法,可以生成任意数量的真实和合成身份的合成图像。因此,使用这种方法可以根据所表示的身份数量和每个身份的图像数量来扩展人脸图像数据集,而无需从网络上下载图像所带来的身份标签和隐私复杂性。为了衡量合成人脸图像和身份的视觉保真度和唯一性,我们在260万张真实人脸图像的数据集上,对人类参与者和预训练的CNN进行了人脸匹配实验。为了评估这些合成人脸的稳定性,我们使用包含近20万个合成人脸的增强数据集训练了一个CNN模型。我们使用经过训练的CNN的快照来识别极具挑战性的正面(真实)人脸图像。实验表明,增强人脸训练提高了CNN的人脸识别性能。
{"title":"SREFI: Synthesis of realistic example face images","authors":"Sandipan Banerjee, John S. Bernhard, W. Scheirer, K. Bowyer, P. Flynn","doi":"10.1109/BTAS.2017.8272680","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272680","url":null,"abstract":"In this paper, we propose a novel face synthesis approach that can generate an arbitrarily large number of synthetic images of both real and synthetic identities. Thus a face image dataset can be expanded in terms of the number of identities represented and the number of images per identity using this approach, without the identity-labeling and privacy complications that come from downloading images from the web. To measure the visual fidelity and uniqueness of the synthetic face images and identities, we conducted face matching experiments with both human participants and a CNN pre-trained on a dataset of 2.6M real face images. To evaluate the stability of these synthetic faces, we trained a CNN model with an augmented dataset containing close to 200,000 synthetic faces. We used a snapshot of this trained CNN to recognize extremely challenging frontal (real) face images. Experiments showed training with the augmented faces boosted the face recognition performance of the CNN.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132707788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Deep 3D face identification 深度三维人脸识别
Pub Date : 2017-03-30 DOI: 10.1109/BTAS.2017.8272691
Donghyun Kim, Matthias Hernandez, Jongmoo Choi, G. Medioni
We propose a novel 3D face recognition algorithm using a deep convolutional neural network (DCNN) and a 3D face expression augmentation technique. The performance of 2D face recognition algorithms has significantly increased by leveraging the representational power of deep neural networks and the use of large-scale labeled training data. In this paper, we show that transfer learning from a CNN trained on 2D face images can effectively work for 3D face recognition by fine-tuning the CNN with an extremely small number of 3D facial scans. We also propose a 3D face expression augmentation technique which synthesizes a number of different facial expressions from a single 3D face scan. Our proposed method shows excellent recognition results on Bosphorus, BU-3DFE, and 3D-TEC datasets without using hand-crafted features. The 3D face identification using our deep features also scales well for large databases.
本文提出了一种基于深度卷积神经网络(DCNN)和三维面部表情增强技术的三维人脸识别算法。利用深度神经网络的表征能力和大规模标记训练数据的使用,显著提高了二维人脸识别算法的性能。在本文中,我们表明,通过使用极少量的3D面部扫描对CNN进行微调,从2D面部图像上训练的CNN迁移学习可以有效地用于3D面部识别。我们还提出了一种3D面部表情增强技术,该技术可以从单个3D面部扫描中合成许多不同的面部表情。该方法在不使用手工特征的情况下,对博斯普鲁斯、BU-3DFE和3D-TEC数据集显示了良好的识别效果。使用我们的深度特征的3D人脸识别也可以很好地扩展到大型数据库。
{"title":"Deep 3D face identification","authors":"Donghyun Kim, Matthias Hernandez, Jongmoo Choi, G. Medioni","doi":"10.1109/BTAS.2017.8272691","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272691","url":null,"abstract":"We propose a novel 3D face recognition algorithm using a deep convolutional neural network (DCNN) and a 3D face expression augmentation technique. The performance of 2D face recognition algorithms has significantly increased by leveraging the representational power of deep neural networks and the use of large-scale labeled training data. In this paper, we show that transfer learning from a CNN trained on 2D face images can effectively work for 3D face recognition by fine-tuning the CNN with an extremely small number of 3D facial scans. We also propose a 3D face expression augmentation technique which synthesizes a number of different facial expressions from a single 3D face scan. Our proposed method shows excellent recognition results on Bosphorus, BU-3DFE, and 3D-TEC datasets without using hand-crafted features. The 3D face identification using our deep features also scales well for large databases.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2017-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122249494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 110
LOTS about attacking deep features 很多关于攻击深层特征的内容
Pub Date : 2016-11-18 DOI: 10.1109/BTAS.2017.8272695
Andras Rozsa, Manuel Günther, T. Boult
Deep neural networks provide state-of-the-art performance on various tasks and are, therefore, widely used in real world applications. DNNs are becoming frequently utilized in biometrics for extracting deep features, which can be used in recognition systems for enrolling and recognizing new individuals. It was revealed that deep neural networks suffer from a fundamental problem, namely, they can unexpectedly misclassify examples formed by slightly perturbing correctly recognized inputs. Various approaches have been developed for generating these so-called adversarial examples, but they aim at attacking end-to-end networks. For biometrics, it is natural to ask whether systems using deep features are immune to or, at least, more resilient to attacks than end-to-end networks. In this paper, we introduce a general technique called the layerwise origin-target synthesis (LOTS) that can be efficiently used to form adversarial examples that mimic the deep features of the target. We analyze and compare the adversarial robustness of the end-to-end VGG Face network with systems that use Euclidean or cosine distance between gallery templates and extracted deep features. We demonstrate that iterative LOTS is very effective and show that systems utilizing deep features are easier to attack than the end-to-end network.
深度神经网络在各种任务上提供最先进的性能,因此在现实世界的应用中得到广泛应用。深度神经网络在生物识别中被广泛用于提取深层特征,这些特征可以用于识别系统中招募和识别新个体。研究表明,深度神经网络存在一个基本问题,即对正确识别的输入进行轻微干扰就会产生意想不到的错误分类。已经开发了各种方法来生成这些所谓的对抗性示例,但它们的目标是攻击端到端网络。对于生物识别技术,人们很自然地会问,使用深度特征的系统是否比端到端网络更能抵御攻击,或者至少更能抵御攻击。在本文中,我们介绍了一种称为分层原点-目标合成(LOTS)的通用技术,该技术可以有效地用于形成模拟目标深层特征的对抗性示例。我们分析并比较了端到端VGG Face网络与使用欧几里得或余弦距离的库模板和提取深度特征的系统的对抗鲁棒性。我们证明了迭代lot是非常有效的,并且表明利用深度特征的系统比端到端网络更容易被攻击。
{"title":"LOTS about attacking deep features","authors":"Andras Rozsa, Manuel Günther, T. Boult","doi":"10.1109/BTAS.2017.8272695","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272695","url":null,"abstract":"Deep neural networks provide state-of-the-art performance on various tasks and are, therefore, widely used in real world applications. DNNs are becoming frequently utilized in biometrics for extracting deep features, which can be used in recognition systems for enrolling and recognizing new individuals. It was revealed that deep neural networks suffer from a fundamental problem, namely, they can unexpectedly misclassify examples formed by slightly perturbing correctly recognized inputs. Various approaches have been developed for generating these so-called adversarial examples, but they aim at attacking end-to-end networks. For biometrics, it is natural to ask whether systems using deep features are immune to or, at least, more resilient to attacks than end-to-end networks. In this paper, we introduce a general technique called the layerwise origin-target synthesis (LOTS) that can be efficiently used to form adversarial examples that mimic the deep features of the target. We analyze and compare the adversarial robustness of the end-to-end VGG Face network with systems that use Euclidean or cosine distance between gallery templates and extracted deep features. We demonstrate that iterative LOTS is very effective and show that systems utilizing deep features are easier to attack than the end-to-end network.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131471680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
AFFACT: Alignment-free facial attribute classification technique 影响:无对齐面部属性分类技术
Pub Date : 2016-11-18 DOI: 10.1109/BTAS.2017.8272686
Manuel Günther, Andras Rozsa, T. Boult
Facial attributes are soft-biometrics that allow limiting the search space, e.g., by rejecting identities with non-matching facial characteristics such as nose sizes or eyebrow shapes. In this paper, we investigate how the latest versions of deep convolutional neural networks, ResNets, perform on the facial attribute classification task. We test two loss functions: the sigmoid cross-entropy loss and the Euclidean loss, and find that for classification performance there is little difference between these two. Using an ensemble of three ResNets, we obtain the new state-of-the-art facial attribute classification error of 8.00 % on the aligned images of the CelebA dataset. More significantly, we introduce the Alignment-Free Facial Attribute Classification Technique (AFFACT), a data augmentation technique that allows a network to classify facial attributes without requiring alignment beyond detected face bounding boxes. To our best knowledge, we are the first to report similar accuracy when using only the detected bounding boxes — rather than requiring alignment based on automatically detected facial landmarks — and who can improve classification accuracy with rotating and scaling test images. We show that this approach outperforms the CelebA baseline on unaligned images with a relative improvement of 36.8 %.
面部属性是允许限制搜索空间的软生物识别技术,例如,通过拒绝具有不匹配的面部特征(如鼻子大小或眉毛形状)的身份。在本文中,我们研究了最新版本的深度卷积神经网络ResNets在面部属性分类任务上的表现。我们测试了两种损失函数:s型交叉熵损失和欧几里得损失,发现两者在分类性能上差别不大。使用三个resnet的集合,我们在CelebA数据集的对齐图像上获得了新的最先进的面部属性分类误差为8.00%。更重要的是,我们引入了无对齐面部属性分类技术(AFFACT),这是一种数据增强技术,允许网络对面部属性进行分类,而不需要在检测到的面部边界框之外进行对齐。据我们所知,我们是第一个在仅使用检测到的边界框时报告相似精度的人-而不是要求基于自动检测到的面部地标进行对齐-并且可以通过旋转和缩放测试图像来提高分类精度。我们表明,这种方法在未对齐图像上的性能优于CelebA基线,相对提高了36.8%。
{"title":"AFFACT: Alignment-free facial attribute classification technique","authors":"Manuel Günther, Andras Rozsa, T. Boult","doi":"10.1109/BTAS.2017.8272686","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272686","url":null,"abstract":"Facial attributes are soft-biometrics that allow limiting the search space, e.g., by rejecting identities with non-matching facial characteristics such as nose sizes or eyebrow shapes. In this paper, we investigate how the latest versions of deep convolutional neural networks, ResNets, perform on the facial attribute classification task. We test two loss functions: the sigmoid cross-entropy loss and the Euclidean loss, and find that for classification performance there is little difference between these two. Using an ensemble of three ResNets, we obtain the new state-of-the-art facial attribute classification error of 8.00 % on the aligned images of the CelebA dataset. More significantly, we introduce the Alignment-Free Facial Attribute Classification Technique (AFFACT), a data augmentation technique that allows a network to classify facial attributes without requiring alignment beyond detected face bounding boxes. To our best knowledge, we are the first to report similar accuracy when using only the detected bounding boxes — rather than requiring alignment based on automatically detected facial landmarks — and who can improve classification accuracy with rotating and scaling test images. We show that this approach outperforms the CelebA baseline on unaligned images with a relative improvement of 36.8 %.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129603558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
UMDFaces: An annotated face dataset for training deep networks UMDFaces:用于训练深度网络的带注释的人脸数据集
Pub Date : 2016-11-04 DOI: 10.1109/BTAS.2017.8272731
Ankan Bansal, Anirudh Nanduri, C. Castillo, Rajeev Ranjan, R. Chellappa
Recent progress in face detection (including keypoint detection), and recognition is mainly being driven by (i) deeper convolutional neural network architectures, and (ii) larger datasets. However, most of the large datasets are maintained by private companies and are not publicly available. The academic computer vision community needs larger and more varied datasets to make further progress. In this paper, we introduce a new face dataset, called UMDFaces, which has 367,888 annotated faces of 8,277 subjects. We also introduce a new face recognition evaluation protocol which will help advance the state-of-the-art in this area. We discuss how a large dataset can be collected and annotated using human annotators and deep networks. We provide human curated bounding boxes for faces. We also provide estimated pose (roll, pitch and yaw), locations of twenty-one key-points and gender information generated by a pre-trained neural network. In addition, the quality of keypoint annotations has been verified by humans for about 115,000 images. Finally, we compare the quality of the dataset with other publicly available face datasets at similar scales.
人脸检测(包括关键点检测)和识别的最新进展主要是由(i)更深的卷积神经网络架构和(ii)更大的数据集驱动的。然而,大多数大型数据集是由私营公司维护的,不向公众开放。学术计算机视觉社区需要更大、更多样化的数据集来取得进一步的进展。在本文中,我们引入了一个新的人脸数据集,称为UMDFaces,该数据集包含8277个受试者的367,888张标注的人脸。我们还介绍了一种新的人脸识别评估协议,这将有助于推进这一领域的最新技术。我们讨论了如何使用人工注释器和深度网络收集和注释大型数据集。我们为人脸提供人类策划的边界框。我们还提供了预估姿态(滚动、俯仰和偏航)、21个关键点的位置以及由预训练的神经网络生成的性别信息。此外,关键点标注的质量已经被人类对大约11.5万张图片进行了验证。最后,我们将数据集的质量与其他公开可用的类似规模的人脸数据集进行比较。
{"title":"UMDFaces: An annotated face dataset for training deep networks","authors":"Ankan Bansal, Anirudh Nanduri, C. Castillo, Rajeev Ranjan, R. Chellappa","doi":"10.1109/BTAS.2017.8272731","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272731","url":null,"abstract":"Recent progress in face detection (including keypoint detection), and recognition is mainly being driven by (i) deeper convolutional neural network architectures, and (ii) larger datasets. However, most of the large datasets are maintained by private companies and are not publicly available. The academic computer vision community needs larger and more varied datasets to make further progress. In this paper, we introduce a new face dataset, called UMDFaces, which has 367,888 annotated faces of 8,277 subjects. We also introduce a new face recognition evaluation protocol which will help advance the state-of-the-art in this area. We discuss how a large dataset can be collected and annotated using human annotators and deep networks. We provide human curated bounding boxes for faces. We also provide estimated pose (roll, pitch and yaw), locations of twenty-one key-points and gender information generated by a pre-trained neural network. In addition, the quality of keypoint annotations has been verified by humans for about 115,000 images. Finally, we compare the quality of the dataset with other publicly available face datasets at similar scales.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124701441","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 191
Latent fingerprint minutia extraction using fully convolutional network 基于全卷积网络的潜在指纹细节提取
Pub Date : 2016-09-30 DOI: 10.1109/BTAS.2017.8272689
Yao Tang, Fei Gao, Jufu Feng
Minutiae play a major role in fingerprint identification. Extracting reliable minutiae is difficult for latent fingerprints which are usually of poor quality. As the limitation of traditional handcrafted features, a fully convolutional network (FCN) is utilized to learn features directly from data to overcome complex background noises. Raw fingerprints are mapped to a correspondingly-sized minutia-score map with a fixed stride. And thus a large number of minutiae will be extracted through a given threshold. Then small regions centering at these minutia points are entered into a convolutional neural network (CNN) to reclassify these minutiae and calculate their orientations. The CNN shares convolutional layers with the fully convolutional network to speed up. 0.45 second is used on average to detect one fingerprint on a GPU. On the NIST SD27 database, we achieve 53% recall rate and 53% precise rate that outperform many other algorithms. Our trained model is also visualized to show that we have successfully extracted features preserving ridge information of a latent fingerprint.
细节在指纹识别中起着重要的作用。潜在指纹通常质量较差,难以提取出可靠的细节信息。针对传统手工特征的局限性,利用全卷积网络(FCN)直接从数据中学习特征,克服复杂的背景噪声。原始指纹被映射到具有固定步幅的相应大小的细节分数映射。因此,通过给定的阈值可以提取大量的细节。然后将以这些细节点为中心的小区域输入到卷积神经网络(CNN)中,对这些细节点进行重新分类并计算它们的方向。CNN与全卷积网络共享卷积层以加快速度。在GPU上检测一个指纹平均耗时0.45秒。在NIST SD27数据库上,我们实现了53%的召回率和53%的精确率,优于许多其他算法。我们的训练模型还被可视化,表明我们成功地提取了保留潜在指纹脊信息的特征。
{"title":"Latent fingerprint minutia extraction using fully convolutional network","authors":"Yao Tang, Fei Gao, Jufu Feng","doi":"10.1109/BTAS.2017.8272689","DOIUrl":"https://doi.org/10.1109/BTAS.2017.8272689","url":null,"abstract":"Minutiae play a major role in fingerprint identification. Extracting reliable minutiae is difficult for latent fingerprints which are usually of poor quality. As the limitation of traditional handcrafted features, a fully convolutional network (FCN) is utilized to learn features directly from data to overcome complex background noises. Raw fingerprints are mapped to a correspondingly-sized minutia-score map with a fixed stride. And thus a large number of minutiae will be extracted through a given threshold. Then small regions centering at these minutia points are entered into a convolutional neural network (CNN) to reclassify these minutiae and calculate their orientations. The CNN shares convolutional layers with the fully convolutional network to speed up. 0.45 second is used on average to detect one fingerprint on a GPU. On the NIST SD27 database, we achieve 53% recall rate and 53% precise rate that outperform many other algorithms. Our trained model is also visualized to show that we have successfully extracted features preserving ridge information of a latent fingerprint.","PeriodicalId":372008,"journal":{"name":"2017 IEEE International Joint Conference on Biometrics (IJCB)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-09-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129264132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
期刊
2017 IEEE International Joint Conference on Biometrics (IJCB)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1