首页 > 最新文献

2019 International Conference on Biometrics (ICB)最新文献

英文 中文
Gait-Based Age Estimation with Deep Convolutional Neural Network 基于步态的深度卷积神经网络年龄估计
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987240
Shaoxiong Zhang, Yunhong Wang, Annan Li
Gait is a unique biometric identifier for its non-invasive and low-cooperative features. Gait-based attribute recognition can play a crucial role in a wide range of applications, such as intelligent surveillance and criminal retrieval. However, due to the lack of data, there are relatively few studies which apply deep convolutional neural networks on gait attribute recognition. In this study, with the new progress in public gait dataset, we proposed a deep convolutional neural network with multi-task learning for gait-based human age estimation. Gait energy images are directly fed into our model for age estimation while gender information is also integrated for improving the performance of age estimation. The experiments on large-scale OULP-Age dataset show that our model outperforms the state-of-the-art.
步态具有非侵入性和低协作性的特点,是一种独特的生物特征识别方法。基于步态的属性识别在智能监控、罪犯检索等领域具有广泛的应用前景。然而,由于数据的缺乏,将深度卷积神经网络应用于步态属性识别的研究相对较少。在这项研究中,我们利用公共步态数据集的新进展,提出了一种基于多任务学习的深度卷积神经网络,用于基于步态的人类年龄估计。将步态能量图像直接输入到模型中进行年龄估计,同时将性别信息集成到模型中以提高年龄估计的性能。在大规模OULP-Age数据集上的实验表明,我们的模型优于最先进的模型。
{"title":"Gait-Based Age Estimation with Deep Convolutional Neural Network","authors":"Shaoxiong Zhang, Yunhong Wang, Annan Li","doi":"10.1109/ICB45273.2019.8987240","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987240","url":null,"abstract":"Gait is a unique biometric identifier for its non-invasive and low-cooperative features. Gait-based attribute recognition can play a crucial role in a wide range of applications, such as intelligent surveillance and criminal retrieval. However, due to the lack of data, there are relatively few studies which apply deep convolutional neural networks on gait attribute recognition. In this study, with the new progress in public gait dataset, we proposed a deep convolutional neural network with multi-task learning for gait-based human age estimation. Gait energy images are directly fed into our model for age estimation while gender information is also integrated for improving the performance of age estimation. The experiments on large-scale OULP-Age dataset show that our model outperforms the state-of-the-art.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127403944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
End-to-End Protocols and Performance Metrics For Unconstrained Face Recognition 无约束人脸识别的端到端协议和性能指标
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987345
James A. Duncan, N. Kalka, Brianna Maze, Anil K. Jain
Face recognition algorithms have received substantial attention over the past decade resulting in significant performance improvements. Arguably, improvement can be attributed to the wide spread availability of large face training sets, GPU computing to train state-of-the-art deep learning algorithms, and curation of challenging test sets that continue to push the state-of-the-art. Traditionally, protocol design and algorithm evaluation have primarily focused on measuring performance of specific stages of the biometric pipeline (e.g., face detection, feature extraction, or recognition) and do not capture errors that may propagate from face input to identification output in an end-to-end (E2E) manner. In this paper, we address this problem by expanding upon the novel open-set E2E identification protocols created for the IARPA Janus program. In particular, we describe in detail the joint detection, tracking, clustering, and recognition protocols, introduce novel E2E performance metrics, and provide rigorous evaluation using the IARPA Janus Benchmark C (IJB-C) and S (IJB-S) datasets.
在过去的十年中,人脸识别算法受到了极大的关注,导致了显着的性能改进。可以说,这种进步可以归功于大型人脸训练集的广泛可用性、用于训练最先进深度学习算法的GPU计算,以及不断推动最先进技术的具有挑战性的测试集的管理。传统上,协议设计和算法评估主要侧重于测量生物识别管道特定阶段的性能(例如,人脸检测、特征提取或识别),而不捕捉可能以端到端(E2E)方式从人脸输入传播到识别输出的错误。在本文中,我们通过扩展为IARPA Janus计划创建的新型开集端到端识别协议来解决这个问题。特别是,我们详细描述了联合检测、跟踪、聚类和识别协议,引入了新的端到端性能指标,并使用IARPA Janus基准C (IJB-C)和S (IJB-S)数据集进行了严格的评估。
{"title":"End-to-End Protocols and Performance Metrics For Unconstrained Face Recognition","authors":"James A. Duncan, N. Kalka, Brianna Maze, Anil K. Jain","doi":"10.1109/ICB45273.2019.8987345","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987345","url":null,"abstract":"Face recognition algorithms have received substantial attention over the past decade resulting in significant performance improvements. Arguably, improvement can be attributed to the wide spread availability of large face training sets, GPU computing to train state-of-the-art deep learning algorithms, and curation of challenging test sets that continue to push the state-of-the-art. Traditionally, protocol design and algorithm evaluation have primarily focused on measuring performance of specific stages of the biometric pipeline (e.g., face detection, feature extraction, or recognition) and do not capture errors that may propagate from face input to identification output in an end-to-end (E2E) manner. In this paper, we address this problem by expanding upon the novel open-set E2E identification protocols created for the IARPA Janus program. In particular, we describe in detail the joint detection, tracking, clustering, and recognition protocols, introduce novel E2E performance metrics, and provide rigorous evaluation using the IARPA Janus Benchmark C (IJB-C) and S (IJB-S) datasets.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122685824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Multi-sample Compression of Finger Vein Images using H.265 Video Coding 基于H.265视频编码的手指静脉图像多样本压缩
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987412
Kevin Schörgnhofer, Sami Dafir, A. Uhl
A new video-compression based approach extending traditional biometric sample data compression techniques is evaluated in the context of finger vein recognition. The proposed scheme is implemented in HEVC / H.265 in different settings and compared to (i) compressing each sample individually with JPEG2000 according to ISO/IEC 19794-9:2011 and to (ii) compressing each users’ data into an individual video file. Compression efficiency and implications on recognition accuracy are determined using 4 recognition schemes and 2 data sets, both based on publicly available data. Results obtained using the proposed approach are fairly stable across different recognition schemes and data sets and indicate a significant improvement over the current state of the art.
在手指静脉识别的背景下,研究了一种基于视频压缩的新方法,扩展了传统的生物特征样本数据压缩技术。提出的方案在不同设置的HEVC / H.265中实现,并比较了(i)根据ISO/IEC 19794-9:2011使用JPEG2000单独压缩每个样本和(ii)将每个用户的数据压缩到单个视频文件中。压缩效率和对识别精度的影响使用4种识别方案和2个数据集,均基于公开可用的数据。使用所提出的方法获得的结果在不同的识别方案和数据集上相当稳定,并且表明比当前技术状态有了显着改进。
{"title":"Multi-sample Compression of Finger Vein Images using H.265 Video Coding","authors":"Kevin Schörgnhofer, Sami Dafir, A. Uhl","doi":"10.1109/ICB45273.2019.8987412","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987412","url":null,"abstract":"A new video-compression based approach extending traditional biometric sample data compression techniques is evaluated in the context of finger vein recognition. The proposed scheme is implemented in HEVC / H.265 in different settings and compared to (i) compressing each sample individually with JPEG2000 according to ISO/IEC 19794-9:2011 and to (ii) compressing each users’ data into an individual video file. Compression efficiency and implications on recognition accuracy are determined using 4 recognition schemes and 2 data sets, both based on publicly available data. Results obtained using the proposed approach are fairly stable across different recognition schemes and data sets and indicate a significant improvement over the current state of the art.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128572187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SeLENet: A Semi-Supervised Low Light Face Enhancement Method for Mobile Face Unlock SeLENet:一种用于移动人脸解锁的半监督低光人脸增强方法
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987344
Ha A. Le, I. Kakadiaris
Facial recognition is becoming a standard feature on new smartphones. However, the face unlocking feature of devices using regular 2D camera sensors exhibits poor performance in low light environments. In this paper, we propose a semi-supervised low light face enhancement method to improve face verification performance on low light face images. The proposed method is a network with two components: decomposition and reconstruction. The decomposition component splits an input low light face image into face normals and face albedo, while the reconstruction component enhances and reconstructs the lighting condition of the input image using the spherical harmonic lighting coefficients of a direct ambient white light. The network is trained in a semi-supervised manner using both labeled synthetic data and unlabeled real data. Qualitative results demonstrate that the proposed method produces more realistic images than the state-of-the-art low light enhancement algorithms. Quantitative experiments confirm the effectiveness of our low light face enhancement method for face verification. By applying the proposed method, the gap of verification accuracy between extreme low light and neutral light face images is reduced from approximately 3% to 0.5%.
面部识别正在成为新型智能手机的标准功能。然而,使用常规2D相机传感器的设备的面部解锁功能在弱光环境下表现不佳。本文提出了一种半监督弱光人脸增强方法,以提高弱光人脸图像的人脸验证性能。该方法是一个由分解和重构两部分组成的网络。分解分量将输入的弱光人脸图像分解为人脸法线和人脸反照率,重构分量利用环境白光直射的球谐光照系数增强和重构输入图像的光照条件。该网络以半监督的方式使用标记的合成数据和未标记的真实数据进行训练。定性结果表明,该方法比目前最先进的弱光增强算法产生更真实的图像。定量实验验证了该方法在人脸验证中的有效性。应用该方法,将极弱光人脸图像与中性光人脸图像的验证精度差距从约3%减小到0.5%。
{"title":"SeLENet: A Semi-Supervised Low Light Face Enhancement Method for Mobile Face Unlock","authors":"Ha A. Le, I. Kakadiaris","doi":"10.1109/ICB45273.2019.8987344","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987344","url":null,"abstract":"Facial recognition is becoming a standard feature on new smartphones. However, the face unlocking feature of devices using regular 2D camera sensors exhibits poor performance in low light environments. In this paper, we propose a semi-supervised low light face enhancement method to improve face verification performance on low light face images. The proposed method is a network with two components: decomposition and reconstruction. The decomposition component splits an input low light face image into face normals and face albedo, while the reconstruction component enhances and reconstructs the lighting condition of the input image using the spherical harmonic lighting coefficients of a direct ambient white light. The network is trained in a semi-supervised manner using both labeled synthetic data and unlabeled real data. Qualitative results demonstrate that the proposed method produces more realistic images than the state-of-the-art low light enhancement algorithms. Quantitative experiments confirm the effectiveness of our low light face enhancement method for face verification. By applying the proposed method, the gap of verification accuracy between extreme low light and neutral light face images is reduced from approximately 3% to 0.5%.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116731589","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
BioPass-UFPB: a Novel Multibiometric Database biopass - fpb:一种新型的多生物特征数据库
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987313
A. Silva, H. Gomes, H. N. Oliveira, P. B. Lins, Diego F. S. Lima, L. Batista
The baseline of a new multibiometric database is presented. The database consists of hand-and-fingerprint images acquired under a controlled environment to produce 5184 × 3456 pixels (hand) and 800×750 with 500 dpi (finger) images. The BioPass-UFPB includes data from 100 individuals and data acquisition, setup and protocols are described, as well as population statistics. Verification and classification results are presented for the hand images in order to provide a baseline for other research projects using these data. BioPass-UFPB is publicly available for research purposes in a conscious effort to improve reproducibility in multimodal biometrics.
提出了一种新的多生物特征数据库的基线。该数据库由在受控环境下获取的手和指纹图像组成,生成5184 × 3456像素(手)和800×750 500 dpi(手指)图像。biopass - uffb包括来自100个人的数据,并描述了数据采集,设置和协议,以及人口统计。验证和分类结果的手图像提出,以便为其他研究项目使用这些数据提供基线。biopass - fpb公开用于研究目的,有意识地努力提高多模态生物识别技术的可重复性。
{"title":"BioPass-UFPB: a Novel Multibiometric Database","authors":"A. Silva, H. Gomes, H. N. Oliveira, P. B. Lins, Diego F. S. Lima, L. Batista","doi":"10.1109/ICB45273.2019.8987313","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987313","url":null,"abstract":"The baseline of a new multibiometric database is presented. The database consists of hand-and-fingerprint images acquired under a controlled environment to produce 5184 × 3456 pixels (hand) and 800×750 with 500 dpi (finger) images. The BioPass-UFPB includes data from 100 individuals and data acquisition, setup and protocols are described, as well as population statistics. Verification and classification results are presented for the hand images in order to provide a baseline for other research projects using these data. BioPass-UFPB is publicly available for research purposes in a conscious effort to improve reproducibility in multimodal biometrics.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121699409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Universal Material Translator: Towards Spoof Fingerprint Generalization 通用材料翻译器:迈向欺骗指纹泛化
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987320
Rohit Gajawada, Additya Popli, T. Chugh, A. Namboodiri, Anil K. Jain
Spoof detectors are classifiers that are trained to distinguish spoof fingerprints from bonafide ones. However, state of the art spoof detectors do not generalize well on unseen spoof materials. This study proposes a style transfer based augmentation wrapper that can be used on any existing spoof detector and can dynamically improve the robustness of the spoof detection system on spoof materials for which we have very low data. Our method is an approach for synthesizing new spoof images from a few spoof examples that transfers the style or material properties of the spoof examples to the content of bonafide fingerprints to generate a larger number of examples to train the classifier on. We demonstrate the effectiveness of our approach on materials in the publicly available LivDet 2015 dataset and show that the proposed approach leads to robustness to fingerprint spoofs of the target material.
欺骗检测器是经过训练来区分欺骗指纹和真实指纹的分类器。然而,最先进的欺骗探测器不能很好地推广看不见的欺骗材料。本研究提出了一种基于风格转移的增强包装器,该包装器可用于任何现有的欺骗检测器,并且可以动态地提高欺骗检测系统对我们拥有非常低数据的欺骗材料的鲁棒性。我们的方法是一种从几个欺骗样本中合成新的欺骗图像的方法,该方法将欺骗样本的风格或材料属性转移到真实指纹的内容中,以生成更多的样本来训练分类器。我们在公开可用的LivDet 2015数据集中证明了我们的方法在材料上的有效性,并表明所提出的方法对目标材料的指纹欺骗具有鲁棒性。
{"title":"Universal Material Translator: Towards Spoof Fingerprint Generalization","authors":"Rohit Gajawada, Additya Popli, T. Chugh, A. Namboodiri, Anil K. Jain","doi":"10.1109/ICB45273.2019.8987320","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987320","url":null,"abstract":"Spoof detectors are classifiers that are trained to distinguish spoof fingerprints from bonafide ones. However, state of the art spoof detectors do not generalize well on unseen spoof materials. This study proposes a style transfer based augmentation wrapper that can be used on any existing spoof detector and can dynamically improve the robustness of the spoof detection system on spoof materials for which we have very low data. Our method is an approach for synthesizing new spoof images from a few spoof examples that transfers the style or material properties of the spoof examples to the content of bonafide fingerprints to generate a larger number of examples to train the classifier on. We demonstrate the effectiveness of our approach on materials in the publicly available LivDet 2015 dataset and show that the proposed approach leads to robustness to fingerprint spoofs of the target material.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129968284","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Adversarial Examples to Fool Iris Recognition Systems 欺骗虹膜识别系统的对抗性示例
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987389
Sobhan Soleymani, Ali Dabouei, J. Dawson, N. Nasrabadi
Adversarial examples have recently proven to be able to fool deep learning methods by adding carefully crafted small perturbation to the input space image. In this paper, we study the possibility of generating adversarial examples for code-based iris recognition systems. Since generating adversarial examples requires back-propagation of the adversarial loss, conventional filter bank-based iris-code generation frameworks cannot be employed in such a setup. Therefore, to compensate for this shortcoming, we propose to train a deep auto-encoder surrogate network to mimic the conventional iris code generation procedure. This trained surrogate network is then deployed to generate the adversarial examples using the iterative gradient sign method algorithm [15]. We consider non-targeted and targeted attacks through three attack scenarios. Considering these attacks, we study the possibility of fooling an iris recognition system in white-box and black-box frameworks.
对抗性示例最近被证明能够通过在输入空间图像中添加精心制作的小扰动来欺骗深度学习方法。在本文中,我们研究了为基于代码的虹膜识别系统生成对抗性示例的可能性。由于生成对抗性示例需要对抗性损失的反向传播,因此传统的基于滤波器组的虹膜代码生成框架不能用于这种设置。因此,为了弥补这一缺点,我们建议训练一个深度自动编码器代理网络来模拟传统的虹膜代码生成过程。然后使用迭代梯度符号法算法部署这个训练好的代理网络来生成对抗性示例[15]。我们通过三种攻击场景来考虑非目标攻击和目标攻击。考虑到这些攻击,我们研究了在白盒和黑盒框架下欺骗虹膜识别系统的可能性。
{"title":"Adversarial Examples to Fool Iris Recognition Systems","authors":"Sobhan Soleymani, Ali Dabouei, J. Dawson, N. Nasrabadi","doi":"10.1109/ICB45273.2019.8987389","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987389","url":null,"abstract":"Adversarial examples have recently proven to be able to fool deep learning methods by adding carefully crafted small perturbation to the input space image. In this paper, we study the possibility of generating adversarial examples for code-based iris recognition systems. Since generating adversarial examples requires back-propagation of the adversarial loss, conventional filter bank-based iris-code generation frameworks cannot be employed in such a setup. Therefore, to compensate for this shortcoming, we propose to train a deep auto-encoder surrogate network to mimic the conventional iris code generation procedure. This trained surrogate network is then deployed to generate the adversarial examples using the iterative gradient sign method algorithm [15]. We consider non-targeted and targeted attacks through three attack scenarios. Considering these attacks, we study the possibility of fooling an iris recognition system in white-box and black-box frameworks.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130951963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Merged Multi-CNN with Parameter Reduction for Face Attribute Estimation 融合参数约简的Multi-CNN人脸属性估计
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987397
Hiroya Kawai, Koichi Ito, T. Aoki
This paper proposes a face attribute estimation method using Merged Multi-CNN (MM-CNN). The proposed method merges single-task CNNs into one CNN by adding merging points and reduces the number of parameters by removing the fully-connected layers. We also propose a new idea of reducing parameters of CNN called Convolutionalization for Parameter Reduction (CPR), which estimates attributes using only convolution layers, in other words, does not need any fully-connected layers to estimate attributes from extracted features. Through a set of experiments using the Celeb A and LFW-a datasets, we demonstrated that MM- CNN with CPR exhibits higher efficiency of face attribute estimation than conventional methods.
提出了一种基于融合多cnn (MM-CNN)的人脸属性估计方法。该方法通过增加合并点将单任务CNN合并为一个CNN,并通过去除全连接层来减少参数的数量。我们还提出了一种减少CNN参数的新思路,称为卷积化参数减少(CPR),它只使用卷积层来估计属性,换句话说,不需要任何全连接层来从提取的特征中估计属性。通过Celeb a和LFW-a数据集的一组实验,我们证明了基于CPR的MM- CNN在人脸属性估计方面比传统方法具有更高的效率。
{"title":"Merged Multi-CNN with Parameter Reduction for Face Attribute Estimation","authors":"Hiroya Kawai, Koichi Ito, T. Aoki","doi":"10.1109/ICB45273.2019.8987397","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987397","url":null,"abstract":"This paper proposes a face attribute estimation method using Merged Multi-CNN (MM-CNN). The proposed method merges single-task CNNs into one CNN by adding merging points and reduces the number of parameters by removing the fully-connected layers. We also propose a new idea of reducing parameters of CNN called Convolutionalization for Parameter Reduction (CPR), which estimates attributes using only convolution layers, in other words, does not need any fully-connected layers to estimate attributes from extracted features. Through a set of experiments using the Celeb A and LFW-a datasets, we demonstrated that MM- CNN with CPR exhibits higher efficiency of face attribute estimation than conventional methods.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133340563","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Conditional Perceptual Adversarial Variational Autoencoder for Age Progression and Regression on Child Face 儿童面部年龄进退的条件知觉对抗变分自编码器
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987410
Praveen Kumar Chandaliya, N. Nain
Recent works have shown that Generative Adversarial Networks (GAN) and Variational Auto-Encoder (VAE) can construct synthetic images of remarkable visual fidelity. In this paper, we propose a novel architecture based on GAN and VAE with Perceptual loss termed as Conditional Perceptual Adversarial Variational Autoencoder (CPAVAE), a model for face aging and rejuvenation on children face. CPAVAE performs face aging and rejuvenation by learning manifold constrained with conditions such as age and gender, which allows it to preserve face identity. CPAVAE uses six networks; these networks are an Encoder (E) and Sampling (S) which maps the child face to latent vector, Generator (G) takes the latent vector z as input along with age conditioned vector and tries to reconstruct the input image, a perceptual loss network Φ, a pre-trained very deep convolution network, discriminator on the encoder (Dz) smoothen’s the age transformation, discriminator on the image (Dimg) forces the generator to produce human realistic images. Here D and E are based on Variational Auto-encoder (VAE) architecture, VGGNet is used as perceptual loss network (Ploss), Dz and Dimg are convolutional neural networks. We represent child face progression and regression on the Children Longitudinal Face(CLF) dataset containing 10752 faces images in the age group [0 : 20]. This dataset contains 6164 and 4588 images of boys and girls respectively.
最近的研究表明,生成对抗网络(GAN)和变分自编码器(VAE)可以构建具有显著视觉保真度的合成图像。在本文中,我们提出了一种新的基于GAN和VAE的具有感知损失的结构,称为条件感知对抗变分自编码器(CPAVAE),这是一种用于儿童面部衰老和年轻化的模型。CPAVAE通过学习受年龄和性别等条件限制的多样性来实现面部衰老和年轻化,从而使其能够保持面部身份。CPAVAE使用六个网络;这些网络是编码器(E)和采样(S),它们将儿童面部映射到潜在向量,生成器(G)将潜在向量z作为输入以及年龄条件向量,并试图重建输入图像,感知损失网络Φ,预训练的非常深的卷积网络,编码器上的鉴别器(Dz)平滑年龄转换,图像上的鉴别器(Dimg)迫使生成器生成人类逼真的图像。其中D和E基于变分自编码器(VAE)架构,VGGNet作为感知损失网络(Ploss), Dz和Dimg是卷积神经网络。我们在儿童纵向面部(Children Longitudinal face, CLF)数据集上表示儿童面部的进展和回归,该数据集包含10752张年龄组的面部图像[0:20]。该数据集分别包含6164和4588张男孩和女孩的图像。
{"title":"Conditional Perceptual Adversarial Variational Autoencoder for Age Progression and Regression on Child Face","authors":"Praveen Kumar Chandaliya, N. Nain","doi":"10.1109/ICB45273.2019.8987410","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987410","url":null,"abstract":"Recent works have shown that Generative Adversarial Networks (GAN) and Variational Auto-Encoder (VAE) can construct synthetic images of remarkable visual fidelity. In this paper, we propose a novel architecture based on GAN and VAE with Perceptual loss termed as Conditional Perceptual Adversarial Variational Autoencoder (CPAVAE), a model for face aging and rejuvenation on children face. CPAVAE performs face aging and rejuvenation by learning manifold constrained with conditions such as age and gender, which allows it to preserve face identity. CPAVAE uses six networks; these networks are an Encoder (E) and Sampling (S) which maps the child face to latent vector, Generator (G) takes the latent vector z as input along with age conditioned vector and tries to reconstruct the input image, a perceptual loss network Φ, a pre-trained very deep convolution network, discriminator on the encoder (Dz) smoothen’s the age transformation, discriminator on the image (Dimg) forces the generator to produce human realistic images. Here D and E are based on Variational Auto-encoder (VAE) architecture, VGGNet is used as perceptual loss network (Ploss), Dz and Dimg are convolutional neural networks. We represent child face progression and regression on the Children Longitudinal Face(CLF) dataset containing 10752 faces images in the age group [0 : 20]. This dataset contains 6164 and 4588 images of boys and girls respectively.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128459554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Hyperspectral Band Selection for Face Recognition Based on a Structurally Sparsified Deep Convolutional Neural Networks 基于结构稀疏化深度卷积神经网络的人脸识别高光谱波段选择
Pub Date : 2019-06-01 DOI: 10.1109/ICB45273.2019.8987360
Fariborz Taherkhani, J. Dawson, N. Nasrabadi
Hyperspectral imaging systems collect and process information from specific wavelengths across the electromagnetic spectrum. The fusion of multi-spectral bands in the visible spectrum has been exploited to improve face recognition performance over conventional broad band face images. In this paper, we propose a new Convolutional Neural Network (CNN) framework which adopts a structural sparsity learning technique to select the optimal spectral bands to obtain the best face recognition performance over all of the spectral bands. Specifically, in this method, all the bands are fed to a CNN and the convolutional filters in the first layer of the CNN are then regularized by employing a group Lasso algorithm to zero out the redundant bands during the training of the network. Contrary to other methods which usually select the bands manually or in a greedy fashion, our method selects the optimal spectral bands automatically to achieve the best face recognition performance over all the spectral bands. Moreover, experimental results demonstrate that our method outperforms state of the art band selection methods for face recognition on several publicly-available hyperspectral face image datasets.
高光谱成像系统收集和处理电磁波谱中特定波长的信息。利用可见光谱中多光谱波段的融合来提高传统宽带人脸图像的识别性能。在本文中,我们提出了一种新的卷积神经网络(CNN)框架,该框架采用结构稀疏性学习技术来选择最优的光谱带,从而在所有光谱带上获得最佳的人脸识别性能。具体来说,在该方法中,将所有的频带都馈送到一个CNN中,然后在CNN的第一层卷积滤波器中使用一组Lasso算法进行正则化,在网络的训练过程中剔除冗余频带。与其他方法通常手动或贪婪地选择波段不同,该方法自动选择最优的光谱波段,从而在所有光谱波段中获得最佳的人脸识别性能。此外,实验结果表明,我们的方法在几个公开可用的高光谱人脸图像数据集上优于最先进的人脸识别波段选择方法。
{"title":"Hyperspectral Band Selection for Face Recognition Based on a Structurally Sparsified Deep Convolutional Neural Networks","authors":"Fariborz Taherkhani, J. Dawson, N. Nasrabadi","doi":"10.1109/ICB45273.2019.8987360","DOIUrl":"https://doi.org/10.1109/ICB45273.2019.8987360","url":null,"abstract":"Hyperspectral imaging systems collect and process information from specific wavelengths across the electromagnetic spectrum. The fusion of multi-spectral bands in the visible spectrum has been exploited to improve face recognition performance over conventional broad band face images. In this paper, we propose a new Convolutional Neural Network (CNN) framework which adopts a structural sparsity learning technique to select the optimal spectral bands to obtain the best face recognition performance over all of the spectral bands. Specifically, in this method, all the bands are fed to a CNN and the convolutional filters in the first layer of the CNN are then regularized by employing a group Lasso algorithm to zero out the redundant bands during the training of the network. Contrary to other methods which usually select the bands manually or in a greedy fashion, our method selects the optimal spectral bands automatically to achieve the best face recognition performance over all the spectral bands. Moreover, experimental results demonstrate that our method outperforms state of the art band selection methods for face recognition on several publicly-available hyperspectral face image datasets.","PeriodicalId":430846,"journal":{"name":"2019 International Conference on Biometrics (ICB)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134457588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2019 International Conference on Biometrics (ICB)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1