首页 > 最新文献

2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)最新文献

英文 中文
Keystroke dynamics recognition based on personal data: A comparative experimental evaluation implementing reproducible research 基于个人数据的击键动力学识别:实现可重复性研究的对比实验评价
A. Morales, Mario Falanga, Julian Fierrez, Carlo Sansone, J. Ortega-Garcia
This work proposes a new benchmark for keystroke dynamics recognition on the basis of fully reproducible research. Instead of traditional authentication approaches based on complex passwords, we propose a novel keystroke recognition based on typing patterns from personal data. We present a new database made up with the keystroke patterns of 63 users and 7560 samples. The proposed approach eliminates the necessity to memorize complex passwords (something that we know) by replacing them by personal data (something that we are). The results encourage to further explore this new application scenario and the availability of data and source code represent a new valuable resource for the research community.
这项工作在完全可重复研究的基础上提出了击键动力学识别的新基准。本文提出了一种基于个人数据输入模式的击键识别方法,取代了传统的基于复杂密码的身份验证方法。我们提出了一个由63个用户和7560个样本组成的新数据库。这个提议的方法通过用个人数据(我们知道的东西)代替它们,消除了记忆复杂密码(我们知道的东西)的必要性。研究结果鼓励进一步探索这种新的应用场景,数据和源代码的可用性为研究界提供了新的有价值的资源。
{"title":"Keystroke dynamics recognition based on personal data: A comparative experimental evaluation implementing reproducible research","authors":"A. Morales, Mario Falanga, Julian Fierrez, Carlo Sansone, J. Ortega-Garcia","doi":"10.1109/BTAS.2015.7358772","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358772","url":null,"abstract":"This work proposes a new benchmark for keystroke dynamics recognition on the basis of fully reproducible research. Instead of traditional authentication approaches based on complex passwords, we propose a novel keystroke recognition based on typing patterns from personal data. We present a new database made up with the keystroke patterns of 63 users and 7560 samples. The proposed approach eliminates the necessity to memorize complex passwords (something that we know) by replacing them by personal data (something that we are). The results encourage to further explore this new application scenario and the availability of data and source code represent a new valuable resource for the research community.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129726058","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Bin-based weak classifier fusion of iris and face biometrics 基于人脸识别的弱分类器虹膜融合
Di Miao, Man Zhang, Haiqing Li, Zhenan Sun, T. Tan
Both high accuracy of iris biometrics and friendly interface of face recognition are important issues to a biometric recognition system. So an open problem is how to combine iris and face biometrics for reliable personal identification. This paper proposes a bin-based weak classifier fusion method for Multibiometrics of Iris and Face. The matching scores of iris and face image patches are partitioned into multiple bins so that the weak classifiers are learned on the bins. Such a non-linear score mapping is simple and efficient but it can discover detailed and distinctive information hidden in matching scores. So that pattern classification performance of the matching scores can be significantly improved. In addition, an ensemble learning method based on boosting is used to select the most discriminant and robust bin-based weak classifiers for identity verification. The excellent performance on the CASIA-Iris-Distance demonstrates the advantages of the proposed method over other multibiometric fusion methods.
虹膜生物识别的高精度和人脸识别的友好界面是生物识别系统的重要组成部分。因此,一个悬而未决的问题是如何结合虹膜和面部生物识别技术来进行可靠的个人识别。提出了一种基于样本的弱分类器融合虹膜和人脸多生物特征的方法。将虹膜和人脸图像贴片的匹配分数划分为多个bin,在这些bin上学习弱分类器。这种非线性的分数映射虽然简单有效,但可以发现隐藏在匹配分数中的详细而独特的信息。从而显著提高匹配分数的模式分类性能。此外,采用基于增强的集成学习方法选择最具判别性和鲁棒性的基于bin的弱分类器进行身份验证。在CASIA-Iris-Distance上的优异性能证明了该方法相对于其他多生物特征融合方法的优越性。
{"title":"Bin-based weak classifier fusion of iris and face biometrics","authors":"Di Miao, Man Zhang, Haiqing Li, Zhenan Sun, T. Tan","doi":"10.1109/BTAS.2015.7358749","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358749","url":null,"abstract":"Both high accuracy of iris biometrics and friendly interface of face recognition are important issues to a biometric recognition system. So an open problem is how to combine iris and face biometrics for reliable personal identification. This paper proposes a bin-based weak classifier fusion method for Multibiometrics of Iris and Face. The matching scores of iris and face image patches are partitioned into multiple bins so that the weak classifiers are learned on the bins. Such a non-linear score mapping is simple and efficient but it can discover detailed and distinctive information hidden in matching scores. So that pattern classification performance of the matching scores can be significantly improved. In addition, an ensemble learning method based on boosting is used to select the most discriminant and robust bin-based weak classifiers for identity verification. The excellent performance on the CASIA-Iris-Distance demonstrates the advantages of the proposed method over other multibiometric fusion methods.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125843231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Iris imaging in visible spectrum using white LED 利用白光LED进行可见光谱虹膜成像
K. Raja, Ramachandra Raghavendra, C. Busch
Iris recognition in the visible spectrum has many challenging aspects. Especially, for subjects with dark iris color, which is caused by higher melanin pigmentation and collagen fibrils, the pattern is not clearly observable under visible light. Thus, the verification performance is generally lowered due to limited texture visibility in the captured iris samples. In this work, we propose a novel method of employing a white light-emitting-diode (LED) to obtain high-quality iris images with detailed texture. To evaluate the proposed set-up with LED light, we have acquired a new database of dark iris images comprising of 62 unique iris instances with ten samples each that were captured in different sessions. The database is acquired using three different smartphones - iPhone 5S, Nokia Lumia 1020 and Samsung Active S4. We also provide a benchmark of the proposed method with conventional to Near-Infra-Red (NIR) images, which are available for a subset of the database. Extensive experiments were carried out using five different well-established iris recognition algorithms and one commercial-of-the-shelf algorithm. They demonstrate the reliable performance of the proposed image capturing setup with GMR of 91.01% at FMR = 0.01% indicating the applicability in real-life authentication scenarios.
虹膜识别在可见光谱中具有许多挑战性。特别是对于虹膜颜色较深的受试者,这是由于黑色素色素沉着和胶原原纤维较高造成的,在可见光下不能清楚地观察到这种模式。因此,由于捕获的虹膜样本中纹理可见性有限,通常会降低验证性能。在这项工作中,我们提出了一种利用白光发光二极管(LED)获得具有详细纹理的高质量虹膜图像的新方法。为了评估使用LED灯的建议设置,我们获得了一个新的暗虹膜图像数据库,其中包括62个独特的虹膜实例,每个实例有10个样本,这些样本是在不同的会话中捕获的。该数据库是通过三种不同的智能手机——iPhone 5S、诺基亚Lumia 1020和三星Active S4获得的。我们还提供了常规到近红外(NIR)图像的基准测试,这些图像可用于数据库的一个子集。使用五种不同的成熟虹膜识别算法和一种商用货架算法进行了广泛的实验。他们证明了所提出的图像捕获设置的可靠性能,在FMR = 0.01%时GMR为91.01%,表明在现实生活中的认证场景中的适用性。
{"title":"Iris imaging in visible spectrum using white LED","authors":"K. Raja, Ramachandra Raghavendra, C. Busch","doi":"10.1109/BTAS.2015.7358769","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358769","url":null,"abstract":"Iris recognition in the visible spectrum has many challenging aspects. Especially, for subjects with dark iris color, which is caused by higher melanin pigmentation and collagen fibrils, the pattern is not clearly observable under visible light. Thus, the verification performance is generally lowered due to limited texture visibility in the captured iris samples. In this work, we propose a novel method of employing a white light-emitting-diode (LED) to obtain high-quality iris images with detailed texture. To evaluate the proposed set-up with LED light, we have acquired a new database of dark iris images comprising of 62 unique iris instances with ten samples each that were captured in different sessions. The database is acquired using three different smartphones - iPhone 5S, Nokia Lumia 1020 and Samsung Active S4. We also provide a benchmark of the proposed method with conventional to Near-Infra-Red (NIR) images, which are available for a subset of the database. Extensive experiments were carried out using five different well-established iris recognition algorithms and one commercial-of-the-shelf algorithm. They demonstrate the reliable performance of the proposed image capturing setup with GMR of 91.01% at FMR = 0.01% indicating the applicability in real-life authentication scenarios.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"115 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115000531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Finger vein liveness detection using motion magnification 用运动放大技术检测手指静脉活动性
Ramachandra Raghavendra, M. Avinash, S. Marcel, C. Busch
Finger vein recognition has emerged as an accurate and reliable biometric modality that was deployed in various security applications. However, the use of finger vein recognition also indicated its vulnerability to presentation attacks (or direct attacks). In this work, we present a novel algorithm to identify the liveness of the finger vein characteristic that is presented to the sensor. The core idea of the proposed approach is to magnify the blood flow through the finger vein to measure its liveness. To this extent, we employ the Eulerian Video Magnification (EVM) approach to enhance the motion of the blood in the recorded finger vein video. Next, we further process the magnified video to extract the motion-based features using optical flow to identify the finger vein artefacts. Extensive experiments are carried out on a relatively large database that is comprised of normal presentations vein videos from 300 unique finger instances corresponding to 100 subjects. The finger vein artefact database is captured by printing 300 real (or normal) presentation image of the finger vein sample on a high-quality paper using two different kinds of printers namely laser and inkjet. Extensive comparative evaluation with four different well-established state-of-the-art schemes demonstrated the efficacy of the proposed scheme.
手指静脉识别已成为一种准确可靠的生物识别技术,在各种安全应用中得到广泛应用。然而,使用手指静脉识别也表明它容易受到演示攻击(或直接攻击)。在这项工作中,我们提出了一种新的算法来识别呈现给传感器的手指静脉特征的活动性。该方法的核心思想是通过放大手指静脉的血流来测量其活动性。在这种程度上,我们采用欧拉视频放大(EVM)方法来增强所记录的手指静脉视频中的血液运动。接下来,我们进一步对放大后的视频进行处理,利用光流提取基于运动的特征来识别手指静脉伪影。广泛的实验是在一个相对较大的数据库中进行的,该数据库由来自100个受试者的300个独特手指实例的正常呈现静脉视频组成。通过使用激光和喷墨两种不同类型的打印机在高质量纸张上打印300个手指静脉样本的真实(或正常)呈现图像来捕获手指静脉人工制品数据库。与四种不同的成熟的最先进的方案进行了广泛的比较评估,证明了拟议方案的有效性。
{"title":"Finger vein liveness detection using motion magnification","authors":"Ramachandra Raghavendra, M. Avinash, S. Marcel, C. Busch","doi":"10.1109/BTAS.2015.7358762","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358762","url":null,"abstract":"Finger vein recognition has emerged as an accurate and reliable biometric modality that was deployed in various security applications. However, the use of finger vein recognition also indicated its vulnerability to presentation attacks (or direct attacks). In this work, we present a novel algorithm to identify the liveness of the finger vein characteristic that is presented to the sensor. The core idea of the proposed approach is to magnify the blood flow through the finger vein to measure its liveness. To this extent, we employ the Eulerian Video Magnification (EVM) approach to enhance the motion of the blood in the recorded finger vein video. Next, we further process the magnified video to extract the motion-based features using optical flow to identify the finger vein artefacts. Extensive experiments are carried out on a relatively large database that is comprised of normal presentations vein videos from 300 unique finger instances corresponding to 100 subjects. The finger vein artefact database is captured by printing 300 real (or normal) presentation image of the finger vein sample on a high-quality paper using two different kinds of printers namely laser and inkjet. Extensive comparative evaluation with four different well-established state-of-the-art schemes demonstrated the efficacy of the proposed scheme.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"535 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128468107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 44
A deep neural network for audio-visual person recognition 一种用于视听人物识别的深度神经网络
Mohammad Rafiqul Alam, Bennamoun, R. Togneri, Ferdous Sohel
This paper presents applications of special types of deep neural networks (DNNs) for audio-visual biometrics. A common example is the DBN-DNN that uses the generative weights of deep belief networks (DBNs) to initialize the feature detecting layers of deterministic feed forward DNNs. In this paper, we propose the DBM-DNN that uses the generative weights of deep Boltzmann machines (DBMs) for initialization of DNNs. Then, a softmax layer is added on top and the DNNs are trained discriminatively. Our experimental results show that lower error rates can be achieved using the DBM-DNN compared to the support vector machine (SVM), linear regression-based classifier (LRC) and the DBN-DNN. Experiments were carried out on two publicly available audio-visual datasets: the VidTIMIT and MOBIO.
本文介绍了特殊类型的深度神经网络(dnn)在视听生物识别中的应用。一个常见的例子是DBN-DNN,它使用深度信念网络(dbn)的生成权值来初始化确定性前馈dnn的特征检测层。在本文中,我们提出了使用深度玻尔兹曼机(DBMs)的生成权值来初始化dnn的DBM-DNN。然后,在上面添加一个softmax层,并对dnn进行判别训练。实验结果表明,与支持向量机(SVM)、基于线性回归的分类器(LRC)和DBN-DNN相比,DBM-DNN可以实现更低的错误率。实验是在VidTIMIT和MOBIO两个公开的视听数据集上进行的。
{"title":"A deep neural network for audio-visual person recognition","authors":"Mohammad Rafiqul Alam, Bennamoun, R. Togneri, Ferdous Sohel","doi":"10.1109/BTAS.2015.7358754","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358754","url":null,"abstract":"This paper presents applications of special types of deep neural networks (DNNs) for audio-visual biometrics. A common example is the DBN-DNN that uses the generative weights of deep belief networks (DBNs) to initialize the feature detecting layers of deterministic feed forward DNNs. In this paper, we propose the DBM-DNN that uses the generative weights of deep Boltzmann machines (DBMs) for initialization of DNNs. Then, a softmax layer is added on top and the DNNs are trained discriminatively. Our experimental results show that lower error rates can be achieved using the DBM-DNN compared to the support vector machine (SVM), linear regression-based classifier (LRC) and the DBN-DNN. Experiments were carried out on two publicly available audio-visual datasets: the VidTIMIT and MOBIO.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132817899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Presentation attack detection using Laplacian decomposed frequency response for visible spectrum and Near-Infra-Red iris systems 基于拉普拉斯分解频率响应的可见光谱和近红外虹膜系统呈现攻击检测
K. Raja, Ramachandra Raghavendra, C. Busch
Biometrics systems are being challenged at the sensor level using artefact presentation such as printed artefacts or electronic screen attacks. In this work, we propose a novel technique to detect the artefact iris images by decomposing the images into Laplacian pyramids of various scales and obtain frequency responses in different orientations. The obtained features are classified using a support vector machine with a polynomial kernel. Further, we extend the same technique with majority voting rule to provide the decision on artefact detection for video based iris recognition in the visible spectrum. The proposed technique is evaluated on the newly created visible spectrum iris video database and also Near-Infra-Red (NIR) images. The newly constructed visible spectrum iris video database is specifically tailored to study the vulnerability of presentation attacks on visible spectrum iris recognition using videos on a smartphone. The newly constructed database is referred as `Presentation Attack Video Iris Database' (PAVID) and consists of 152 unique iris patterns obtained from two different smartphone - iPhone 5S and Nokia Lumia 1020. The proposed technique has provided an Attack Classificiation Error Rate (ACER) of 0.64% on PAVID database and 1.37% on LiveDet iris dataset validating the robustness and applicability of the proposed presentation attack detection (PAD) algorithm in real life scenarios.
生物识别系统在传感器层面受到人工制品呈现的挑战,如印刷人工制品或电子屏幕攻击。在这项工作中,我们提出了一种新的检测人工虹膜图像的技术,通过将图像分解成不同尺度的拉普拉斯金字塔,获得不同方向的频率响应。使用具有多项式核的支持向量机对得到的特征进行分类。此外,我们将相同的技术扩展为多数投票规则,为可见光谱中基于视频的虹膜识别提供伪影检测决策。在新创建的可见光谱虹膜视频数据库和近红外(NIR)图像上对该技术进行了评估。新构建的可见光谱虹膜视频数据库专门针对智能手机上的视频进行可见光谱虹膜识别的呈现攻击漏洞研究。新构建的数据库被称为“演示攻击视频虹膜数据库”(PAVID),由152种独特的虹膜模式组成,这些虹膜模式来自两款不同的智能手机——iPhone 5S和诺基亚Lumia 1020。该技术在PAVID数据库和LiveDet虹膜数据集上的攻击分类错误率(ACER)分别为0.64%和1.37%,验证了所提出的呈现攻击检测(PAD)算法在现实场景中的鲁棒性和适用性。
{"title":"Presentation attack detection using Laplacian decomposed frequency response for visible spectrum and Near-Infra-Red iris systems","authors":"K. Raja, Ramachandra Raghavendra, C. Busch","doi":"10.1109/BTAS.2015.7358790","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358790","url":null,"abstract":"Biometrics systems are being challenged at the sensor level using artefact presentation such as printed artefacts or electronic screen attacks. In this work, we propose a novel technique to detect the artefact iris images by decomposing the images into Laplacian pyramids of various scales and obtain frequency responses in different orientations. The obtained features are classified using a support vector machine with a polynomial kernel. Further, we extend the same technique with majority voting rule to provide the decision on artefact detection for video based iris recognition in the visible spectrum. The proposed technique is evaluated on the newly created visible spectrum iris video database and also Near-Infra-Red (NIR) images. The newly constructed visible spectrum iris video database is specifically tailored to study the vulnerability of presentation attacks on visible spectrum iris recognition using videos on a smartphone. The newly constructed database is referred as `Presentation Attack Video Iris Database' (PAVID) and consists of 152 unique iris patterns obtained from two different smartphone - iPhone 5S and Nokia Lumia 1020. The proposed technique has provided an Attack Classificiation Error Rate (ACER) of 0.64% on PAVID database and 1.37% on LiveDet iris dataset validating the robustness and applicability of the proposed presentation attack detection (PAD) algorithm in real life scenarios.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133281580","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Pokerface: Partial order keeping and energy repressing method for extreme face illumination normalization 极值面照度归一化的偏序保持和能量抑制方法
Felix Juefei-Xu, M. Savvides
We propose a new method called the Pokerface for extreme face illumination normalization. The Pokerface is a two-phase approach. It first aims at maximizing the minimum gap between adjacently-valued pixels while keeping the partial ordering of the pixels in the face image under extreme illumination condition, an intuitive effort based on order theory to unveil the underlying structure of a dark image. This optimization can be formulated as a feasibility search problem and can be efficiently solved by linear programming. It then smooths the intermediate representation by repressing the energy of the gradient map. The smoothing step is carried out by total variation minimization and sparse approximation. The illumination normalized faces using our proposed Pokerface not only exhibit very high fidelity against neutrally illuminated face, but also allow for a significant improvement in face verification experiments using even the simplest classifier. Simultaneously achieving high level of faithfulness and expressiveness is very rare among other methods. These conclusions are drawn after benchmarking our algorithm against 22 prevailing illumination normalization techniques on both the CMU Multi-PIE database and Extended YaleB database that are widely adopted for face illumination problems.
我们提出了一种名为Pokerface的极端人脸光照归一化方法。Pokerface是一种两阶段的方法。首先,它的目标是在极端光照条件下,在保持人脸图像中像素的偏序的同时,最大化相邻像素之间的最小间隙,这是一种基于序理论的直观的努力,以揭示黑暗图像的底层结构。该优化问题可表述为可行性搜索问题,并可通过线性规划有效地求解。然后通过抑制梯度映射的能量来平滑中间表示。平滑步骤采用总变差最小化和稀疏逼近方法。使用我们提出的Pokerface的光照归一化人脸不仅对中性光照的人脸表现出非常高的保真度,而且即使使用最简单的分类器,也可以显著改善人脸验证实验。在其他方法中,同时达到高水平的忠实性和表现力是非常罕见的。这些结论是在CMU Multi-PIE数据库和Extended YaleB数据库上与22种主流照明归一化技术进行基准测试后得出的,这些技术被广泛用于人脸照明问题。
{"title":"Pokerface: Partial order keeping and energy repressing method for extreme face illumination normalization","authors":"Felix Juefei-Xu, M. Savvides","doi":"10.1109/BTAS.2015.7358787","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358787","url":null,"abstract":"We propose a new method called the Pokerface for extreme face illumination normalization. The Pokerface is a two-phase approach. It first aims at maximizing the minimum gap between adjacently-valued pixels while keeping the partial ordering of the pixels in the face image under extreme illumination condition, an intuitive effort based on order theory to unveil the underlying structure of a dark image. This optimization can be formulated as a feasibility search problem and can be efficiently solved by linear programming. It then smooths the intermediate representation by repressing the energy of the gradient map. The smoothing step is carried out by total variation minimization and sparse approximation. The illumination normalized faces using our proposed Pokerface not only exhibit very high fidelity against neutrally illuminated face, but also allow for a significant improvement in face verification experiments using even the simplest classifier. Simultaneously achieving high level of faithfulness and expressiveness is very rare among other methods. These conclusions are drawn after benchmarking our algorithm against 22 prevailing illumination normalization techniques on both the CMU Multi-PIE database and Extended YaleB database that are widely adopted for face illumination problems.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133592764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
On the vulnerability of speaker verification to realistic voice spoofing 说话人验证对现实语音欺骗的脆弱性研究
Serife Seda Kucur Ergunay, E. Khoury, Alexandros Lazaridis, S. Marcel
Automatic speaker verification (ASV) systems are subject to various kinds of malicious attacks. Replay, voice conversion and speech synthesis attacks drastically degrade the performance of a standard ASV system by increasing its false acceptance rates. This issue raised a high level of interest in the speech research community where the possible voice spoofing attacks and their related countermeasures have been investigated. However, much less effort has been devoted in creating realistic and diverse spoofing attack databases that foster researchers to correctly evaluate their countermeasures against attacks. The existing studies are not complete in terms of types of attacks, and often difficult to reproduce because of unavailability of public databases. In this paper we introduce the voice spoofing data-set of AVspoof, a public audio-visual spoofing database. AVspoof includes ten realistic spoofing threats generated using replay, speech synthesis and voice conversion. In addition, we provide a set of experimental results that show the effect of such attacks on current state-of-the-art ASV systems.
自动说话人验证(ASV)系统经常受到各种恶意攻击。重放、语音转换和语音合成攻击通过增加其错误接受率,大大降低了标准ASV系统的性能。这个问题引起了语音研究界的高度关注,他们正在研究可能的语音欺骗攻击及其相关对策。然而,在创建现实的和多样化的欺骗攻击数据库方面投入的努力要少得多,这些数据库可以促进研究人员正确评估针对攻击的对策。就攻击类型而言,现有的研究并不完整,而且由于缺乏公共数据库,往往难以复制。本文介绍了公共视听欺骗数据库AVspoof的语音欺骗数据集。AVspoof包括使用重放、语音合成和语音转换生成的十个现实欺骗威胁。此外,我们提供了一组实验结果,显示了这种攻击对当前最先进的ASV系统的影响。
{"title":"On the vulnerability of speaker verification to realistic voice spoofing","authors":"Serife Seda Kucur Ergunay, E. Khoury, Alexandros Lazaridis, S. Marcel","doi":"10.1109/BTAS.2015.7358783","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358783","url":null,"abstract":"Automatic speaker verification (ASV) systems are subject to various kinds of malicious attacks. Replay, voice conversion and speech synthesis attacks drastically degrade the performance of a standard ASV system by increasing its false acceptance rates. This issue raised a high level of interest in the speech research community where the possible voice spoofing attacks and their related countermeasures have been investigated. However, much less effort has been devoted in creating realistic and diverse spoofing attack databases that foster researchers to correctly evaluate their countermeasures against attacks. The existing studies are not complete in terms of types of attacks, and often difficult to reproduce because of unavailability of public databases. In this paper we introduce the voice spoofing data-set of AVspoof, a public audio-visual spoofing database. AVspoof includes ten realistic spoofing threats generated using replay, speech synthesis and voice conversion. In addition, we provide a set of experimental results that show the effect of such attacks on current state-of-the-art ASV systems.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130694688","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 115
Towards fitting a 3D dense facial model to a 2D image: A landmark-free approach 拟合三维密集面部模型到二维图像:无地标方法
Yuhang Wu, Xiang Xu, S. Shah, I. Kakadiaris
Head pose estimation helps to align a 3D face model to a 2D image, which is critical to research requiring dense 2D-to-2D or 3D-to-2D correspondence. Traditional pose estimation relies strongly on the accuracy of landmarks, so it is sensitive to missing or incorrect landmarks. In this paper, we propose a landmark-free approach to estimate the pose projection matrix. The method can be used to estimate this matrix in unconstrained scenarios and we demonstrate its effectiveness through multiple head pose estimation experiments.
头部姿态估计有助于将3D面部模型与2D图像对齐,这对于需要密集2D到2D或3D到2D对应的研究至关重要。传统的姿态估计强烈依赖于地标的准确性,因此对缺失或错误的地标很敏感。在本文中,我们提出了一种无地标的姿态投影矩阵估计方法。该方法可用于在无约束情况下估计该矩阵,并通过多个头姿估计实验证明了其有效性。
{"title":"Towards fitting a 3D dense facial model to a 2D image: A landmark-free approach","authors":"Yuhang Wu, Xiang Xu, S. Shah, I. Kakadiaris","doi":"10.1109/BTAS.2015.7358799","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358799","url":null,"abstract":"Head pose estimation helps to align a 3D face model to a 2D image, which is critical to research requiring dense 2D-to-2D or 3D-to-2D correspondence. Traditional pose estimation relies strongly on the accuracy of landmarks, so it is sensitive to missing or incorrect landmarks. In this paper, we propose a landmark-free approach to estimate the pose projection matrix. The method can be used to estimate this matrix in unconstrained scenarios and we demonstrate its effectiveness through multiple head pose estimation experiments.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"30 14","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114116027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Human and algorithm performance on the PaSC face Recognition Challenge 人类和算法在PaSC人脸识别挑战中的表现
P. Phillips, Matthew Q. Hill, Jake A. Swindle, A. O’Toole
Face recognition by machines has improved substantially in the past decade and now is at a level that compares favorably with humans for frontal faces acquired by digital single lens reflex cameras. We expand the comparison between humans and algorithms to still images and videos taken with digital point and shoot cameras. The data used for this comparison are from the Point and Shoot Face Recognition Challenge (PaSC). For videos, human performance was compared with the four top performers in the Face and Gesture 2015 Person Recognition Evaluation. In the literature, there are two methods for computing human performance: aggregation and fusion. We show that the fusion method produces higher performance estimates. We report performance for two levels of difficulty: challenging and extremely-difficult. Our results provide additional evidence that human performance shines relative to algorithms on extremely-difficult comparisons. To improve the community's understanding of the state of human and algorithm performance, we update the cross-modal performance analysis in Phillips and O'Toole [22] with these new results.
在过去的十年里,机器的面部识别已经有了很大的进步,现在已经达到了与人类相比的水平,可以通过数码单镜头反光相机获得正面的面部。我们将人与算法之间的比较扩展到用数码相机拍摄的静态图像和视频。用于比较的数据来自Point and Shoot Face Recognition Challenge (PaSC)。对于视频,人类的表现与2015年人脸和手势识别评估中的前四名表现最好的人进行了比较。在文献中,有两种计算人类表现的方法:聚合和融合。我们表明,融合方法产生更高的性能估计。我们根据两个难度等级来报告游戏表现:具有挑战性的和极难的。我们的研究结果提供了额外的证据,证明在极其困难的比较中,人类的表现相对于算法更出色。为了提高社区对人和算法性能状态的理解,我们用这些新结果更新了Phillips和O'Toole[22]的跨模态性能分析。
{"title":"Human and algorithm performance on the PaSC face Recognition Challenge","authors":"P. Phillips, Matthew Q. Hill, Jake A. Swindle, A. O’Toole","doi":"10.1109/BTAS.2015.7358765","DOIUrl":"https://doi.org/10.1109/BTAS.2015.7358765","url":null,"abstract":"Face recognition by machines has improved substantially in the past decade and now is at a level that compares favorably with humans for frontal faces acquired by digital single lens reflex cameras. We expand the comparison between humans and algorithms to still images and videos taken with digital point and shoot cameras. The data used for this comparison are from the Point and Shoot Face Recognition Challenge (PaSC). For videos, human performance was compared with the four top performers in the Face and Gesture 2015 Person Recognition Evaluation. In the literature, there are two methods for computing human performance: aggregation and fusion. We show that the fusion method produces higher performance estimates. We report performance for two levels of difficulty: challenging and extremely-difficult. Our results provide additional evidence that human performance shines relative to algorithms on extremely-difficult comparisons. To improve the community's understanding of the state of human and algorithm performance, we update the cross-modal performance analysis in Phillips and O'Toole [22] with these new results.","PeriodicalId":404972,"journal":{"name":"2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)","volume":"98 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124887130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
期刊
2015 IEEE 7th International Conference on Biometrics Theory, Applications and Systems (BTAS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1