首页 > 最新文献

2023 11th International Workshop on Biometrics and Forensics (IWBF)最新文献

英文 中文
CP-Net: Multi-Scale Core Point Localization in Fingerprints Using Hourglass Network CP-Net:基于沙漏网络的指纹多尺度核心点定位
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157521
Geetika Arora, Arsh Kumbhat, Ashutosh Bhatia, Kamlesh Tiwari
Core point is a location that exhibits high curvature properties in a fingerprint. Detecting the accurate location of a core point is useful for efficient fingerprint matching, classification, and identification tasks. This paper proposes CP-Net, a novel core point detection network that comprises the Macro Localization Network (MLN) and the Micro-Regression Network (MRN). MLN is a specialized autoencoder network with an hourglass network at its bottleneck. It takes an input fingerprint image and outputs a region of interest that could be the most probable region containing the core point. The second component, MRN, regresses the RoI and locates the coordinates of the core point in the given fingerprint sample. Introducing an hourglass network in the MLN bottleneck ensures multi-scale spatial attention that captures local and global contexts and facilitates a higher localization accuracy for that area. Unlike existing multi-stage models, the components are stacked and trained in an end-to-end manner. Experiments have been performed on three widely used publicly available fingerprint datasets, namely, FVC2002 DB1A, FVC2004 DB1A, and FVC2006 DB2A. The proposed model achieved a true detection rate (TDR) of 98%, 100%, and 99.04% respectively, while considering 20 pixels distance from the ground truth as correct. Obtained experimental results on the considered datasets demonstrate that CP-Net outperforms the state-of-the-art core point detection techniques.
核心点是指纹中显示高曲率特性的位置。检测核心点的准确位置对于高效的指纹匹配、分类和识别任务非常有用。本文提出了一种新的核心点检测网络CP-Net,该网络由宏观定位网络(MLN)和微回归网络(MRN)组成。MLN是一种特殊的自编码器网络,其瓶颈是沙漏网络。它输入指纹图像并输出感兴趣的区域,该区域可能是包含核心点的最可能区域。第二个组件MRN对RoI进行回归并定位给定指纹样本中核心点的坐标。在MLN瓶颈中引入沙漏网络可确保捕获局部和全局上下文的多尺度空间注意力,并促进该区域的更高定位精度。与现有的多阶段模型不同,组件以端到端的方式堆叠和训练。在FVC2002 DB1A、FVC2004 DB1A和FVC2006 DB2A三个广泛使用的公开指纹数据集上进行了实验。在考虑距离地面20像素距离的情况下,该模型的真实检测率(TDR)分别为98%、100%和99.04%。在考虑的数据集上获得的实验结果表明,CP-Net优于最先进的核心点检测技术。
{"title":"CP-Net: Multi-Scale Core Point Localization in Fingerprints Using Hourglass Network","authors":"Geetika Arora, Arsh Kumbhat, Ashutosh Bhatia, Kamlesh Tiwari","doi":"10.1109/IWBF57495.2023.10157521","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157521","url":null,"abstract":"Core point is a location that exhibits high curvature properties in a fingerprint. Detecting the accurate location of a core point is useful for efficient fingerprint matching, classification, and identification tasks. This paper proposes CP-Net, a novel core point detection network that comprises the Macro Localization Network (MLN) and the Micro-Regression Network (MRN). MLN is a specialized autoencoder network with an hourglass network at its bottleneck. It takes an input fingerprint image and outputs a region of interest that could be the most probable region containing the core point. The second component, MRN, regresses the RoI and locates the coordinates of the core point in the given fingerprint sample. Introducing an hourglass network in the MLN bottleneck ensures multi-scale spatial attention that captures local and global contexts and facilitates a higher localization accuracy for that area. Unlike existing multi-stage models, the components are stacked and trained in an end-to-end manner. Experiments have been performed on three widely used publicly available fingerprint datasets, namely, FVC2002 DB1A, FVC2004 DB1A, and FVC2006 DB2A. The proposed model achieved a true detection rate (TDR) of 98%, 100%, and 99.04% respectively, while considering 20 pixels distance from the ground truth as correct. Obtained experimental results on the considered datasets demonstrate that CP-Net outperforms the state-of-the-art core point detection techniques.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123730504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty-aware Comparison Scores for Face Recognition 人脸识别的不确定性感知比较分数
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157282
Marco Huber, Philipp Terhörst, Florian Kirchbuchner, Arjan Kuijper, N. Damer
Estimating and understanding uncertainty in face recognition systems is receiving increasing attention as face recognition systems spread worldwide and process privacy and security-related data. In this work, we investigate how such uncertainties can be further utilized to increase the accuracy and therefore the trust of automatic face recognition systems. We propose to use the uncertainties of extracted face features to compute a new uncertainty-aware comparison score (UACS). This score takes into account the estimated uncertainty during the calculation of the comparison score, leading to a reduction in verification errors. To achieve this, we model the comparison score and its uncertainty as a probability distribution and measure its distance to a distribution of an ideal genuine comparison. In extended experiments with three face recognition models and on six benchmarks, we investigated the impact of our approach and demonstrated its benefits in enhancing the verification performance and the genuine-imposter comparison scores separability.
随着人脸识别系统在全球范围内的普及以及处理隐私和安全相关数据,对人脸识别系统中不确定性的估计和理解越来越受到关注。在这项工作中,我们研究了如何进一步利用这些不确定性来提高准确性,从而提高自动人脸识别系统的信任度。我们提出利用提取的人脸特征的不确定性来计算新的不确定性感知比较分数(UACS)。该分数在计算比较分数时考虑了估计的不确定度,从而减少了验证误差。为了实现这一点,我们将比较分数及其不确定性建模为概率分布,并测量其与理想真实比较分布的距离。在三个人脸识别模型和六个基准的扩展实验中,我们研究了我们的方法的影响,并证明了它在提高验证性能和真假比较分数可分性方面的好处。
{"title":"Uncertainty-aware Comparison Scores for Face Recognition","authors":"Marco Huber, Philipp Terhörst, Florian Kirchbuchner, Arjan Kuijper, N. Damer","doi":"10.1109/IWBF57495.2023.10157282","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157282","url":null,"abstract":"Estimating and understanding uncertainty in face recognition systems is receiving increasing attention as face recognition systems spread worldwide and process privacy and security-related data. In this work, we investigate how such uncertainties can be further utilized to increase the accuracy and therefore the trust of automatic face recognition systems. We propose to use the uncertainties of extracted face features to compute a new uncertainty-aware comparison score (UACS). This score takes into account the estimated uncertainty during the calculation of the comparison score, leading to a reduction in verification errors. To achieve this, we model the comparison score and its uncertainty as a probability distribution and measure its distance to a distribution of an ideal genuine comparison. In extended experiments with three face recognition models and on six benchmarks, we investigated the impact of our approach and demonstrated its benefits in enhancing the verification performance and the genuine-imposter comparison scores separability.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132021649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Optimization-Based Improvement of Face Image Quality Assessment Techniques 基于优化的人脸图像质量评估技术改进
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157796
Ziga Babnik, N. Damer, V. Štruc
Contemporary face recognition (FR) models achieve near-ideal recognition performance in constrained settings, yet do not fully translate the performance to unconstrained (real-world) scenarios. To help improve the performance and stability of FR systems in such unconstrained settings, face image quality assessment (FIQA) techniques try to infer sample-quality information from the input face images that can aid with the recognition process. While existing FIQA techniques are able to efficiently capture the differences between high and low quality images, they typically cannot fully distinguish between images of similar quality, leading to lower performance in many scenarios. To address this issue, we present in this paper a supervised quality-label optimization approach, aimed at improving the performance of existing FIQA techniques. The developed optimization procedure infuses additional information (computed with a selected FR model) into the initial quality scores generated with a given FIQA technique to produce better estimates of the “actual” image quality. We evaluate the proposed approach in comprehensive experiments with six state-of-the-art FIQA approaches (CR-FIQA, FaceQAN, SER-FIQ, PCNet, MagFace, SER-FIQ) on five commonly used benchmarks (LFW, CFP-FP, CPLFW, CALFW, XQLFW) using three targeted FR models (ArcFace, ElasticFace, CurricularFace) with highly encouraging results.
当代人脸识别(FR)模型在受限环境中实现了近乎理想的识别性能,但不能完全将性能转化为无约束(现实世界)场景。为了帮助提高人脸识别系统在这种无约束环境下的性能和稳定性,人脸图像质量评估(FIQA)技术试图从输入的人脸图像中推断出有助于识别过程的样本质量信息。虽然现有的FIQA技术能够有效地捕获高质量和低质量图像之间的差异,但它们通常不能完全区分相似质量的图像,导致在许多情况下性能较低。为了解决这个问题,我们提出了一种监督质量标签优化方法,旨在提高现有FIQA技术的性能。开发的优化程序将附加信息(用选定的FR模型计算)注入到使用给定FIQA技术生成的初始质量分数中,以产生对“实际”图像质量的更好估计。我们使用六种最先进的FIQA方法(CR-FIQA, FaceQAN, SER-FIQ, PCNet, MagFace, SER-FIQ)在五个常用基准(LFW, CFP-FP, CPLFW, CALFW, XQLFW)上使用三种目标FR模型(ArcFace, ElasticFace, CurricularFace)在综合实验中评估了所提出的方法,并获得了非常令人鼓舞的结果。
{"title":"Optimization-Based Improvement of Face Image Quality Assessment Techniques","authors":"Ziga Babnik, N. Damer, V. Štruc","doi":"10.1109/IWBF57495.2023.10157796","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157796","url":null,"abstract":"Contemporary face recognition (FR) models achieve near-ideal recognition performance in constrained settings, yet do not fully translate the performance to unconstrained (real-world) scenarios. To help improve the performance and stability of FR systems in such unconstrained settings, face image quality assessment (FIQA) techniques try to infer sample-quality information from the input face images that can aid with the recognition process. While existing FIQA techniques are able to efficiently capture the differences between high and low quality images, they typically cannot fully distinguish between images of similar quality, leading to lower performance in many scenarios. To address this issue, we present in this paper a supervised quality-label optimization approach, aimed at improving the performance of existing FIQA techniques. The developed optimization procedure infuses additional information (computed with a selected FR model) into the initial quality scores generated with a given FIQA technique to produce better estimates of the “actual” image quality. We evaluate the proposed approach in comprehensive experiments with six state-of-the-art FIQA approaches (CR-FIQA, FaceQAN, SER-FIQ, PCNet, MagFace, SER-FIQ) on five commonly used benchmarks (LFW, CFP-FP, CPLFW, CALFW, XQLFW) using three targeted FR models (ArcFace, ElasticFace, CurricularFace) with highly encouraging results.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121948022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
On the influence of the quality of pseudo-labels on the self-supervised speaker verification task: a thorough analysis 伪标签质量对自监督说话人验证任务的影响分析
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157651
A. Fathan, J. Alam
One of the most widely used self-supervised (SS) speaker verification (SV) system training methods is to optimize the speaker embedding network in a discriminative fashion using clustering algorithm (CA)-driven Pseudo-Labels (PLs). Although the PL-based SS training scheme showed impressive performance, recent studies have shown that label noise can significantly impact performance. In this paper, we have explored various PLs driven by different CAs and conducted a fine-grained analysis of the relationship between the quality of the PLs and the SV performance. Experimentally, we shed light on several previously overlooked aspects of the PLs that can impact SV performance. Moreover, we could observe that the SS-SV performance is heavily dependent on multiple qualitative aspects of the CA used to generate the PLs. Furthermore, we show that SV performance can be severely degraded from overfitting the noisy PLs and that the mixup strategy can mitigate the memorization effects of label noise.
自监督说话人验证(SV)系统训练中应用最广泛的一种方法是利用聚类算法(CA)驱动的伪标签(PLs)以判别方式优化说话人嵌入网络。尽管基于pl的SS训练方案表现出令人印象深刻的性能,但最近的研究表明,标签噪声会显著影响性能。在本文中,我们探讨了由不同ca驱动的各种PLs,并对PLs质量与SV性能之间的关系进行了细致的分析。通过实验,我们揭示了几个以前被忽视的可能影响SV性能的PLs方面。此外,我们可以观察到,SS-SV的性能严重依赖于用于生成PLs的CA的多个定性方面。此外,我们表明,SV的性能可能会因过度拟合有噪声的PLs而严重下降,并且混合策略可以减轻标签噪声的记忆影响。
{"title":"On the influence of the quality of pseudo-labels on the self-supervised speaker verification task: a thorough analysis","authors":"A. Fathan, J. Alam","doi":"10.1109/IWBF57495.2023.10157651","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157651","url":null,"abstract":"One of the most widely used self-supervised (SS) speaker verification (SV) system training methods is to optimize the speaker embedding network in a discriminative fashion using clustering algorithm (CA)-driven Pseudo-Labels (PLs). Although the PL-based SS training scheme showed impressive performance, recent studies have shown that label noise can significantly impact performance. In this paper, we have explored various PLs driven by different CAs and conducted a fine-grained analysis of the relationship between the quality of the PLs and the SV performance. Experimentally, we shed light on several previously overlooked aspects of the PLs that can impact SV performance. Moreover, we could observe that the SS-SV performance is heavily dependent on multiple qualitative aspects of the CA used to generate the PLs. Furthermore, we show that SV performance can be severely degraded from overfitting the noisy PLs and that the mixup strategy can mitigate the memorization effects of label noise.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127200154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved Likelihood Ratios for Surveillance Video Face Recognition with Multimodal Feature Pairing 基于多模态特征配对的改进似然比监控视频人脸识别
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157791
Andrea Macarulla Rodriguez, Z. Geradts, M. Worring, Luis Unzueta
The accuracy of face recognition in real-world surveillance videos plays a crucial role in forensic investigation and security monitoring systems. Despite advancements in technology, face recognition methods can be influenced by variations in pose, illumination, and facial expression that often occur in these videos. To address this issue, we propose a new method for images-to-video face recognition that pairs face images with multiple attributes (soft labels) and face image quality (FIQ). This is followed by the application of three calibration methods to estimate the likelihood ratio, which is a statistical measure commonly used in forensic investigations. To validate the results, we test our method on the ENFSI proficiency test 2015 dataset, using SCFace and ForenFace as calibration datasets and three embedding models: ArcFace, FaceNet, and QMagFace. Our results indicate that using only high quality frames can improve face recognition performance for forensic purposes compared to using all frames. The best results were achieved when using the highest number of common attributes between the reference image and selected frames, or by creating a single common embedding from the selected frames, weighted by the quality of each frame’s face image.
在现实世界的监控视频中,人脸识别的准确性在法医调查和安全监控系统中起着至关重要的作用。尽管技术不断进步,但人脸识别方法可能会受到这些视频中经常出现的姿势、照明和面部表情变化的影响。为了解决这个问题,我们提出了一种新的图像到视频的人脸识别方法,该方法将具有多个属性(软标签)和人脸图像质量(FIQ)的人脸图像配对。其次是应用三种校准方法来估计似然比,这是法医调查中常用的统计度量。为了验证结果,我们在2015年ENFSI熟练测试数据集上测试了我们的方法,使用SCFace和ForenFace作为校准数据集和三种嵌入模型:ArcFace, FaceNet和QMagFace。我们的研究结果表明,与使用所有帧相比,仅使用高质量帧可以提高用于法医目的的人脸识别性能。当在参考图像和选定帧之间使用最大数量的共同属性时,或者通过从选定帧中创建单个共同嵌入,并根据每帧人脸图像的质量加权时,可以获得最佳结果。
{"title":"Improved Likelihood Ratios for Surveillance Video Face Recognition with Multimodal Feature Pairing","authors":"Andrea Macarulla Rodriguez, Z. Geradts, M. Worring, Luis Unzueta","doi":"10.1109/IWBF57495.2023.10157791","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157791","url":null,"abstract":"The accuracy of face recognition in real-world surveillance videos plays a crucial role in forensic investigation and security monitoring systems. Despite advancements in technology, face recognition methods can be influenced by variations in pose, illumination, and facial expression that often occur in these videos. To address this issue, we propose a new method for images-to-video face recognition that pairs face images with multiple attributes (soft labels) and face image quality (FIQ). This is followed by the application of three calibration methods to estimate the likelihood ratio, which is a statistical measure commonly used in forensic investigations. To validate the results, we test our method on the ENFSI proficiency test 2015 dataset, using SCFace and ForenFace as calibration datasets and three embedding models: ArcFace, FaceNet, and QMagFace. Our results indicate that using only high quality frames can improve face recognition performance for forensic purposes compared to using all frames. The best results were achieved when using the highest number of common attributes between the reference image and selected frames, or by creating a single common embedding from the selected frames, weighted by the quality of each frame’s face image.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129119377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fingerprint quality per individual finger type: A large-scale study on real operational data 每个手指类型的指纹质量:对真实操作数据的大规模研究
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157166
Javier Galbally, A. Cepilovs, R. Blanco-Gonzalo, G. Ormiston, O. Miguel-Hurtado, I. S. Racz
Even though some initial works have shown on small sets of data that not all fingerprints present the same level of utility for recognition purposes, there is still insufficient data-supported evidence to understand the impact that finger type may have on fingerprint quality and, in turn, also on fingerprint comparison. The present work addresses this still under-researched topic, on a large-scale database of operational data containing 10-print impressions of over 18,000 subjects. The results show a noticeable difference in the quality level of fingerprints produced by each of the 10 fingers and also between the dominant and non-dominant hands. Based on these observations, several recommendations are made regarding: 1) the selection of fingers to be captured depending on the context of the application; 2) improvement in the usability of scanners and the capturing protocols; 3) improvement in the development, ergonomics and positioning of the acquisition devices; and 4) improvement of recognition algorithms by incorporating information on finger type and handedness.
尽管一些初步的工作已经在小数据集上表明,并非所有指纹在识别目的上都具有相同的效用水平,但仍然没有足够的数据支持证据来理解手指类型可能对指纹质量的影响,进而对指纹比较的影响。目前的工作解决了这一尚未充分研究的主题,在一个包含超过18,000个主题的10个印刷印象的大型操作数据数据库上。结果显示,这10个手指产生的指纹质量水平存在显著差异,而且优势手和非优势手之间的指纹质量水平也存在显著差异。基于这些观察,提出了以下几点建议:1)根据应用程序的上下文选择要捕获的手指;2)提高扫描仪和捕获协议的可用性;3)改进采集设备的开发、人机工程学和定位;4)结合手指类型和惯用手性信息改进识别算法。
{"title":"Fingerprint quality per individual finger type: A large-scale study on real operational data","authors":"Javier Galbally, A. Cepilovs, R. Blanco-Gonzalo, G. Ormiston, O. Miguel-Hurtado, I. S. Racz","doi":"10.1109/IWBF57495.2023.10157166","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157166","url":null,"abstract":"Even though some initial works have shown on small sets of data that not all fingerprints present the same level of utility for recognition purposes, there is still insufficient data-supported evidence to understand the impact that finger type may have on fingerprint quality and, in turn, also on fingerprint comparison. The present work addresses this still under-researched topic, on a large-scale database of operational data containing 10-print impressions of over 18,000 subjects. The results show a noticeable difference in the quality level of fingerprints produced by each of the 10 fingers and also between the dominant and non-dominant hands. Based on these observations, several recommendations are made regarding: 1) the selection of fingers to be captured depending on the context of the application; 2) improvement in the usability of scanners and the capturing protocols; 3) improvement in the development, ergonomics and positioning of the acquisition devices; and 4) improvement of recognition algorithms by incorporating information on finger type and handedness.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"155 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123296506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detecting Deepfakes in Alternative Color Spaces to Withstand Unseen Corruptions 检测深度伪造在替代颜色空间,以抵御看不见的腐败
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157416
Kai Zeng, Xiangyu Yu, Beibei Liu, Yu Guan, Yongjian Hu
The adverse impact of deepfakes has recently raised world-wide concerns. Many ways of deepfake detection are published in the literature. The reported results of existing methods are generally good under known settings. However, the robustness challenge in deepfake detection is not well addressed. Most detectors fail to identify deepfakes that have undergone post-processing. Observing that the conventionally adopted RGB space does not guarantee the best performance, we propose other color spaces that prove to be more effective in detecting corrupted deepfake videos. We design a robust detection approach that leverages an adaptive manipulation trace extraction network to reveal artifacts from two color spaces. To mimic practical scenarios, we conduct experiments to detect images with post-processings that are not seen in the training stage. The results demonstrate that our approach outperforms state-of-the-art methods, boosting the average detection accuracy by 7% ~ 17%.
深度造假的负面影响最近引起了全世界的关注。文献中发表了许多深度伪造检测方法。现有方法报告的结果在已知条件下通常是好的。然而,深度伪造检测中的鲁棒性挑战并没有得到很好的解决。大多数检测器无法识别经过后处理的深度伪造。观察到传统采用的RGB空间并不能保证最佳性能,我们提出了其他被证明在检测损坏的深度假视频方面更有效的颜色空间。我们设计了一种鲁棒的检测方法,利用自适应操作跟踪提取网络来揭示来自两个颜色空间的工件。为了模拟实际场景,我们进行了实验来检测经过后处理的图像,这些图像在训练阶段是看不到的。结果表明,我们的方法优于目前最先进的方法,平均检测精度提高了7% ~ 17%。
{"title":"Detecting Deepfakes in Alternative Color Spaces to Withstand Unseen Corruptions","authors":"Kai Zeng, Xiangyu Yu, Beibei Liu, Yu Guan, Yongjian Hu","doi":"10.1109/IWBF57495.2023.10157416","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157416","url":null,"abstract":"The adverse impact of deepfakes has recently raised world-wide concerns. Many ways of deepfake detection are published in the literature. The reported results of existing methods are generally good under known settings. However, the robustness challenge in deepfake detection is not well addressed. Most detectors fail to identify deepfakes that have undergone post-processing. Observing that the conventionally adopted RGB space does not guarantee the best performance, we propose other color spaces that prove to be more effective in detecting corrupted deepfake videos. We design a robust detection approach that leverages an adaptive manipulation trace extraction network to reveal artifacts from two color spaces. To mimic practical scenarios, we conduct experiments to detect images with post-processings that are not seen in the training stage. The results demonstrate that our approach outperforms state-of-the-art methods, boosting the average detection accuracy by 7% ~ 17%.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129436642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Video superresolution in real forensic cases 真实法医案件中的视频超分辨率
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157760
J. Flusser, F. Šroubek, J. Kamenický, B. Zitová
Visual data, such as images and videos, are frequently used as evidence in court trials. If the data quality is insufficient to convince the court, a carefully tailored data processing algorithm supported with expert’s opinion is necessary. We present two real cases from our forensic expertise practice, in which we demonstrate a successful application of video superresolution that helped to convict offenders. The most important feature of image processing algorithms to be legally accepted by the court, is to rule out artifacts with realistic details, which are known to appear for example in deep learning methods.
视觉数据,如图像和视频,经常被用作法庭审判的证据。如果数据质量不足以说服法院,则需要精心定制的数据处理算法,并得到专家意见的支持。我们展示了两个来自我们法医专业实践的真实案例,在这些案例中,我们展示了视频超分辨率的成功应用,帮助罪犯定罪。被法院合法接受的图像处理算法的最重要的特征是,排除在深度学习方法中出现的具有现实细节的人工制品。
{"title":"Video superresolution in real forensic cases","authors":"J. Flusser, F. Šroubek, J. Kamenický, B. Zitová","doi":"10.1109/IWBF57495.2023.10157760","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157760","url":null,"abstract":"Visual data, such as images and videos, are frequently used as evidence in court trials. If the data quality is insufficient to convince the court, a carefully tailored data processing algorithm supported with expert’s opinion is necessary. We present two real cases from our forensic expertise practice, in which we demonstrate a successful application of video superresolution that helped to convict offenders. The most important feature of image processing algorithms to be legally accepted by the court, is to rule out artifacts with realistic details, which are known to appear for example in deep learning methods.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"164 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116293600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Study on Diffusion Modelling For Sensor-based Human Activity Recognition 基于传感器的人体活动识别扩散模型研究
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157482
Shuai Shao, Victor Sanchez
Human activity recognition (HAR) is a core research topic in mobile and wearable computing, and has been applied in many applications including biometrics, health monitoring and sports coaching. In recent years, researchers have focused more attention on sensor-based HAR due to the popularity of sensor devices. However, sensor-based HAR faces the challenge of limited data size caused by the high cost of data collection and labelling work, resulting in low performance for HAR tasks. Data transformation and generative adversarial network (GAN) have been proposed as data augmentation approaches to enrich sensor data, thereby addressing the problem of data size limitations. In this paper, we studied the effectiveness of diffusion-based generative models for generating synthetic sensor data as compared to the other data augmentation approaches in sensor-based HAR. In addition, UNet has been redesigned in order to improve the efficiency and practicality of diffusion modelling. Experiments on two public datasets showed the performance of diffusion modelling compared with different data augmentation methods, indicating the feasibility of synthetic sensor data generated using diffusion modelling.
人体活动识别(Human activity recognition, HAR)是移动和可穿戴计算领域的核心研究课题,在生物识别、健康监测、运动指导等领域都有广泛的应用。近年来,由于传感器设备的普及,研究人员越来越关注基于传感器的HAR。然而,基于传感器的HAR面临着数据大小有限的挑战,这是由于数据收集和标记工作的高成本造成的,从而导致HAR任务的性能较低。数据转换和生成对抗网络(GAN)已被提出作为数据增强方法来丰富传感器数据,从而解决数据大小限制的问题。在本文中,我们研究了基于扩散的生成模型在生成合成传感器数据方面的有效性,并与基于传感器的HAR中的其他数据增强方法进行了比较。此外,为了提高扩散建模的效率和实用性,对UNet进行了重新设计。在两个公共数据集上的实验表明,扩散建模与不同数据增强方法的性能进行了比较,表明利用扩散建模生成合成传感器数据的可行性。
{"title":"A Study on Diffusion Modelling For Sensor-based Human Activity Recognition","authors":"Shuai Shao, Victor Sanchez","doi":"10.1109/IWBF57495.2023.10157482","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157482","url":null,"abstract":"Human activity recognition (HAR) is a core research topic in mobile and wearable computing, and has been applied in many applications including biometrics, health monitoring and sports coaching. In recent years, researchers have focused more attention on sensor-based HAR due to the popularity of sensor devices. However, sensor-based HAR faces the challenge of limited data size caused by the high cost of data collection and labelling work, resulting in low performance for HAR tasks. Data transformation and generative adversarial network (GAN) have been proposed as data augmentation approaches to enrich sensor data, thereby addressing the problem of data size limitations. In this paper, we studied the effectiveness of diffusion-based generative models for generating synthetic sensor data as compared to the other data augmentation approaches in sensor-based HAR. In addition, UNet has been redesigned in order to improve the efficiency and practicality of diffusion modelling. Experiments on two public datasets showed the performance of diffusion modelling compared with different data augmentation methods, indicating the feasibility of synthetic sensor data generated using diffusion modelling.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114380231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Multi-Tasked Approach Towards Face Detection and Face Quality Assessment 一种多任务人脸检测和人脸质量评估方法
Pub Date : 2023-04-19 DOI: 10.1109/IWBF57495.2023.10157540
Rod Izadi, Chen Liu
Face in Video Recognition (FiVR) commonly follows a sequential pipeline of face detection, face quality assessment, and face recognition. However, performing these often machine learning-based tasks sequentially in real-time is a challenge when considering the excessive overhead caused by convolution and other feature extraction operations typically seen in neural networks employed across these stages. To overcome this challenge, a process that can perform these operations in parallel is needed. In this paper, we propose a methodology that can alleviate the constraints of real-time processing found in the sequential pipeline of FiVR. We exploit the similarities in features used in face detection and face quality assessment, hence designing a multi-tasked face detection and quality assessment network which can perform our FiVR operations with less inference time without sparing prediction accuracy. We evaluated the face quality prediction performance of our proposed approach in comparison with a stand-alone face quality network. We also evaluated the reduction in inference time by comparing the prediction speed of our multi-tasked face detection and quality network against its sequential counterparts. Our experimental results show that our multi-tasked model can successfully meet real-time processing demand while performing at the same level of accuracy as the sequential stand-alone models.
视频人脸识别(FiVR)通常遵循人脸检测、人脸质量评估和人脸识别的顺序流程。然而,考虑到卷积和其他特征提取操作通常在这些阶段使用的神经网络中造成的过度开销,实时执行这些通常基于机器学习的任务是一个挑战。为了克服这一挑战,需要一个能够并行执行这些操作的进程。在本文中,我们提出了一种方法,可以减轻实时处理的限制,发现在顺序管道的FiVR。我们利用人脸检测和人脸质量评估中使用的特征的相似性,从而设计了一个多任务的人脸检测和质量评估网络,该网络可以在不影响预测精度的情况下以更少的推理时间执行我们的FiVR操作。我们与独立的人脸质量网络比较,评估了我们提出的方法的人脸质量预测性能。我们还通过比较多任务人脸检测和质量网络的预测速度与顺序网络的预测速度来评估推理时间的减少。实验结果表明,我们的多任务模型可以成功地满足实时处理需求,同时具有与顺序独立模型相同的精度。
{"title":"A Multi-Tasked Approach Towards Face Detection and Face Quality Assessment","authors":"Rod Izadi, Chen Liu","doi":"10.1109/IWBF57495.2023.10157540","DOIUrl":"https://doi.org/10.1109/IWBF57495.2023.10157540","url":null,"abstract":"Face in Video Recognition (FiVR) commonly follows a sequential pipeline of face detection, face quality assessment, and face recognition. However, performing these often machine learning-based tasks sequentially in real-time is a challenge when considering the excessive overhead caused by convolution and other feature extraction operations typically seen in neural networks employed across these stages. To overcome this challenge, a process that can perform these operations in parallel is needed. In this paper, we propose a methodology that can alleviate the constraints of real-time processing found in the sequential pipeline of FiVR. We exploit the similarities in features used in face detection and face quality assessment, hence designing a multi-tasked face detection and quality assessment network which can perform our FiVR operations with less inference time without sparing prediction accuracy. We evaluated the face quality prediction performance of our proposed approach in comparison with a stand-alone face quality network. We also evaluated the reduction in inference time by comparing the prediction speed of our multi-tasked face detection and quality network against its sequential counterparts. Our experimental results show that our multi-tasked model can successfully meet real-time processing demand while performing at the same level of accuracy as the sequential stand-alone models.","PeriodicalId":273412,"journal":{"name":"2023 11th International Workshop on Biometrics and Forensics (IWBF)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129647894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2023 11th International Workshop on Biometrics and Forensics (IWBF)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1