首页 > 最新文献

IET Biometrics最新文献

英文 中文
Unified Physical–Digital Face Attack Detection Challenge: A Review 统一物理-数字面部攻击检测挑战:综述
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-02-04 DOI: 10.1049/bme2/9653627
Junze Zheng, Xinping Gao, Ajian Liu, Haocheng Yuan, Jun Wan, Yanyan Liang, Jiankang Deng, Sergio Escalera, Hugo Jair Escalante, Zhen Lei, Isabelle Guyon, Du Zhang

Face antispoofing (FAS) technologies play a pivotal role in safeguarding face recognition (FR) systems against potential security loopholes. The biometric community has witnessed significant advancements lately, largely due to the exceptional performance of deep learning architectures and the abundance of substantial datasets. Despite these progress, FR systems remain susceptible to both physical and digital forgery attacks. However, most existing detection methods do not address both types of threats concurrently. To bridge this gap and foster the development of a comprehensive detection framework, we have compiled a unified dataset named UniAttackData. This dataset incorporates both physical and digital spoofing attacks while maintaining identity consistency, encompassing 1800 participants each subjected to two different physical attacks (PAs) and 12 different digital attacks (DAs), respectively. This effort has resulted in a comprehensive collection of 29,706 video samples. We organized the Chalearn FAS face attack detection challenge based on this novel resource to boost research aiming to promote joint antispoofing efforts. The Chalearn unified antispoofing attack detection challenge drew 136 teams during the development phase, with 13 teams advancing to the final round. The organizing team revalidated and re-executed the submitted code to determine the final rankings. This paper provides a summary of the challenge, covering the dataset used, the protocol definition, the evaluation metrics, and the competition results. Additionally, we discuss the top-ranked algorithms and the research insights offered by the competition for attack detection.

人脸反欺骗(FAS)技术在保护人脸识别(FR)系统不受潜在安全漏洞的影响方面发挥着关键作用。生物识别社区最近取得了重大进展,这主要归功于深度学习架构的卓越性能和大量数据集的丰富。尽管取得了这些进展,FR系统仍然容易受到物理和数字伪造攻击。然而,大多数现有的检测方法不能同时处理这两种类型的威胁。为了弥合这一差距并促进全面检测框架的发展,我们编制了一个名为UniAttackData的统一数据集。该数据集结合了物理和数字欺骗攻击,同时保持身份一致性,包括1800名参与者,每个参与者分别遭受两种不同的物理攻击(pa)和12种不同的数字攻击(da)。这一努力的结果是全面收集了29,706个视频样本。我们组织了基于这种新资源的Chalearn FAS人脸攻击检测挑战,以促进旨在促进联合反欺骗努力的研究。Chalearn统一反欺骗攻击检测挑战在开发阶段吸引了136个团队,其中13个团队进入了最后一轮。组织团队重新验证并重新执行提交的代码以确定最终排名。本文提供了挑战的总结,包括使用的数据集,协议定义,评估指标和竞争结果。此外,我们还讨论了排名靠前的算法和攻击检测竞争提供的研究见解。
{"title":"Unified Physical–Digital Face Attack Detection Challenge: A Review","authors":"Junze Zheng,&nbsp;Xinping Gao,&nbsp;Ajian Liu,&nbsp;Haocheng Yuan,&nbsp;Jun Wan,&nbsp;Yanyan Liang,&nbsp;Jiankang Deng,&nbsp;Sergio Escalera,&nbsp;Hugo Jair Escalante,&nbsp;Zhen Lei,&nbsp;Isabelle Guyon,&nbsp;Du Zhang","doi":"10.1049/bme2/9653627","DOIUrl":"https://doi.org/10.1049/bme2/9653627","url":null,"abstract":"<p>Face antispoofing (FAS) technologies play a pivotal role in safeguarding face recognition (FR) systems against potential security loopholes. The biometric community has witnessed significant advancements lately, largely due to the exceptional performance of deep learning architectures and the abundance of substantial datasets. Despite these progress, FR systems remain susceptible to both physical and digital forgery attacks. However, most existing detection methods do not address both types of threats concurrently. To bridge this gap and foster the development of a comprehensive detection framework, we have compiled a unified dataset named UniAttackData. This dataset incorporates both physical and digital spoofing attacks while maintaining identity consistency, encompassing 1800 participants each subjected to two different physical attacks (PAs) and 12 different digital attacks (DAs), respectively. This effort has resulted in a comprehensive collection of 29,706 video samples. We organized the Chalearn FAS face attack detection challenge based on this novel resource to boost research aiming to promote joint antispoofing efforts. The Chalearn unified antispoofing attack detection challenge drew 136 teams during the development phase, with 13 teams advancing to the final round. The organizing team revalidated and re-executed the submitted code to determine the final rankings. This paper provides a summary of the challenge, covering the dataset used, the protocol definition, the evaluation metrics, and the competition results. Additionally, we discuss the top-ranked algorithms and the research insights offered by the competition for attack detection.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2026 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2026-02-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/9653627","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146139148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robustness Analysis of Distributed CNN Model Training in Expression Recognition 分布式CNN模型训练在表情识别中的鲁棒性分析
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-31 DOI: 10.1049/bme2/4107824
Jun Li

Facial expression recognition is vital in pattern recognition and affective computing. With the advancement of deep learning, its performance has improved, yet challenges remain in nonlaboratory environments due to occlusion, poor lighting, and varying head poses. This study explores a robust facial expression recognition approach using a CNN-based model integrated with key point localization techniques. Instead of relying on a dense set of landmarks, the proposed method focuses on fewer but more informative expression key points. Each point is analyzed for local shape features, and contour consistency is verified using indexing along the normal direction. This strategy enhances robustness while reducing computational complexity. Specifically, the hybrid active shape model (ASM) + structure method significantly lowers the processing load compared to the traditional ASM approach. Experimental results demonstrate a 3.02% improvement in recognition accuracy over one-to-many SVM classifiers when dealing with clear facial images. Furthermore, the system shows strong resilience to partial occlusions and maintains real-time performance, making it suitable for real-world applications. The proposed framework highlights the importance of selecting effective key points and optimizing feature extraction to enhance both accuracy and efficiency in facial expression recognition tasks under challenging conditions.

面部表情识别是模式识别和情感计算的重要组成部分。随着深度学习的进步,其性能得到了提高,但在非实验室环境中,由于遮挡、光线不足和头部姿势的变化,仍然存在挑战。本研究探索了一种基于cnn的面部表情识别方法,该方法结合了关键点定位技术。该方法不依赖于密集的地标集,而是专注于更少但更有信息量的表达关键点。分析每个点的局部形状特征,并沿法线方向标度验证轮廓一致性。该策略增强了鲁棒性,同时降低了计算复杂度。具体而言,与传统的主动形状模型(ASM)方法相比,混合主动形状模型(ASM) +结构方法显著降低了处理负荷。实验结果表明,在处理清晰的人脸图像时,与一对多SVM分类器相比,识别精度提高了3.02%。此外,该系统显示出对部分遮挡的强弹性,并保持实时性能,使其适合实际应用。该框架强调了选择有效的关键点和优化特征提取的重要性,以提高具有挑战性条件下面部表情识别任务的准确性和效率。
{"title":"Robustness Analysis of Distributed CNN Model Training in Expression Recognition","authors":"Jun Li","doi":"10.1049/bme2/4107824","DOIUrl":"https://doi.org/10.1049/bme2/4107824","url":null,"abstract":"<p>Facial expression recognition is vital in pattern recognition and affective computing. With the advancement of deep learning, its performance has improved, yet challenges remain in nonlaboratory environments due to occlusion, poor lighting, and varying head poses. This study explores a robust facial expression recognition approach using a CNN-based model integrated with key point localization techniques. Instead of relying on a dense set of landmarks, the proposed method focuses on fewer but more informative expression key points. Each point is analyzed for local shape features, and contour consistency is verified using indexing along the normal direction. This strategy enhances robustness while reducing computational complexity. Specifically, the hybrid active shape model (ASM) + structure method significantly lowers the processing load compared to the traditional ASM approach. Experimental results demonstrate a 3.02% improvement in recognition accuracy over one-to-many SVM classifiers when dealing with clear facial images. Furthermore, the system shows strong resilience to partial occlusions and maintains real-time performance, making it suitable for real-world applications. The proposed framework highlights the importance of selecting effective key points and optimizing feature extraction to enhance both accuracy and efficiency in facial expression recognition tasks under challenging conditions.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2026 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2026-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/4107824","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146148306","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Introducing Learnable Gaussian Noise Into Defed for Enhanced Defense Against Adversarial Attacks in Fingerprint Liveness Detection 在指纹活体检测中引入可学习高斯噪声增强对抗性攻击的防御
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-17 DOI: 10.1049/bme2/5664546
Shuifa Sun, Shaohua Hu, Yifei Wang, Wanyi Zheng, Jianming Lin, Sani M. Abdullahi, Jian Zhang

Deep learning has significantly improved the performance of fingerprint liveness detection, while susceptibility to adversarial attacks remains a critical security challenge. Existing input transformation–based defense methods, including JPEG compression, total variance minimization (TVM), high-level representation guided denoiser (HGD), and Defed, are typically designed for specific attacks, resulting in limited generalization across diverse adversarial scenarios. Experimental analysis indicates that among the four defense methods based on input transformation, Defed achieves the best overall performance when evaluated against both momentum iterative fast gradient sign method (MI-FGSM) and DeepFool attacks. However, Defed exhibits strong robustness against MI-FGSM attacks but demonstrates insufficient defense effectiveness against DeepFool attacks. To address this issue, an improved method of Defed has been proposed by integrating a learnable Gaussian noise module into the core structure to enable adaptive suppression of adversarial perturbations, and by employing 1 × 1 convolutions to allow cross-channel information interaction, thereby enhancing feature consistency and overall robustness. Experimental results on the LivDet 2015 dataset demonstrate that the defense success rate against DeepFool attacks has increased by 3%–5%, while strong robustness against MI-FGSM attacks has been maintained, substantially improving the security and reliability of fingerprint liveness detection systems.

深度学习显著提高了指纹活性检测的性能,但对对抗性攻击的敏感性仍然是一个关键的安全挑战。现有的基于输入变换的防御方法,包括JPEG压缩、总方差最小化(TVM)、高级表示引导去噪(HGD)和Defed,通常是为特定的攻击而设计的,导致在不同的对抗场景中泛化有限。实验分析表明,在四种基于输入变换的防御方法中,Defed在对抗动量迭代快速梯度符号法(MI-FGSM)和DeepFool攻击时均取得了最佳的综合性能。然而,Defed对MI-FGSM攻击表现出很强的鲁棒性,但对DeepFool攻击的防御效果不够。为了解决这一问题,提出了一种改进的Defed方法,将一个可学习的高斯噪声模块集成到核心结构中,以实现对抗性扰动的自适应抑制,并采用1 × 1卷积允许跨通道信息交互,从而增强特征一致性和整体鲁棒性。在LivDet 2015数据集上的实验结果表明,对DeepFool攻击的防御成功率提高了3%-5%,同时对MI-FGSM攻击保持了较强的鲁棒性,大大提高了指纹活体检测系统的安全性和可靠性。
{"title":"Introducing Learnable Gaussian Noise Into Defed for Enhanced Defense Against Adversarial Attacks in Fingerprint Liveness Detection","authors":"Shuifa Sun,&nbsp;Shaohua Hu,&nbsp;Yifei Wang,&nbsp;Wanyi Zheng,&nbsp;Jianming Lin,&nbsp;Sani M. Abdullahi,&nbsp;Jian Zhang","doi":"10.1049/bme2/5664546","DOIUrl":"https://doi.org/10.1049/bme2/5664546","url":null,"abstract":"<p>Deep learning has significantly improved the performance of fingerprint liveness detection, while susceptibility to adversarial attacks remains a critical security challenge. Existing input transformation–based defense methods, including JPEG compression, total variance minimization (TVM), high-level representation guided denoiser (HGD), and Defed, are typically designed for specific attacks, resulting in limited generalization across diverse adversarial scenarios. Experimental analysis indicates that among the four defense methods based on input transformation, Defed achieves the best overall performance when evaluated against both momentum iterative fast gradient sign method (MI-FGSM) and DeepFool attacks. However, Defed exhibits strong robustness against MI-FGSM attacks but demonstrates insufficient defense effectiveness against DeepFool attacks. To address this issue, an improved method of Defed has been proposed by integrating a learnable Gaussian noise module into the core structure to enable adaptive suppression of adversarial perturbations, and by employing 1 × 1 convolutions to allow cross-channel information interaction, thereby enhancing feature consistency and overall robustness. Experimental results on the LivDet 2015 dataset demonstrate that the defense success rate against DeepFool attacks has increased by 3%–5%, while strong robustness against MI-FGSM attacks has been maintained, substantially improving the security and reliability of fingerprint liveness detection systems.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2026 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2026-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/5664546","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146002409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Conditional Synthetic Live and Spoof Fingerprint Generation 条件合成活和欺骗指纹生成
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-10 DOI: 10.1049/bme2/7736489
Syed Konain Abbas, Sandip Purnapatra, M. G. Sarwar Murshed, Conor Miller-Lynch, Lambert Igene, Soumyabrata Dey, Stephanie Schuckers, Faraz Hussain

Large fingerprint datasets, while important for training and evaluation, are time-consuming and expensive to collect and require strict privacy measures. Researchers are exploring the use of synthetic fingerprint data to address these issues. This article presents a novel approach for generating synthetic fingerprint images (both spoof and live), addressing concerns related to privacy, cost, and accessibility in biometric data collection. Our approach utilizes conditional StyleGAN2-ADA and StyleGAN3 architectures to produce high-resolution synthetic live fingerprints, conditioned on specific finger identities (thumb through little finger). Additionally, we employ CycleGANs to translate these into realistic spoof fingerprints, simulating a variety of presentation attack materials (e.g., EcoFlex, Play-Doh). These synthetic spoof fingerprints are crucial for developing robust spoof detection systems. Through these generative models, we created two synthetic datasets (DB2 and DB3), each containing 1500 fingerprint images of all 10 fingers with multiple impressions per finger, and including corresponding spoofs in eight material types. The results indicate robust performance: our StyleGAN3 model achieves a Fréchet inception distance (FID) as low as 5, and the generated fingerprints achieve a true acceptance rate (TAR) of 99.47% at a 0.01% false acceptance rate (FAR). The StyleGAN2-ADA model achieved a TAR of 98.67% at the same 0.01% FAR. We assess fingerprint quality using standard metrics (NFIQ2, MINDTCT), and notably, matching experiments confirm strong privacy preservation, with no significant evidence of identity leakage, confirming the strong privacy-preserving properties of our synthetic datasets.

大型指纹数据集虽然对训练和评估很重要,但收集起来既耗时又昂贵,还需要严格的隐私保护措施。研究人员正在探索使用合成指纹数据来解决这些问题。本文提出了一种生成合成指纹图像(包括假指纹和真实指纹)的新方法,解决了生物识别数据收集中与隐私、成本和可访问性相关的问题。我们的方法利用条件StyleGAN2-ADA和StyleGAN3架构来生成高分辨率的合成活指纹,条件是特定的手指身份(拇指到小指)。此外,我们使用cyclegan将这些转化为逼真的欺骗指纹,模拟各种演示攻击材料(例如,EcoFlex, Play-Doh)。这些合成欺骗指纹对于开发强大的欺骗检测系统至关重要。通过这些生成模型,我们创建了两个合成数据集(DB2和DB3),每个数据集包含所有10个手指的1500个指纹图像,每个手指有多个印痕,并包括8种材料类型的相应欺骗。结果显示了稳健的性能:我们的StyleGAN3模型实现了低至5的fr起始距离(FID),生成的指纹在0.01%的错误接受率(FAR)下实现了99.47%的真实接受率(TAR)。StyleGAN2-ADA模型在相同的0.01% FAR下实现了98.67%的TAR。我们使用标准指标(NFIQ2, MINDTCT)评估指纹质量,值得注意的是,匹配实验证实了强大的隐私保护,没有明显的身份泄露证据,证实了我们的合成数据集的强大隐私保护特性。
{"title":"Conditional Synthetic Live and Spoof Fingerprint Generation","authors":"Syed Konain Abbas,&nbsp;Sandip Purnapatra,&nbsp;M. G. Sarwar Murshed,&nbsp;Conor Miller-Lynch,&nbsp;Lambert Igene,&nbsp;Soumyabrata Dey,&nbsp;Stephanie Schuckers,&nbsp;Faraz Hussain","doi":"10.1049/bme2/7736489","DOIUrl":"https://doi.org/10.1049/bme2/7736489","url":null,"abstract":"<p>Large fingerprint datasets, while important for training and evaluation, are time-consuming and expensive to collect and require strict privacy measures. Researchers are exploring the use of synthetic fingerprint data to address these issues. This article presents a novel approach for generating synthetic fingerprint images (both spoof and live), addressing concerns related to privacy, cost, and accessibility in biometric data collection. Our approach utilizes conditional StyleGAN2-ADA and StyleGAN3 architectures to produce high-resolution synthetic live fingerprints, conditioned on specific finger identities (thumb through little finger). Additionally, we employ CycleGANs to translate these into realistic spoof fingerprints, simulating a variety of presentation attack materials (e.g., EcoFlex, Play-Doh). These synthetic spoof fingerprints are crucial for developing robust spoof detection systems. Through these generative models, we created two synthetic datasets (DB2 and DB3), each containing 1500 fingerprint images of all 10 fingers with multiple impressions per finger, and including corresponding spoofs in eight material types. The results indicate robust performance: our StyleGAN3 model achieves a Fréchet inception distance (FID) as low as 5, and the generated fingerprints achieve a true acceptance rate (TAR) of 99.47% at a 0.01% false acceptance rate (FAR). The StyleGAN2-ADA model achieved a TAR of 98.67% at the same 0.01% FAR. We assess fingerprint quality using standard metrics (NFIQ2, MINDTCT), and notably, matching experiments confirm strong privacy preservation, with no significant evidence of identity leakage, confirming the strong privacy-preserving properties of our synthetic datasets.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2026 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2026-01-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/7736489","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145964221","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reinforcement Training of Face Recognition Systems Using Morphing and XAI Methods 基于变形和XAI方法的人脸识别系统强化训练
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-01-06 DOI: 10.1049/bme2/7897011
Roberto Gallardo-Cava, David Ortega-DelCampo, Daniel Palacios-Alonso, Javier M. Moguerza, Cristina Conde, Enrique Cabello

This study introduces a reinforcement training framework for face recognition systems (FRSs) that leverages facial morphing techniques to generate counterfactual visual instances for model enhancement. Two complementary morphing strategies were employed: a geometric approach based on Delaunay–Voronoi triangulation (DVT-Morph) and a generative approach using latent diffusion and autoencoder-based models (diffusion-based morphing [MorDIFF]). The generated morphs act as controlled counterfactuals, representing minimally modified facial images that induce changes in FRS verification decisions. The proposed method integrates these counterfactuals into the training process of two state-of-the-art recognition systems, ArcFace and MagFace, to strengthen their decision boundaries and improve their robustness, calibration, and explainability. By combining morphing-based counterfactual generation with eXplainable Artificial Intelligence (XAI) techniques, the framework enables a more interpretable embedding space and increased resilience against morphing and adversarial perturbations. The experimental results demonstrate that the inclusion of morph-based counterfactuals significantly enhances the verification accuracy and decision transparency of modern FRSs. Moreover, the methodology is model- and morphing-agnostic and can be applied to any FRS architecture, regardless of the morphing generation technique.

本研究为人脸识别系统(FRSs)引入了一个强化训练框架,该框架利用面部变形技术为模型增强生成反事实视觉实例。采用了两种互补的变形策略:基于Delaunay-Voronoi三角剖分(DVT-Morph)的几何方法和基于潜在扩散和自编码器模型的生成方法(基于扩散的变形[MorDIFF])。生成的变形作为受控的反事实,表示引起FRS验证决策变化的最小修改的面部图像。提出的方法将这些反事实集成到两个最先进的识别系统ArcFace和MagFace的训练过程中,以增强其决策边界并提高其鲁棒性、校准性和可解释性。通过将基于变形的反事实生成与可解释的人工智能(XAI)技术相结合,该框架实现了更可解释的嵌入空间,并增强了对变形和对抗性扰动的弹性。实验结果表明,包含基于形态的反事实显著提高了现代frs的验证精度和决策透明度。此外,该方法与模型和变形无关,可以应用于任何FRS体系结构,而不考虑变形生成技术。
{"title":"Reinforcement Training of Face Recognition Systems Using Morphing and XAI Methods","authors":"Roberto Gallardo-Cava,&nbsp;David Ortega-DelCampo,&nbsp;Daniel Palacios-Alonso,&nbsp;Javier M. Moguerza,&nbsp;Cristina Conde,&nbsp;Enrique Cabello","doi":"10.1049/bme2/7897011","DOIUrl":"https://doi.org/10.1049/bme2/7897011","url":null,"abstract":"<p>This study introduces a reinforcement training framework for face recognition systems (FRSs) that leverages facial morphing techniques to generate counterfactual visual instances for model enhancement. Two complementary morphing strategies were employed: a geometric approach based on Delaunay–Voronoi triangulation (DVT-Morph) and a generative approach using latent diffusion and autoencoder-based models (diffusion-based morphing [MorDIFF]). The generated morphs act as controlled counterfactuals, representing minimally modified facial images that induce changes in FRS verification decisions. The proposed method integrates these counterfactuals into the training process of two state-of-the-art recognition systems, ArcFace and MagFace, to strengthen their decision boundaries and improve their robustness, calibration, and explainability. By combining morphing-based counterfactual generation with eXplainable Artificial Intelligence (XAI) techniques, the framework enables a more interpretable embedding space and increased resilience against morphing and adversarial perturbations. The experimental results demonstrate that the inclusion of morph-based counterfactuals significantly enhances the verification accuracy and decision transparency of modern FRSs. Moreover, the methodology is model- and morphing-agnostic and can be applied to any FRS architecture, regardless of the morphing generation technique.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2026 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2026-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/7897011","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145909230","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CTFN: Multistage CNN-Transformer Fusion Network for ECG Authentication 用于心电认证的多级cnn -变压器融合网络
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-31 DOI: 10.1049/bme2/8757767
Heng Jia, Zhidong Zhao, Yefei Zhang, Xianfei Zhang, Yanjun Deng, Yongguang Wang, Hao Wang, Pengfei Jiao

In the face of the mounting challenges posed by cybersecurity threats, there is an imperative for the development of robust identity authentication systems to safeguard sensitive user data. Conventional biometric authentication methods, such as fingerprinting and facial recognition, are vulnerable to spoofing attacks. In contrast, electrocardiogram (ECG) signals offer distinct advantages as dynamic, “liveness”-assured biomarkers, exhibiting individual specificity. This study proposes a novel fusion network model, the convolutional neural network (CNN)-transformer fusion network (CTFN), to achieve high-precision ECG-based identity authentication by synergizing local feature extraction and global signal correlation analysis. The proposed framework integrates a multistage enhanced CNN to capture fine-grained local patterns in ECG morphology and a transformer encoder to model long-range dependencies in heartbeat sequences. An adaptive weighting mechanism dynamically optimizes the contributions of both modules during feature fusion. The efficacy of CTFN was evaluated in three critical real-world scenarios: single/multi-heartbeat authentication, cross-temporal consistency, and emotional variability resistance. The evaluation was conducted on 283 subjects from four public ECG databases: CYBHi, PTB, ECG-ID, and MIT-BIH. The CYBHi dataset revealed that CTFN exhibited a state-of-the-art recognition accuracy of 98.46%, 80.95%, and 90.76%, respectively, signifying its remarkable performance. Notably, the model attained a 100% authentication accuracy rate using only six heartbeats. This represents a 25% decrease in input requirements when compared to prior works, while concurrently maintaining its robust performance against physiological variations induced by emotional states or temporal gaps. These results demonstrate that CTFN significantly advances the practicality of ECG biometrics by balancing high accuracy with minimal data acquisition demands, offering a scalable and spoof-resistant solution for secure authentication systems.

面对网络安全威胁带来的日益严峻的挑战,开发强大的身份认证系统来保护敏感的用户数据势在必行。传统的生物识别认证方法,如指纹和面部识别,容易受到欺骗攻击。相比之下,心电图(ECG)信号作为动态的、“活性”保证的生物标志物具有明显的优势,表现出个体特异性。本文提出了一种新颖的融合网络模型——卷积神经网络(CNN)-变压器融合网络(CTFN),通过局部特征提取和全局信号相关分析的协同作用,实现基于ecg的高精度身份认证。该框架集成了一个多级增强型CNN来捕获ECG形态学中的细粒度局部模式,以及一个变压器编码器来模拟心跳序列中的远程依赖关系。一种自适应加权机制在特征融合过程中动态优化两个模块的贡献。CTFN的有效性在三个关键的现实世界场景中进行了评估:单次/多次心跳认证、跨时间一致性和情绪变异性抵抗。对来自四个公共心电图数据库(CYBHi、PTB、ECG- id和MIT-BIH)的283名受试者进行评估。CYBHi数据集显示,CTFN的识别准确率分别为98.46%、80.95%和90.76%,显示了其卓越的性能。值得注意的是,该模型仅使用6次心跳就获得了100%的身份验证准确率。与之前的工作相比,这意味着输入需求减少了25%,同时保持其对情绪状态或时间间隔引起的生理变化的强劲表现。这些结果表明,CTFN通过平衡高精度和最小数据采集需求,显着提高了心电生物识别的实用性,为安全认证系统提供了可扩展和防欺骗的解决方案。
{"title":"CTFN: Multistage CNN-Transformer Fusion Network for ECG Authentication","authors":"Heng Jia,&nbsp;Zhidong Zhao,&nbsp;Yefei Zhang,&nbsp;Xianfei Zhang,&nbsp;Yanjun Deng,&nbsp;Yongguang Wang,&nbsp;Hao Wang,&nbsp;Pengfei Jiao","doi":"10.1049/bme2/8757767","DOIUrl":"https://doi.org/10.1049/bme2/8757767","url":null,"abstract":"<p>In the face of the mounting challenges posed by cybersecurity threats, there is an imperative for the development of robust identity authentication systems to safeguard sensitive user data. Conventional biometric authentication methods, such as fingerprinting and facial recognition, are vulnerable to spoofing attacks. In contrast, electrocardiogram (ECG) signals offer distinct advantages as dynamic, “liveness”-assured biomarkers, exhibiting individual specificity. This study proposes a novel fusion network model, the convolutional neural network (CNN)-transformer fusion network (CTFN), to achieve high-precision ECG-based identity authentication by synergizing local feature extraction and global signal correlation analysis. The proposed framework integrates a multistage enhanced CNN to capture fine-grained local patterns in ECG morphology and a transformer encoder to model long-range dependencies in heartbeat sequences. An adaptive weighting mechanism dynamically optimizes the contributions of both modules during feature fusion. The efficacy of CTFN was evaluated in three critical real-world scenarios: single/multi-heartbeat authentication, cross-temporal consistency, and emotional variability resistance. The evaluation was conducted on 283 subjects from four public ECG databases: CYBHi, PTB, ECG-ID, and MIT-BIH. The CYBHi dataset revealed that CTFN exhibited a state-of-the-art recognition accuracy of 98.46%, 80.95%, and 90.76%, respectively, signifying its remarkable performance. Notably, the model attained a 100% authentication accuracy rate using only six heartbeats. This represents a 25% decrease in input requirements when compared to prior works, while concurrently maintaining its robust performance against physiological variations induced by emotional states or temporal gaps. These results demonstrate that CTFN significantly advances the practicality of ECG biometrics by balancing high accuracy with minimal data acquisition demands, offering a scalable and spoof-resistant solution for secure authentication systems.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2025 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/8757767","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145905238","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Spatiotextural Feature Learning-Driven Multimetric Biometric Authentication System for Strategic Smart Infrastructures: An Iris–Fingerprint Multimodality Solution 基于深度空间纹理特征学习的战略智能基础设施多尺度生物特征认证系统:虹膜-指纹多模态解决方案
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-19 DOI: 10.1049/bme2/9919250
Chethana J., Ravi J.

The last few years have witnessed an exponential rise in smart infrastructures including internet of things (IoT)-driven human–machine interface (HMI), automatic tailoring machine (ATM), smart home, data access control interfaces, and varied other e-infrastructures that demand robust person authentication systems. However, guaranteeing scalability with minimum latency and power consumption in the aforesaid applications remains a challenge. Unlike cryptographic methods, the biometric authentication methods seem to be more efficient, especially in terms of their ability toward non-repudiation, low-computational cost, low complexity, and minimum latency. The world has witnessed cyber breaches due to mimicking tools or techniques. In addition, standalone biometrics might be prone to false positives over a large application environment. It indicates the need of a multi-metric biometrics solution to guarantee robust and reliable personal authentication and verification task. To cope up the aforesaid demands, in this paper, a novel deep spatiotextural (DST) feature learning-driven multimodal biometric is proposed. Unlike traditional biometric solutions, we applied iris and fingerprint images altogether to achieve a robust person authentication solution. Here, the both biometrics input images (i.e., fingerprint and iris) were processed for preprocessing tasks such asintensity and histogram equalization, and Z-score normalization and resizing. Subsequently, firefly heuristic-driven fuzzy C-means (FCMs) clustering (FFCM) algorithm is developed to segment region-of-interest (ROI) from the input fingerprint and iris images. The segmented ROI-specific color images were processed for the DST feature extraction by using gray level co-occurrence metrics (GLCMs) and ResNet101 deep network. The extracted DST features were processed for feature-level fusion, and thus, the composite feature vector obtained was processed for multiclass classification by using random forest (RF) ensemble classifier. The simulation results confirmed (user) verification accuracy of 99.74%, 98.86%, recall 98.49%, and F-measure 98.67%, signifying its superiority over other state-of-the-arts. The feature learning robustness over the targeted multimetric biometrics confirms its suitability for real-world person authentication tasks.

过去几年见证了智能基础设施的指数级增长,包括物联网(IoT)驱动的人机界面(HMI),自动剪裁机(ATM),智能家居,数据访问控制接口以及需要强大的人员身份验证系统的各种其他电子基础设施。然而,在上述应用程序中以最小的延迟和功耗保证可伸缩性仍然是一个挑战。与加密方法不同,生物识别身份验证方法似乎更有效,特别是在不可否认性、低计算成本、低复杂性和最小延迟方面。由于模仿工具或技术,世界目睹了网络漏洞。此外,在大型应用程序环境中,独立的生物识别技术可能容易出现误报。这表明需要一种多度量生物识别解决方案来保证鲁棒性和可靠性的个人身份验证和验证任务。针对上述需求,本文提出了一种基于深度空间纹理特征学习的多模态生物识别方法。与传统的生物识别解决方案不同,我们将虹膜和指纹图像一起应用,以实现健壮的身份验证解决方案。在这里,两个生物识别输入图像(即指纹和虹膜)被处理用于预处理任务,如强度和直方图均衡化,以及z分数归一化和调整大小。随后,开发了萤火虫启发式驱动模糊c均值聚类算法(FFCM),从输入指纹和虹膜图像中分割感兴趣区域(ROI)。利用灰度共生度量(glcm)和ResNet101深度网络对分割后的roi特定彩色图像进行DST特征提取。对提取的DST特征进行特征级融合处理,得到的复合特征向量通过随机森林集成分类器进行多类分类。仿真结果表明,(用户)验证准确率为99.74%、98.86%,召回率为98.49%,F-measure为98.67%,优于其他先进技术。特征学习鲁棒性优于目标多度量生物特征,证实了其适用于现实世界的身份验证任务。
{"title":"Deep Spatiotextural Feature Learning-Driven Multimetric Biometric Authentication System for Strategic Smart Infrastructures: An Iris–Fingerprint Multimodality Solution","authors":"Chethana J.,&nbsp;Ravi J.","doi":"10.1049/bme2/9919250","DOIUrl":"10.1049/bme2/9919250","url":null,"abstract":"<p>The last few years have witnessed an exponential rise in smart infrastructures including internet of things (IoT)-driven human–machine interface (HMI), automatic tailoring machine (ATM), smart home, data access control interfaces, and varied other e-infrastructures that demand robust person authentication systems. However, guaranteeing scalability with minimum latency and power consumption in the aforesaid applications remains a challenge. Unlike cryptographic methods, the biometric authentication methods seem to be more efficient, especially in terms of their ability toward non-repudiation, low-computational cost, low complexity, and minimum latency. The world has witnessed cyber breaches due to mimicking tools or techniques. In addition, standalone biometrics might be prone to false positives over a large application environment. It indicates the need of a multi-metric biometrics solution to guarantee robust and reliable personal authentication and verification task. To cope up the aforesaid demands, in this paper, a novel deep spatiotextural (DST) feature learning-driven multimodal biometric is proposed. Unlike traditional biometric solutions, we applied iris and fingerprint images altogether to achieve a robust person authentication solution. Here, the both biometrics input images (i.e., fingerprint and iris) were processed for preprocessing tasks such asintensity and histogram equalization, and <i>Z</i>-score normalization and resizing. Subsequently, firefly heuristic-driven fuzzy <i>C</i>-means (FCMs) clustering (FFCM) algorithm is developed to segment region-of-interest (ROI) from the input fingerprint and iris images. The segmented ROI-specific color images were processed for the DST feature extraction by using gray level co-occurrence metrics (GLCMs) and ResNet101 deep network. The extracted DST features were processed for feature-level fusion, and thus, the composite feature vector obtained was processed for multiclass classification by using random forest (RF) ensemble classifier. The simulation results confirmed (user) verification accuracy of 99.74%, 98.86%, recall 98.49%, and <i>F</i>-measure 98.67%, signifying its superiority over other state-of-the-arts. The feature learning robustness over the targeted multimetric biometrics confirms its suitability for real-world person authentication tasks.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2025 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/9919250","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145824695","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Layer-Wise Cue Alignment for Unified Face Attack Detection With Vision-Language Model 基于视觉语言模型的统一人脸攻击检测分层线索对齐
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-17 DOI: 10.1049/bme2/3954107
Yongze Li, Jing Gu, Ning Li, Bo-Han Li, Xiaoyuan Yu, Zhiyao Liang, Bin Li, Zhen Lei

Unified face attack detection (UAD), which simultaneously addresses physical presentation attacks (PAs) and digital forgery attacks (DAs) with a vision-language model, remains challenging due to the difficulty of effectively separating live and fake cues. The challenges mainly arise from two aspects: (1) text prompts are insufficiently aligned with visual features across layers, and (2) patch tokens containing live and fake cues often overlap, leading to ambiguous attribution and small decision margins. To address these problems, we propose a novel Layer-wise Cue Alignment framework (LCA) that leverages textual features to extract both layer-wise and global cues from patch tokens, and we further introduce a new training strategy to improve the separation of live and fake cues. Specifically, the layer-wise prompts are obtained by the cue matching block (CMB), which matches textual features with patch embeddings at each transformer layer, and the layer-level cues are injected into the visual features of each layer and further aggregated by the cue fusion block (CFB) to form comprehensive prompts that enhance the overall visual representation. Moreover, we design a complementary supervision mechanism (CSM) that suppresses forgery cues in live faces while enforcing mutual exclusivity between live and fake cues in attack samples to improve the reliability of cue separation. Extensive experiments on multiple benchmarks demonstrate that our framework achieves state-of-the-art performance on most protocols of the datasets.

统一人脸攻击检测(UAD)通过视觉语言模型同时解决物理表示攻击(PAs)和数字伪造攻击(DAs),但由于难以有效分离真实和虚假线索,因此仍然具有挑战性。挑战主要来自两个方面:(1)文本提示与跨层的视觉特征不够一致;(2)包含真实和虚假线索的补丁令牌经常重叠,导致模糊的归属和较小的决策空间。为了解决这些问题,我们提出了一种新的分层线索对齐框架(LCA),该框架利用文本特征从补丁令牌中提取分层线索和全局线索,并进一步引入了一种新的训练策略来改进真实线索和虚假线索的分离。具体而言,通过线索匹配块(CMB)获得分层提示,CMB将文本特征与每个变压器层的补丁嵌入相匹配,并将层级提示注入每层的视觉特征中,并通过线索融合块(CFB)进一步聚合,形成综合提示,增强整体视觉表示。此外,我们设计了一种互补监督机制(CSM),该机制抑制了真实人脸中的伪造线索,同时在攻击样本中增强了真实和虚假线索之间的互斥性,以提高线索分离的可靠性。在多个基准测试上进行的大量实验表明,我们的框架在大多数数据集协议上实现了最先进的性能。
{"title":"Layer-Wise Cue Alignment for Unified Face Attack Detection With Vision-Language Model","authors":"Yongze Li,&nbsp;Jing Gu,&nbsp;Ning Li,&nbsp;Bo-Han Li,&nbsp;Xiaoyuan Yu,&nbsp;Zhiyao Liang,&nbsp;Bin Li,&nbsp;Zhen Lei","doi":"10.1049/bme2/3954107","DOIUrl":"https://doi.org/10.1049/bme2/3954107","url":null,"abstract":"<p>Unified face attack detection (UAD), which simultaneously addresses physical presentation attacks (PAs) and digital forgery attacks (DAs) with a vision-language model, remains challenging due to the difficulty of effectively separating live and fake cues. The challenges mainly arise from two aspects: (1) text prompts are insufficiently aligned with visual features across layers, and (2) patch tokens containing live and fake cues often overlap, leading to ambiguous attribution and small decision margins. To address these problems, we propose a novel Layer-wise Cue Alignment framework (LCA) that leverages textual features to extract both layer-wise and global cues from patch tokens, and we further introduce a new training strategy to improve the separation of live and fake cues. Specifically, the layer-wise prompts are obtained by the cue matching block (CMB), which matches textual features with patch embeddings at each transformer layer, and the layer-level cues are injected into the visual features of each layer and further aggregated by the cue fusion block (CFB) to form comprehensive prompts that enhance the overall visual representation. Moreover, we design a complementary supervision mechanism (CSM) that suppresses forgery cues in live faces while enforcing mutual exclusivity between live and fake cues in attack samples to improve the reliability of cue separation. Extensive experiments on multiple benchmarks demonstrate that our framework achieves state-of-the-art performance on most protocols of the datasets.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2025 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/3954107","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145845946","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Simple yet Effective ECG Identity Authentication With Low EER and Without Retraining 简单而有效的心电身份认证,低EER,无需再培训
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-12-11 DOI: 10.1049/bme2/7968221
Mingyu Dong, Zhidong Zhao, Yefei Zhang, Yanjun Deng, Hao Wang, Zhe Ye

Recently, electrocardiogram (ECG) signals have garnered significant attention in the field of identity authentication. For identity authentication, the ECG signals collected by wearing smart devices need to be determined whether the signal belongs to an enrolled one. Constrained by the computational efficiency of smart devices in practical scenarios, it is essential to reduce the complexity of the method to lower the computational load. To maintain the accuracy (ACC) of identity authentication, most research efforts rely on both R-wave extraction and segmentation for subsequent authentication. Moreover, many methods constantly require the model to be retrained during the user enrollment stage, leading to performance degradation and waste of training resources. Hence, we propose a simple yet effective ECG identity authentication method that applies blind segmentation and is free from retraining, which greatly simplifies the authentication process. To mitigate the equal error rate (EER) during the verification phase, a combination of AAM-softmax and triplet losses is employed, along with the incorporation of the hard negative mining within batch samples. Extensive experiments demonstrate that our method outperforms competitors by a large margin, e.g., achieving 0.40% EER on the large-scale autonomic dataset. Within models of comparable parameter sizes, our approach demonstrates markedly higher computational efficiency on both CPU and GPU platforms. The source code has been publicly released and is available at: https://github.com/DanMerry/LowEER.

近年来,心电图信号在身份认证领域受到了广泛的关注。对于身份认证,佩戴智能设备采集到的心电信号需要确定该信号是否属于注册的人。受实际场景中智能设备计算效率的限制,降低方法的复杂度以降低计算负荷至关重要。为了保持身份认证的准确性,大多数研究都依赖于r波提取和分割来进行后续认证。此外,许多方法在用户注册阶段不断要求对模型进行再训练,导致性能下降和训练资源的浪费。因此,我们提出了一种简单有效的心电身份认证方法,该方法采用盲分割,无需再训练,大大简化了认证过程。为了减轻验证阶段的等错误率(EER),采用了AAM-softmax和三重态损失的组合,并在批量样本中结合了硬负挖掘。大量的实验表明,我们的方法在很大程度上优于竞争对手,例如,在大规模自主数据集上实现了0.40%的EER。在可比较参数大小的模型中,我们的方法在CPU和GPU平台上都显示出明显更高的计算效率。源代码已经公开发布,可从https://github.com/DanMerry/LowEER获得。
{"title":"Simple yet Effective ECG Identity Authentication With Low EER and Without Retraining","authors":"Mingyu Dong,&nbsp;Zhidong Zhao,&nbsp;Yefei Zhang,&nbsp;Yanjun Deng,&nbsp;Hao Wang,&nbsp;Zhe Ye","doi":"10.1049/bme2/7968221","DOIUrl":"10.1049/bme2/7968221","url":null,"abstract":"<p>Recently, electrocardiogram (ECG) signals have garnered significant attention in the field of identity authentication. For identity authentication, the ECG signals collected by wearing smart devices need to be determined whether the signal belongs to an enrolled one. Constrained by the computational efficiency of smart devices in practical scenarios, it is essential to reduce the complexity of the method to lower the computational load. To maintain the accuracy (ACC) of identity authentication, most research efforts rely on both R-wave extraction and segmentation for subsequent authentication. Moreover, many methods constantly require the model to be retrained during the user enrollment stage, leading to performance degradation and waste of training resources. Hence, we propose a simple yet effective ECG identity authentication method that applies blind segmentation and is free from retraining, which greatly simplifies the authentication process. To mitigate the equal error rate (EER) during the verification phase, a combination of AAM-softmax and triplet losses is employed, along with the incorporation of the hard negative mining within batch samples. Extensive experiments demonstrate that our method outperforms competitors by a large margin, e.g., achieving 0.40% EER on the large-scale autonomic dataset. Within models of comparable parameter sizes, our approach demonstrates markedly higher computational efficiency on both CPU and GPU platforms. The source code has been publicly released and is available at: https://github.com/DanMerry/LowEER.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2025 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/7968221","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145739916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Biometrics: A Review of Handcrafted and AI–Based Fusion Approaches 多模态生物识别:手工和基于人工智能的融合方法综述
IF 1.8 4区 计算机科学 Q3 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2025-10-28 DOI: 10.1049/bme2/5055434
Hind Es-Sobbahi, Mohamed Radouane, Khalid Nafil

As security threats continue to evolve, multimodal biometric recognition systems (MBRSs) have emerged as robust solutions for reliable user authentication. To the best of our knowledge, this study presents the first systematic literature review (SLR) specifically focused on MBRS based on physiological traits, combining traditional image processing techniques (e.g., Gabor filters and edge detection) with artificial intelligence (AI) methods. These include machine learning (ML) approaches (e.g., Softmax classifier and linear discriminant analysis), deep learning (DL) models (e.g., convolutional neural networks [CNNs]), and metaheuristic optimization algorithms (e.g., firefly algorithm, gray wolf optimizer [GWO], and GwPeSOA). We analyze and compare the frequency and effectiveness of various fusion levels (sensor, feature, score, and decision) employed in the literature. Our review synthesizes findings from 29 peer-reviewed studies, highlights commonly used biometric traits and databases (e.g., CASIA and IITD), and categorizes the fusion techniques applied at each stage of the biometric pipeline, from preprocessing and feature extraction to decision-making. Results show that score-level fusion remains the most widely adopted approach. Multimodal systems combining multiple physiological traits (e.g., face, iris, and finger vein) demonstrate significant performance gains, with some studies reporting accuracies reaching 100%. Importantly, no prior review has provided such an integrative perspective combining handcrafted techniques with diverse AI–based approaches across multiple fusion levels. This comprehensive synthesis is intended to guide future research toward more practical, scalable, and accurate multimodal biometric systems.

随着安全威胁的不断发展,多模态生物识别系统(MBRSs)已经成为可靠用户身份验证的强大解决方案。据我们所知,本研究首次针对基于生理特征的MBRS进行了系统的文献综述(SLR),将传统的图像处理技术(如Gabor滤波器和边缘检测)与人工智能(AI)方法相结合。这些包括机器学习(ML)方法(例如,Softmax分类器和线性判别分析),深度学习(DL)模型(例如,卷积神经网络[cnn])和元启发式优化算法(例如,萤火虫算法,灰狼优化器[GWO]和GwPeSOA)。我们分析和比较了文献中使用的各种融合水平(传感器、特征、评分和决策)的频率和有效性。我们的综述综合了29项同行评议的研究结果,重点介绍了常用的生物特征特征和数据库(如CASIA和IITD),并对生物特征管道各个阶段应用的融合技术进行了分类,从预处理、特征提取到决策。结果表明,分数级融合仍然是最广泛采用的方法。结合多种生理特征(如面部、虹膜和手指静脉)的多模态系统显示出显著的性能提升,一些研究报告准确率达到100%。重要的是,没有先前的综述提供了这样的综合视角,将手工制作的技术与跨多个融合水平的各种基于人工智能的方法相结合。这一全面的综合旨在指导未来的研究走向更实用、可扩展和准确的多模态生物识别系统。
{"title":"Multimodal Biometrics: A Review of Handcrafted and AI–Based Fusion Approaches","authors":"Hind Es-Sobbahi,&nbsp;Mohamed Radouane,&nbsp;Khalid Nafil","doi":"10.1049/bme2/5055434","DOIUrl":"https://doi.org/10.1049/bme2/5055434","url":null,"abstract":"<p>As security threats continue to evolve, multimodal biometric recognition systems (MBRSs) have emerged as robust solutions for reliable user authentication. To the best of our knowledge, this study presents the first systematic literature review (SLR) specifically focused on MBRS based on physiological traits, combining traditional image processing techniques (e.g., Gabor filters and edge detection) with artificial intelligence (AI) methods. These include machine learning (ML) approaches (e.g., Softmax classifier and linear discriminant analysis), deep learning (DL) models (e.g., convolutional neural networks [CNNs]), and metaheuristic optimization algorithms (e.g., firefly algorithm, gray wolf optimizer [GWO], and GwPeSOA). We analyze and compare the frequency and effectiveness of various fusion levels (sensor, feature, score, and decision) employed in the literature. Our review synthesizes findings from 29 peer-reviewed studies, highlights commonly used biometric traits and databases (e.g., CASIA and IITD), and categorizes the fusion techniques applied at each stage of the biometric pipeline, from preprocessing and feature extraction to decision-making. Results show that score-level fusion remains the most widely adopted approach. Multimodal systems combining multiple physiological traits (e.g., face, iris, and finger vein) demonstrate significant performance gains, with some studies reporting accuracies reaching 100%. Importantly, no prior review has provided such an integrative perspective combining handcrafted techniques with diverse AI–based approaches across multiple fusion levels. This comprehensive synthesis is intended to guide future research toward more practical, scalable, and accurate multimodal biometric systems.</p>","PeriodicalId":48821,"journal":{"name":"IET Biometrics","volume":"2025 1","pages":""},"PeriodicalIF":1.8,"publicationDate":"2025-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ietresearch.onlinelibrary.wiley.com/doi/epdf/10.1049/bme2/5055434","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145406982","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IET Biometrics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1