首页 > 最新文献

Expert Systems with Applications最新文献

英文 中文
Chrysanthemum image quality assessment via multi-scale feature fusion and meta-learning 基于多尺度特征融合和元学习的菊花图像质量评估
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-02 DOI: 10.1016/j.eswa.2026.131378
Shun Zhu , Xichen Yang , Tianshu Wang , Zhongyuan Mao , Yifan Chen , Jian Jiang , Hui Yan
The origin tracing of chrysanthemum is significant in ensuring the quality of chrysanthemum. With the development of computer vision, it is feasible to utilize vision technology to the origin tracing. This enables intelligent origin tracing, thereby improving efficiency and accuracy. However, image distortions are inevitable while collecting chrysanthemum images. These distortions, such as incomplete chrysanthemum tissue and poor angle, tend to reduce the accuracy of the origin tracing. Thus, it is important to measure the image quality accurately, and then further improve the accuracy of the origin tracing. Considering it, we proposed a chrysanthemum image quality assessment method. First, a two-step screening (TSS) module is designed to screen existing classically distorted images that are suitable for distorted chrysanthemum images. Second, a deep feature extraction module is utilized to extract features at different receptive field levels. Third, the semantic analysis module is used to analyze and fuse the semantic information of different features. Finally, the meta-learning framework is designed to improve the accuracy and robustness of the model. The prior knowledge acquired through meta-learning is utilized to fine-tune the model with few-shot samples. The experimental results demonstrate that the proposed method can accurately judge incomplete and angle distortions, and thus effectively promote the accuracy of origin tracing. Our codes and models are available at https://github.com/dart-into/a-chrysanthemum-Screening-Method.
菊花产地溯源对保证菊花品质具有重要意义。随着计算机视觉技术的发展,将视觉技术应用于物体原点跟踪是可行的。这使得智能溯源成为可能,从而提高效率和准确性。然而,在收集菊花图像时,图像失真是不可避免的。这些畸变,如菊花组织不完整和角度差,往往会降低产地追踪的准确性。因此,准确测量图像质量,进而进一步提高原点跟踪的精度是非常重要的。为此,我们提出了一种菊花图像质量评价方法。首先,设计两步筛选(two-step screening, TSS)模块,筛选现有的适合于菊花畸变图像的经典畸变图像。其次,利用深度特征提取模块提取不同感受野层次的特征;第三,使用语义分析模块对不同特征的语义信息进行分析和融合。最后,设计了元学习框架来提高模型的准确性和鲁棒性。利用元学习获得的先验知识对模型进行少量采样微调。实验结果表明,该方法能够准确地判断不完全和角度畸变,从而有效地提高了原点跟踪的精度。我们的代码和模型可在https://github.com/dart-into/a-chrysanthemum-Screening-Method上获得。
{"title":"Chrysanthemum image quality assessment via multi-scale feature fusion and meta-learning","authors":"Shun Zhu ,&nbsp;Xichen Yang ,&nbsp;Tianshu Wang ,&nbsp;Zhongyuan Mao ,&nbsp;Yifan Chen ,&nbsp;Jian Jiang ,&nbsp;Hui Yan","doi":"10.1016/j.eswa.2026.131378","DOIUrl":"10.1016/j.eswa.2026.131378","url":null,"abstract":"<div><div>The origin tracing of chrysanthemum is significant in ensuring the quality of chrysanthemum. With the development of computer vision, it is feasible to utilize vision technology to the origin tracing. This enables intelligent origin tracing, thereby improving efficiency and accuracy. However, image distortions are inevitable while collecting chrysanthemum images. These distortions, such as incomplete chrysanthemum tissue and poor angle, tend to reduce the accuracy of the origin tracing. Thus, it is important to measure the image quality accurately, and then further improve the accuracy of the origin tracing. Considering it, we proposed a chrysanthemum image quality assessment method. First, a two-step screening (TSS) module is designed to screen existing classically distorted images that are suitable for distorted chrysanthemum images. Second, a deep feature extraction module is utilized to extract features at different receptive field levels. Third, the semantic analysis module is used to analyze and fuse the semantic information of different features. Finally, the meta-learning framework is designed to improve the accuracy and robustness of the model. The prior knowledge acquired through meta-learning is utilized to fine-tune the model with few-shot samples. The experimental results demonstrate that the proposed method can accurately judge incomplete and angle distortions, and thus effectively promote the accuracy of origin tracing. Our codes and models are available at <span><span>https://github.com/dart-into/a-chrysanthemum-Screening-Method</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131378"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146122752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NeuroVision: EEG-to-image reconstruction via progressive neural encoding and cross-modal distillation 神经视觉:通过渐进式神经编码和跨模态蒸馏的脑电图到图像的重建
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-05 DOI: 10.1016/j.eswa.2026.131526
Tianwei Qu , Zexue Yang , Qixian Zhang
Reconstructing visual imagery from electroencephalography (EEG) signals represents a fundamental challenge in brain-computer interfaces, as EEG recordings capture rich neural information about visual perception but lack the spatial resolution necessary for direct image reconstruction. Existing approaches typically employ diffusion models to synthesize images from EEG embeddings, yet these methods often fail to preserve the semantic content encoded in neural signals due to information loss during the encoding-decoding process. In this paper, we propose NeuroVision, a novel framework that adapts neural signal encoders for unified EEG understanding and visual reconstruction through progressive neural encoding and cross-modal distillation. Our approach addresses the fundamental trade-off between preserving EEG semantic information and achieving high-quality image reconstruction by introducing a three-stage progressive training scheme that gradually enhances reconstruction capability while maintaining original neural signal understanding. We develop a temporal-spatial attention fusion mechanism to capture multi-channel temporal dynamics in EEG signals, coupled with adaptive feature alignment that dynamically maps EEG representations to visual feature spaces. Furthermore, we introduce a semantic-preserving loss function that ensures reconstructed images faithfully reflect the semantic content of neural activity rather than generating visually plausible but semantically inconsistent outputs. Extensive experiments demonstrate that NeuroVision achieves superior reconstruction quality compared to existing diffusion-based approaches while significantly better preserving semantic correspondence between neural signals and visual content, establishing a new paradigm for EEG-to-image reconstruction that prioritizes semantic fidelity alongside visual quality.
从脑电图(EEG)信号中重建视觉图像是脑机接口的一个基本挑战,因为脑电图记录捕获了关于视觉感知的丰富神经信息,但缺乏直接图像重建所需的空间分辨率。现有方法通常采用扩散模型从脑电信号嵌入中合成图像,但由于编码-解码过程中的信息丢失,这些方法往往不能保留神经信号中编码的语义内容。在本文中,我们提出了一种新的神经视觉框架,该框架采用神经信号编码器,通过渐进式神经编码和跨模态蒸馏来实现统一的EEG理解和视觉重建。我们的方法解决了保留EEG语义信息和实现高质量图像重建之间的基本权衡,通过引入三阶段渐进式训练方案,在保持原始神经信号理解的同时逐步增强重建能力。我们开发了一种时空注意力融合机制来捕捉脑电图信号中的多通道时间动态,再加上自适应特征对齐,将脑电图表征动态映射到视觉特征空间。此外,我们引入了一个语义保持损失函数,以确保重建图像忠实地反映神经活动的语义内容,而不是产生视觉上似是而非语义上不一致的输出。大量实验表明,与现有的基于扩散的方法相比,NeuroVision实现了更好的重建质量,同时更好地保留了神经信号和视觉内容之间的语义对应关系,为脑电图到图像的重建建立了一个新的范例,优先考虑语义保真度和视觉质量。
{"title":"NeuroVision: EEG-to-image reconstruction via progressive neural encoding and cross-modal distillation","authors":"Tianwei Qu ,&nbsp;Zexue Yang ,&nbsp;Qixian Zhang","doi":"10.1016/j.eswa.2026.131526","DOIUrl":"10.1016/j.eswa.2026.131526","url":null,"abstract":"<div><div>Reconstructing visual imagery from electroencephalography (EEG) signals represents a fundamental challenge in brain-computer interfaces, as EEG recordings capture rich neural information about visual perception but lack the spatial resolution necessary for direct image reconstruction. Existing approaches typically employ diffusion models to synthesize images from EEG embeddings, yet these methods often fail to preserve the semantic content encoded in neural signals due to information loss during the encoding-decoding process. In this paper, we propose NeuroVision, a novel framework that adapts neural signal encoders for unified EEG understanding and visual reconstruction through progressive neural encoding and cross-modal distillation. Our approach addresses the fundamental trade-off between preserving EEG semantic information and achieving high-quality image reconstruction by introducing a three-stage progressive training scheme that gradually enhances reconstruction capability while maintaining original neural signal understanding. We develop a temporal-spatial attention fusion mechanism to capture multi-channel temporal dynamics in EEG signals, coupled with adaptive feature alignment that dynamically maps EEG representations to visual feature spaces. Furthermore, we introduce a semantic-preserving loss function that ensures reconstructed images faithfully reflect the semantic content of neural activity rather than generating visually plausible but semantically inconsistent outputs. Extensive experiments demonstrate that NeuroVision achieves superior reconstruction quality compared to existing diffusion-based approaches while significantly better preserving semantic correspondence between neural signals and visual content, establishing a new paradigm for EEG-to-image reconstruction that prioritizes semantic fidelity alongside visual quality.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131526"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PQS-BFL: A post-quantum secure blockchain-based federated learning framework PQS-BFL:一个后量子安全的基于区块链的联邦学习框架
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-04 DOI: 10.1016/j.eswa.2026.131449
Daniel Commey , Garth V. Crosby
Federated Learning (FL) enables collaborative model training while preserving data privacy, but its classical cryptographic underpinnings are vulnerable to quantum attacks. This vulnerability is particularly critical in sensitive domains like healthcare, where data requires long-term forward secrecy. This paper introduces PQS-BFL (Post-Quantum Secure Blockchain-based Federated Learning), a framework integrating post-quantum cryptography (PQC) with blockchain verification to secure FL against quantum adversaries. We employ ML-DSA-65 (standardized in FIPS 204, formerly Dilithium) signatures to authenticate model updates and leverage optimized smart contracts for decentralized validation. Designed for permissioned consortium environments (e.g., healthcare research networks), our framework ensures update integrity independent of the underlying ledger’s quantum resistance. Extensive evaluations on diverse datasets (MNIST, SVHN, HAR) demonstrate that PQS-BFL achieves feasible cryptographic operations (average PQC sign time: 0.65 ms, verify time: 0.53 ms) with a fixed signature size of 3309 Bytes. While blockchain integration incurs higher gas usage averaging 1.72 × 106 units per update due to PQC verification complexity, we demonstrate that this cost is negligible in private consortium settings where gas fees are nominal. Importantly, the cryptographic overhead relative to transaction time remains minimal (typically  < 0.2%), confirming that PQS-BFL is a viable architecture for securing critical infrastructure against future quantum threats.
联邦学习(FL)支持协作模型训练,同时保护数据隐私,但其经典加密基础容易受到量子攻击。此漏洞在医疗保健等敏感领域尤其严重,因为这些领域的数据需要长期向前保密。本文介绍了PQS-BFL(后量子安全基于区块链的联邦学习),这是一个将后量子密码学(PQC)与区块链验证集成在一起的框架,以保护FL免受量子对手的攻击。我们使用ML-DSA-65(在FIPS 204中标准化,以前是diiliium)签名来验证模型更新,并利用优化的智能合约进行分散验证。我们的框架专为许可的联盟环境(例如,医疗保健研究网络)而设计,可确保独立于底层分类账的量子阻力的更新完整性。在不同数据集(MNIST, SVHN, HAR)上的广泛评估表明,PQS-BFL在固定签名大小为3309字节的情况下实现了可行的加密操作(平均PQC签名时间:0.65 ms,验证时间:0.53 ms)。虽然由于PQC验证的复杂性,区块链集成会导致更高的天然气使用量,平均每次更新1.72 × 106单位,但我们证明,在天然气费用微不足道的私人财团设置中,这种成本可以忽略不计。重要的是,相对于交易时间的加密开销仍然很小(通常为 <; 0.2%),这证实了PQS-BFL是保护关键基础设施免受未来量子威胁的可行架构。
{"title":"PQS-BFL: A post-quantum secure blockchain-based federated learning framework","authors":"Daniel Commey ,&nbsp;Garth V. Crosby","doi":"10.1016/j.eswa.2026.131449","DOIUrl":"10.1016/j.eswa.2026.131449","url":null,"abstract":"<div><div>Federated Learning (FL) enables collaborative model training while preserving data privacy, but its classical cryptographic underpinnings are vulnerable to quantum attacks. This vulnerability is particularly critical in sensitive domains like healthcare, where data requires long-term forward secrecy. This paper introduces PQS-BFL (Post-Quantum Secure Blockchain-based Federated Learning), a framework integrating post-quantum cryptography (PQC) with blockchain verification to secure FL against quantum adversaries. We employ ML-DSA-65 (standardized in FIPS 204, formerly Dilithium) signatures to authenticate model updates and leverage optimized smart contracts for decentralized validation. Designed for permissioned consortium environments (e.g., healthcare research networks), our framework ensures update integrity independent of the underlying ledger’s quantum resistance. Extensive evaluations on diverse datasets (MNIST, SVHN, HAR) demonstrate that PQS-BFL achieves feasible cryptographic operations (average PQC sign time: 0.65 ms, verify time: 0.53 ms) with a fixed signature size of 3309 Bytes. While blockchain integration incurs higher gas usage averaging 1.72 × 10<sup>6</sup> units per update due to PQC verification complexity, we demonstrate that this cost is negligible in private consortium settings where gas fees are nominal. Importantly, the cryptographic overhead relative to transaction time remains minimal (typically  &lt; 0.2%), confirming that PQS-BFL is a viable architecture for securing critical infrastructure against future quantum threats.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131449"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vul2image: A quick image-inspired and CNN-based vulnerability detection system Vul2image:基于cnn的快速图像漏洞检测系统
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-04 DOI: 10.1016/j.eswa.2026.131468
Rong Ren , Mushi Zhou , Ni Liao , Bing Zhang , Guoyan Huang , Haitao He , Qian Wang
Given the accuracy of deep learning (DL) in image classification, some studies have applied DL algorithms to vulnerability detection by characterizing software source code as RGB images. However, effectively utilizing RGB images to store multiple code semantics remains a challenge, impacting the effectiveness of vulnerability detection. To address this, we developed Vul2image, a quick Image-inspired and CNN-based Vulnerability Detection System. By focusing on Potential Vulnerable Code Fragments (PVCFs) and their context code, Vul2image minimized interference from irrelevant information and achieved comprehensive coverage of vulnerability features. It constructed an RGB fine-grained image model incorporating textual, semantic, and structural information from code text, Control Dependency Graphs (CDGs), and Data Dependency Graphs (DDGs), resulting in improved detection efficiency. Evaluated on three datasets with increasing vulnerability types (including our self-collected, VulCNN, and Devign), Vul2image achieved the best results on our dataset, outperforming 9 classic (incl. 4 LLM-based) and 2 SOTA image-based detectors (VulCNN, VulGAI) and demonstrating performance comparable to 7 transformer-encoder-based methods, showing strong precision for specific vulnerability types. In practice, Vul2image was 35 times faster than VulCNN and successfully identified 21 reported and 5 unreported vulnerabilities in various real-world systems and software within 67,352,085 lines of code, showcasing its large-scale vulnerability detection capability.
考虑到深度学习在图像分类中的准确性,一些研究通过将软件源代码表征为RGB图像,将深度学习算法应用于漏洞检测。然而,有效地利用RGB图像存储多种代码语义仍然是一个挑战,影响了漏洞检测的有效性。为了解决这个问题,我们开发了Vul2image,一个快速的图像启发和基于cnn的漏洞检测系统。通过关注潜在脆弱代码片段(pvcf)及其上下文代码,Vul2image最大限度地减少了不相关信息的干扰,实现了对漏洞特征的全面覆盖。它构建了一个RGB细粒度图像模型,结合了来自代码文本、控制依赖图(cdg)和数据依赖图(ddg)的文本、语义和结构信息,从而提高了检测效率。在三个漏洞类型不断增加的数据集(包括我们的自收集、VulCNN和Devign)上进行评估,Vul2image在我们的数据集上取得了最好的结果,优于9个经典(包括4个基于llm的)和2个基于SOTA图像的检测器(VulCNN、VulGAI),并展示了与7种基于变压器编码器的方法相当的性能,对特定漏洞类型显示出很强的精度。在实践中,Vul2image比VulCNN快35倍,在67,352,085行代码中成功识别了各种现实系统和软件中的21个报告漏洞和5个未报告漏洞,展示了其大规模漏洞检测能力。
{"title":"Vul2image: A quick image-inspired and CNN-based vulnerability detection system","authors":"Rong Ren ,&nbsp;Mushi Zhou ,&nbsp;Ni Liao ,&nbsp;Bing Zhang ,&nbsp;Guoyan Huang ,&nbsp;Haitao He ,&nbsp;Qian Wang","doi":"10.1016/j.eswa.2026.131468","DOIUrl":"10.1016/j.eswa.2026.131468","url":null,"abstract":"<div><div>Given the accuracy of deep learning (DL) in image classification, some studies have applied DL algorithms to vulnerability detection by characterizing software source code as RGB images. However, effectively utilizing RGB images to store multiple code semantics remains a challenge, impacting the effectiveness of vulnerability detection. To address this, we developed Vul2image, a quick Image-inspired and CNN-based Vulnerability Detection System. By focusing on Potential Vulnerable Code Fragments (PVCFs) and their context code, Vul2image minimized interference from irrelevant information and achieved comprehensive coverage of vulnerability features. It constructed an RGB fine-grained image model incorporating textual, semantic, and structural information from code text, Control Dependency Graphs (CDGs), and Data Dependency Graphs (DDGs), resulting in improved detection efficiency. Evaluated on three datasets with increasing vulnerability types (including our self-collected, VulCNN, and Devign), Vul2image achieved the best results on our dataset, outperforming 9 classic (incl. 4 LLM-based) and 2 SOTA image-based detectors (VulCNN, VulGAI) and demonstrating performance comparable to 7 transformer-encoder-based methods, showing strong precision for specific vulnerability types. In practice, Vul2image was 35 times faster than VulCNN and successfully identified 21 reported and 5 unreported vulnerabilities in various real-world systems and software within 67,352,085 lines of code, showcasing its large-scale vulnerability detection capability.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131468"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DAWN: Dimension-aware graph contrastive learning for few-shot dissolved gas analysis DAWN:用于少量溶解气体分析的维度感知图对比学习
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-05 DOI: 10.1016/j.eswa.2026.131504
Jiyuan Sun , Huifang Ma , Shuai Yang , Rui Bing , Zhixin Li
Dissolved Gas Analysis (DGA) is a widely used technique for identifying characteristic gas signatures indicative of transformer faults. However, traditional DGA methods struggle to capture the complex interdependencies among gases. Although deep learning has shown promise in this domain, existing approaches face significant limitations under few-shot learning scenarios, primarily due to severe class imbalance and high inter-class similarity, which lead to diagnostic ambiguity. Moreover, these methods often overlook critical inter-gas relationships that are essential for understanding underlying fault mechanisms. To address these challenges, we propose DAWN (Dimension-AWare Graph CoNtrastive Learning), a novel framework that integrates two key components: (1) a contrastive few-shot learning module with clustering consistency loss to enhance the discriminability of similar fault categories through lightweight fine-tuning, and (2) a knowledge-enhanced dimension graph that explicitly models structural dependencies among gas features by combining statistical correlations with expert domain knowledge. Extensive evaluations on DGA datasets demonstrate that DAWN achieves state-of-the-art performance, improving rare fault detection accuracy by over 15% compared to conventional methods. To the best of our knowledge, this work represents the first contrastive few-shot learning framework tailored for DGA-based fault diagnosis.
溶解气体分析(DGA)是一种广泛应用于变压器故障特征气体特征识别的技术。然而,传统的DGA方法难以捕捉气体之间复杂的相互依赖关系。尽管深度学习在这一领域显示出了前景,但现有的方法在少数学习场景下面临着显著的局限性,主要是由于严重的类不平衡和高类间相似性,这导致了诊断的模糊性。此外,这些方法往往忽略了对理解潜在断层机制至关重要的气体间关系。为了解决这些挑战,我们提出了一个新的框架DAWN (dimension - aware Graph CoNtrastive Learning),该框架集成了两个关键组件:(1)具有聚类一致性损失的对比少shot学习模块,通过轻量级微调增强相似故障类别的可分辨性;(2)通过结合统计相关性和专家领域知识来显式建模天然气特征之间的结构依赖关系的知识增强维图。对DGA数据集的广泛评估表明,DAWN达到了最先进的性能,与传统方法相比,将罕见故障检测准确率提高了15%以上。据我们所知,这项工作代表了第一个为基于遗传算法的故障诊断量身定制的对比少镜头学习框架。
{"title":"DAWN: Dimension-aware graph contrastive learning for few-shot dissolved gas analysis","authors":"Jiyuan Sun ,&nbsp;Huifang Ma ,&nbsp;Shuai Yang ,&nbsp;Rui Bing ,&nbsp;Zhixin Li","doi":"10.1016/j.eswa.2026.131504","DOIUrl":"10.1016/j.eswa.2026.131504","url":null,"abstract":"<div><div>Dissolved Gas Analysis (DGA) is a widely used technique for identifying characteristic gas signatures indicative of transformer faults. However, traditional DGA methods struggle to capture the complex interdependencies among gases. Although deep learning has shown promise in this domain, existing approaches face significant limitations under few-shot learning scenarios, primarily due to severe class imbalance and high inter-class similarity, which lead to diagnostic ambiguity. Moreover, these methods often overlook critical inter-gas relationships that are essential for understanding underlying fault mechanisms. To address these challenges, we propose <strong>DAWN</strong> (<strong>D</strong>imension-<strong>AW</strong>are Graph Co<strong>N</strong>trastive Learning), a novel framework that integrates two key components: (1) a contrastive few-shot learning module with clustering consistency loss to enhance the discriminability of similar fault categories through lightweight fine-tuning, and (2) a knowledge-enhanced dimension graph that explicitly models structural dependencies among gas features by combining statistical correlations with expert domain knowledge. Extensive evaluations on DGA datasets demonstrate that DAWN achieves state-of-the-art performance, improving rare fault detection accuracy by over 15% compared to conventional methods. To the best of our knowledge, this work represents the first contrastive few-shot learning framework tailored for DGA-based fault diagnosis.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131504"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MI-DGHCL: Motor imagery EEG domain generalization via hyperbolic contrastive learning MI-DGHCL:基于双曲对比学习的运动意象脑电域泛化
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-02 DOI: 10.1016/j.eswa.2026.131477
Junfu Chen , Dechang Pi , Feng Gao , Cheng Ma , Yang Chen
Cross-subject motor imagery (MI) Electroencephalogram (EEG) decoding remains challenging due to significant domain shifts caused by inter-subject variability. Despite the success of Euclidean domain alignment techniques via deep learning, they fail to capture the potential hierarchical structures inherent in EEG signals. To address this challenge, we propose MI-DGHCL, a domain generalization model for MI-EEG signals that leverages hyperbolic contrastive learning. MI-DGHCL introduces a Mamba-based feature extractor, incorporating slice feature embedding and slice-aware scanning modules that simultaneously capture both global information and local contextual features of EEG signals. Furthermore, it enforces both feature consistency and semantic consistency. Feature consistency is attained by aligning domains through the minimization of feature covariance in Euclidean space. Subsequently, hierarchical representations are derived via hyperbolic embeddings, and supervised contrastive learning pushes intra-class samples across subjects close in hyperbolic space, thereby ensuring semantic consistency. Comprehensive experiments on 3 MI-EEG datasets demonstrate MI-DGHCL yields superior results, compared to advanced methods.
跨主体运动图像(MI)脑电图(EEG)解码仍然具有挑战性,因为主体间可变性引起了显著的域转移。尽管通过深度学习的欧几里得域对齐技术取得了成功,但它们未能捕获脑电图信号中固有的潜在层次结构。为了解决这一挑战,我们提出了MI-DGHCL,这是一种利用双曲对比学习的MI-EEG信号的域泛化模型。MI-DGHCL引入了一种基于曼巴的特征提取器,结合了切片特征嵌入和切片感知扫描模块,可以同时捕获EEG信号的全局信息和局部上下文特征。此外,它还加强了特征一致性和语义一致性。特征一致性是通过在欧几里得空间中最小化特征协方差来对齐域来实现的。随后,通过双曲嵌入推导出层次表示,监督对比学习将类内样本推进到双曲空间中,从而确保语义一致性。在3个MI-EEG数据集上的综合实验表明,与先进的方法相比,MI-DGHCL方法取得了更好的结果。
{"title":"MI-DGHCL: Motor imagery EEG domain generalization via hyperbolic contrastive learning","authors":"Junfu Chen ,&nbsp;Dechang Pi ,&nbsp;Feng Gao ,&nbsp;Cheng Ma ,&nbsp;Yang Chen","doi":"10.1016/j.eswa.2026.131477","DOIUrl":"10.1016/j.eswa.2026.131477","url":null,"abstract":"<div><div>Cross-subject motor imagery (MI) Electroencephalogram (EEG) decoding remains challenging due to significant domain shifts caused by inter-subject variability. Despite the success of Euclidean domain alignment techniques via deep learning, they fail to capture the potential hierarchical structures inherent in EEG signals. To address this challenge, we propose MI-DGHCL, a domain generalization model for MI-EEG signals that leverages hyperbolic contrastive learning. MI-DGHCL introduces a Mamba-based feature extractor, incorporating slice feature embedding and slice-aware scanning modules that simultaneously capture both global information and local contextual features of EEG signals. Furthermore, it enforces both feature consistency and semantic consistency. Feature consistency is attained by aligning domains through the minimization of feature covariance in Euclidean space. Subsequently, hierarchical representations are derived via hyperbolic embeddings, and supervised contrastive learning pushes intra-class samples across subjects close in hyperbolic space, thereby ensuring semantic consistency. Comprehensive experiments on 3 MI-EEG datasets demonstrate MI-DGHCL yields superior results, compared to advanced methods.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131477"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards privacy-preserving and communication-efficient federated distillation 面向隐私保护和通信高效的联邦蒸馏
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-03 DOI: 10.1016/j.eswa.2026.131503
Xinge Ma, Jin Wang, Xuejie Zhang
Federated distillation (FD) has emerged as a promising alternative to federated learning (FL) for collaborative training across decentralized clients to benefit from their private data by exchanging model outputs associated with a large-scale unlabeled public dataset. However, recent studies have revealed that sharing model outputs still poses privacy risks of data exposure when encountering malicious attacks. Although incorporating differential privacy (DP) can provide strong privacy guarantees for FD by perturbing model parameters to produce secure model outputs or directly injecting calibrated noise to model outputs before sharing, it either suffer from inefficient knowledge transfer due to the limited domain-specific knowledge learned by local models or incurs a high privacy cost that significantly compromises the utility of model outputs because the required noise magnitude is proportional to the scale of the perturbed target. To balance the trade-off between knowledge utility and privacy protection, this paper presents FedLA, a privacy-preserving and communication-efficient FD framework empowered by local differential privacy and active data sampling, which proactively selects the most informative subset from the large-scale unlabeled public dataset as a high-quality carrier for local perturbation and knowledge transfer. The resulting reduction in the number of queries to local models minimizes privacy cost and communication overhead while maximizing model performance. Experiments on two popular benchmark datasets across diverse evaluation settings demonstrate the superiority of FedLA in terms of model accuracy, communication efficiency, privacy cost, and attack defense. The code is available at: https://github.com/maxinge8698/FedLA.
联邦蒸馏(FD)已经成为联邦学习(FL)的一个有前途的替代方案,用于跨分散客户端的协作训练,通过交换与大规模未标记的公共数据集相关的模型输出,从他们的私有数据中受益。然而,最近的研究表明,共享模型输出在遭遇恶意攻击时仍然存在数据暴露的隐私风险。虽然结合差分隐私(DP)可以通过扰动模型参数以产生安全的模型输出或在共享之前直接向模型输出注入校准过的噪声,为FD提供强大的隐私保证,由于局部模型学习到的特定领域的知识有限,它要么存在知识转移效率低下的问题,要么由于所需的噪声大小与受干扰目标的规模成正比,从而导致高昂的隐私成本,从而严重损害了模型输出的效用。为了平衡知识效用和隐私保护之间的平衡,本文提出了一种基于局部差分隐私和主动数据采样的隐私保护和通信高效FD框架,该框架主动从大规模未标记的公共数据集中选择最具信息量的子集作为局部扰动和知识转移的高质量载体。对本地模型的查询数量的减少将最小化隐私成本和通信开销,同时最大化模型性能。在两个流行的基准数据集上进行的不同评估设置的实验表明,FedLA在模型精度、通信效率、隐私成本和攻击防御方面具有优势。代码可从https://github.com/maxinge8698/FedLA获得。
{"title":"Towards privacy-preserving and communication-efficient federated distillation","authors":"Xinge Ma,&nbsp;Jin Wang,&nbsp;Xuejie Zhang","doi":"10.1016/j.eswa.2026.131503","DOIUrl":"10.1016/j.eswa.2026.131503","url":null,"abstract":"<div><div>Federated distillation (FD) has emerged as a promising alternative to federated learning (FL) for collaborative training across decentralized clients to benefit from their private data by exchanging model outputs associated with a large-scale unlabeled public dataset. However, recent studies have revealed that sharing model outputs still poses privacy risks of data exposure when encountering malicious attacks. Although incorporating differential privacy (DP) can provide strong privacy guarantees for FD by perturbing model parameters to produce secure model outputs or directly injecting calibrated noise to model outputs before sharing, it either suffer from inefficient knowledge transfer due to the limited domain-specific knowledge learned by local models or incurs a high privacy cost that significantly compromises the utility of model outputs because the required noise magnitude is proportional to the scale of the perturbed target. To balance the trade-off between knowledge utility and privacy protection, this paper presents FedLA, a privacy-preserving and communication-efficient FD framework empowered by local differential privacy and active data sampling, which proactively selects the most informative subset from the large-scale unlabeled public dataset as a high-quality carrier for local perturbation and knowledge transfer. The resulting reduction in the number of queries to local models minimizes privacy cost and communication overhead while maximizing model performance. Experiments on two popular benchmark datasets across diverse evaluation settings demonstrate the superiority of FedLA in terms of model accuracy, communication efficiency, privacy cost, and attack defense. The code is available at: <span><span>https://github.com/maxinge8698/FedLA</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131503"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A dynamic model of rumor propagation based on adversarial behavior and evolutionary games 基于对抗行为和进化博弈的谣言传播动态模型
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-05 DOI: 10.1016/j.eswa.2026.131490
Chaolong Jia , Guicai Deng , Xiaochuan Chen , Kangle Chen , Rong Wang , Tun Li , Yunpeng Xiao
Rumors pose a serious threat to social stability across modern social networks. To address the ongoing confrontation between rumor and anti-rumor groups, this study proposes a framework for analyzing rumor dissemination based on confrontational behavior and evolutionary game theory. This study considers differences in users’ real-time preferences for rumors relative to anti-rumors, which significantly influence rumor spread and are complex to measure. Multivariate linear regression is applied to develop a real-time preference metric and identify multiple factors influencing user preferences. Evolutionary game theory is then employed to construct a game-theoretic mechanism for anti-rumor behavior. The coexistence and confrontation between the two groups were analyzed using the Rosenzweig–MacArthur equations to characterize overall rumor propagation dynamics. For individual users, three states are defined after exposure to both rumor types: a wavering state (I), a rumor-adopting state (P), and an anti-rumor state (O), which form the basis of a rumor dissemination dynamics model. Experimental results show that the proposed model achieves an average SMAPE of 16%, while complementary metrics such as RMSE and AUC further confirm its reliable performance across different evaluation dimensions.
在现代社交网络中,谣言对社会稳定构成严重威胁。针对谣言与反谣言群体之间持续的对抗,本研究提出了一个基于对抗行为和进化博弈论的谣言传播分析框架。本研究考虑了用户对谣言和反谣言的实时偏好的差异,这对谣言的传播有显著影响,并且测量起来很复杂。应用多元线性回归建立实时偏好度量,识别影响用户偏好的多个因素。运用进化博弈论构建了反谣言行为的博弈论机制。使用Rosenzweig-MacArthur方程分析两组之间的共存和对抗,以表征谣言的整体传播动态。对于个人用户而言,在接触两种谣言类型后,定义了三种状态:摇摆状态(I)、接受谣言状态(P)和反谣言状态(O),这三种状态构成了谣言传播动力学模型的基础。实验结果表明,该模型的平均SMAPE达到了16%,而RMSE和AUC等互补指标进一步证实了其在不同评估维度上的可靠性能。
{"title":"A dynamic model of rumor propagation based on adversarial behavior and evolutionary games","authors":"Chaolong Jia ,&nbsp;Guicai Deng ,&nbsp;Xiaochuan Chen ,&nbsp;Kangle Chen ,&nbsp;Rong Wang ,&nbsp;Tun Li ,&nbsp;Yunpeng Xiao","doi":"10.1016/j.eswa.2026.131490","DOIUrl":"10.1016/j.eswa.2026.131490","url":null,"abstract":"<div><div>Rumors pose a serious threat to social stability across modern social networks. To address the ongoing confrontation between rumor and anti-rumor groups, this study proposes a framework for analyzing rumor dissemination based on confrontational behavior and evolutionary game theory. This study considers differences in users’ real-time preferences for rumors relative to anti-rumors, which significantly influence rumor spread and are complex to measure. Multivariate linear regression is applied to develop a real-time preference metric and identify multiple factors influencing user preferences. Evolutionary game theory is then employed to construct a game-theoretic mechanism for anti-rumor behavior. The coexistence and confrontation between the two groups were analyzed using the Rosenzweig–MacArthur equations to characterize overall rumor propagation dynamics. For individual users, three states are defined after exposure to both rumor types: a wavering state (I), a rumor-adopting state (P), and an anti-rumor state (O), which form the basis of a rumor dissemination dynamics model. Experimental results show that the proposed model achieves an average SMAPE of 16%, while complementary metrics such as RMSE and AUC further confirm its reliable performance across different evaluation dimensions.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131490"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep global-ranking hashing via average precision approximation for large-scale image retrieval 基于平均精度近似的深度全局排序哈希算法用于大规模图像检索
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-06 DOI: 10.1016/j.eswa.2026.131557
Lei Wang , Yongyue Fu , Qibing Qin , Lei Huang , Wenfang Zhang
Deep hashing algorithms have become a mainstream solution for large-scale multimedia retrieval due to their advantages in search efficiency and storage space. The algorithm is able to jointly learn semantic features and hash functions to encode raw data into compact binary codes, with significant differentiation. For the retrieval task, the central goal is to learn a ranking relation that can efficiently rank candidate results. Most of the existing hash methods adopt pair-wise or multi-wise strategies to learn the ranking results by minimizing the relative distance between similar samples or maximizing the distance between dissimilar samples. This type of approach can indeed improve the consistency of the local neighborhood, but its optimization objective is essentially limited to the local ranking relationship and fails to globally rank the samples in the entire retrieval set. Ultimately, this leads to the model possibly achieving good performance in local ranking but being unable to guarantee the overall relevance and stability of the global retrieval list. Especially when the sample distribution is complex or there is significant semantic overlap within or between categories; the local ranking method cannot fully capture the relationships among different samples, thereby limiting the discriminative ability and ranking quality of hash codes in actual retrieval scenarios. To address this issue, by introducing the average precision (AP) metric as the optimization objective, a novel Deep Global-ranking Hashing framework via average precision approximation (DGrH) is proposed to learn hash spaces with ranking relationships. Specifically, based on the discrete Heaviside function, a novel Ranking-AP optimization strategy is introduced into deep hashing to learn the global semantic relationships. The Sigmoid function is employed to smoothly approximate the non-differentiable discrete Heaviside function, making it differentiable. On this basis, the overall objective function and the novel Ranking-AP loss could enhance the learning of global ranking information. This helps capture and preserve high-quality high-quality global ranking relationships among samples more effectively in hash code learning. Extensive experiments on several benchmark datasets validate the efficacy of our designed DGrH framework, which consistently outperforms the mainstream deep hashing by large gaps. The code for the implementation of our DGrH framework is available at https://github.com/QinLab-WFU/DGrH.
深度哈希算法以其在搜索效率和存储空间方面的优势,成为大规模多媒体检索的主流解决方案。该算法能够联合学习语义特征和哈希函数,将原始数据编码为紧凑的二进制代码,具有显著的差异性。对于检索任务,中心目标是学习一种排序关系,可以有效地对候选结果进行排序。现有的哈希方法大多采用对智或多智策略,通过最小化相似样本之间的相对距离或最大化不相似样本之间的距离来学习排序结果。这种方法确实可以提高局部邻域的一致性,但其优化目标本质上局限于局部排序关系,无法对整个检索集中的样本进行全局排序。最终,这将导致该模型可能在局部排序中获得良好的性能,但无法保证全局检索列表的整体相关性和稳定性。特别是当样本分布复杂或类别内或类别之间存在明显的语义重叠时;局部排序方法不能充分捕捉不同样本之间的关系,从而限制了实际检索场景中哈希码的判别能力和排序质量。为了解决这一问题,通过引入平均精度(AP)度量作为优化目标,提出了一种新的基于平均精度近似(DGrH)的深度全局排序哈希框架,以学习具有排序关系的哈希空间。具体而言,基于离散的Heaviside函数,在深度哈希中引入了一种新的rank - ap优化策略来学习全局语义关系。利用Sigmoid函数平滑逼近不可微的离散Heaviside函数,使其可微。在此基础上,总体目标函数和新的rank - ap损失可以增强全局排名信息的学习。这有助于在哈希码学习中更有效地捕获和保存样本之间的高质量的高质量全局排序关系。在多个基准数据集上进行的大量实验验证了我们设计的DGrH框架的有效性,该框架始终优于主流深度哈希。实现DGrH框架的代码可从https://github.com/QinLab-WFU/DGrH获得。
{"title":"Deep global-ranking hashing via average precision approximation for large-scale image retrieval","authors":"Lei Wang ,&nbsp;Yongyue Fu ,&nbsp;Qibing Qin ,&nbsp;Lei Huang ,&nbsp;Wenfang Zhang","doi":"10.1016/j.eswa.2026.131557","DOIUrl":"10.1016/j.eswa.2026.131557","url":null,"abstract":"<div><div>Deep hashing algorithms have become a mainstream solution for large-scale multimedia retrieval due to their advantages in search efficiency and storage space. The algorithm is able to jointly learn semantic features and hash functions to encode raw data into compact binary codes, with significant differentiation. For the retrieval task, the central goal is to learn a ranking relation that can efficiently rank candidate results. Most of the existing hash methods adopt pair-wise or multi-wise strategies to learn the ranking results by minimizing the relative distance between similar samples or maximizing the distance between dissimilar samples. This type of approach can indeed improve the consistency of the local neighborhood, but its optimization objective is essentially limited to the local ranking relationship and fails to globally rank the samples in the entire retrieval set. Ultimately, this leads to the model possibly achieving good performance in local ranking but being unable to guarantee the overall relevance and stability of the global retrieval list. Especially when the sample distribution is complex or there is significant semantic overlap within or between categories; the local ranking method cannot fully capture the relationships among different samples, thereby limiting the discriminative ability and ranking quality of hash codes in actual retrieval scenarios. To address this issue, by introducing the average precision (AP) metric as the optimization objective, a novel Deep Global-ranking Hashing framework via average precision approximation (DGrH) is proposed to learn hash spaces with ranking relationships. Specifically, based on the discrete Heaviside function, a novel Ranking-AP optimization strategy is introduced into deep hashing to learn the global semantic relationships. The Sigmoid function is employed to smoothly approximate the non-differentiable discrete Heaviside function, making it differentiable. On this basis, the overall objective function and the novel Ranking-AP loss could enhance the learning of global ranking information. This helps capture and preserve high-quality high-quality global ranking relationships among samples more effectively in hash code learning. Extensive experiments on several benchmark datasets validate the efficacy of our designed DGrH framework, which consistently outperforms the mainstream deep hashing by large gaps. The code for the implementation of our DGrH framework is available at <span><span>https://github.com/QinLab-WFU/DGrH</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131557"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Privacy preservation in face soft biometrics via attribute disentanglement 基于属性解缠的人脸软生物识别隐私保护
IF 7.5 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2026-05-25 Epub Date: 2026-02-04 DOI: 10.1016/j.eswa.2026.131520
Yue Wang , Biao Jin , Zheyu Chen , Jinsen Lin , Zhiqiang Yao
Soft biometric privacy enhancement techniques have been widely adopted in face recognition systems to prevent attackers from inferring sensitive attributes such as gender, age, and ethnicity from facial images. Although existing facial attribute privacy protection methods can conceal multiple attributes simultaneously, they still face three key challenges: (1) precisely modifying target attributes while preserving non-target attributes; (2) achieving a balanced trade-off between privacy preservation and identity recognition utility; and (3) providing flexible and user-controllable options for attribute protection. To address these challenges, this paper proposes a novel face soft biometric privacy protection framework based on attribute disentanglement, which effectively conceals sensitive facial attributes while maximizing identity recognition accuracy. First, a mapping module guided by an attribute supervision loss is introduced to learn a disentangled latent space, where the semantic representations of different attributes are separated for controllable manipulation. Second, a face matcher combined with a dedicated face matching loss enforces identity consistency, enabling the model to preserve recognition utility while suppressing sensitive attribute leakage. Finally, an attribute selection module (ASM) is incorporated during the inference stage, allowing users to flexibly specify which attributes (e.g., gender, age, and smile) to protect, thereby enhancing adaptability and user-level controllability in privacy-sensitive applications. Experimental results demonstrate that the proposed method effectively safeguards the privacy of facial attributes while maintaining high identity recognition utility. Code is available at https://github.com/Forestmumu/PrivAD.
软生物特征隐私增强技术已广泛应用于人脸识别系统中,以防止攻击者从人脸图像中推断出性别、年龄和种族等敏感属性。现有的面部属性隐私保护方法虽然可以同时隐藏多个属性,但仍然面临三个关键挑战:(1)在保留非目标属性的同时精确修改目标属性;(2)在隐私保护和身份识别效用之间取得平衡;(3)提供灵活可控的属性保护选项。针对这些问题,本文提出了一种基于属性解纠缠的人脸软生物特征隐私保护框架,该框架在有效隐藏敏感人脸属性的同时,最大限度地提高了身份识别的准确性。首先,引入以属性监督损失为导向的映射模块学习解纠缠潜空间,在潜空间中分离不同属性的语义表示以进行可控操作;其次,人脸匹配器与专用的人脸匹配损失相结合,增强身份一致性,使模型在保持识别效用的同时抑制敏感属性泄漏。最后,在推理阶段引入属性选择模块(ASM),允许用户灵活指定需要保护的属性(如性别、年龄、微笑),从而增强隐私敏感应用的适应性和用户级可控性。实验结果表明,该方法有效地保护了人脸属性的隐私,同时保持了较高的身份识别效用。代码可从https://github.com/Forestmumu/PrivAD获得。
{"title":"Privacy preservation in face soft biometrics via attribute disentanglement","authors":"Yue Wang ,&nbsp;Biao Jin ,&nbsp;Zheyu Chen ,&nbsp;Jinsen Lin ,&nbsp;Zhiqiang Yao","doi":"10.1016/j.eswa.2026.131520","DOIUrl":"10.1016/j.eswa.2026.131520","url":null,"abstract":"<div><div>Soft biometric privacy enhancement techniques have been widely adopted in face recognition systems to prevent attackers from inferring sensitive attributes such as gender, age, and ethnicity from facial images. Although existing facial attribute privacy protection methods can conceal multiple attributes simultaneously, they still face three key challenges: (1) precisely modifying target attributes while preserving non-target attributes; (2) achieving a balanced trade-off between privacy preservation and identity recognition utility; and (3) providing flexible and user-controllable options for attribute protection. To address these challenges, this paper proposes a novel face soft biometric privacy protection framework based on attribute disentanglement, which effectively conceals sensitive facial attributes while maximizing identity recognition accuracy. First, a mapping module guided by an attribute supervision loss is introduced to learn a disentangled latent space, where the semantic representations of different attributes are separated for controllable manipulation. Second, a face matcher combined with a dedicated face matching loss enforces identity consistency, enabling the model to preserve recognition utility while suppressing sensitive attribute leakage. Finally, an attribute selection module (ASM) is incorporated during the inference stage, allowing users to flexibly specify which attributes (e.g., gender, age, and smile) to protect, thereby enhancing adaptability and user-level controllability in privacy-sensitive applications. Experimental results demonstrate that the proposed method effectively safeguards the privacy of facial attributes while maintaining high identity recognition utility. Code is available at <span><span>https://github.com/Forestmumu/PrivAD</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"312 ","pages":"Article 131520"},"PeriodicalIF":7.5,"publicationDate":"2026-05-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146192851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Expert Systems with Applications
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1