首页 > 最新文献

Neural Networks最新文献

英文 中文
Stability and passivity analysis of delayed neural networks via an improved matrix-valued polynomial inequality 通过改进的矩阵值多项式不等式分析延迟神经网络的稳定性和被动性
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-21 DOI: 10.1016/j.neunet.2024.106637

The stability and passivity of delayed neural networks are addressed in this paper. A novel Lyapunov–Krasovskii functional (LKF) without multiple integrals is constructed. By using an improved matrix-valued polynomial inequality (MVPI), the previous constraint involving skew-symmetric matrices within the MVPI is removed. Then, the stability and passivity criteria for delayed neural networks that are less conservative than the existing ones are proposed. Finally, three examples are employed to demonstrate the meliority and feasibility of the obtained results.

本文探讨了延迟神经网络的稳定性和被动性。本文构建了一个新颖的无多重积分的 Lyapunov-Krasovskii 函数 (LKF)。通过使用改进的矩阵值多项式不等式(MVPI),消除了 MVPI 中以前涉及倾斜对称矩阵的约束。然后,提出了比现有标准更保守的延迟神经网络稳定性和被动性标准。最后,通过三个实例证明了所获结果的优越性和可行性。
{"title":"Stability and passivity analysis of delayed neural networks via an improved matrix-valued polynomial inequality","authors":"","doi":"10.1016/j.neunet.2024.106637","DOIUrl":"10.1016/j.neunet.2024.106637","url":null,"abstract":"<div><p>The stability and passivity of delayed neural networks are addressed in this paper. A novel Lyapunov–Krasovskii functional (LKF) without multiple integrals is constructed. By using an improved matrix-valued polynomial inequality (MVPI), the previous constraint involving skew-symmetric matrices within the MVPI is removed. Then, the stability and passivity criteria for delayed neural networks that are less conservative than the existing ones are proposed. Finally, three examples are employed to demonstrate the meliority and feasibility of the obtained results.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142049203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning the feature distribution similarities for online time series anomaly detection 为在线时间序列异常检测学习特征分布相似性
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-21 DOI: 10.1016/j.neunet.2024.106638

Identifying anomalies in multi-dimensional sequential data is crucial for ensuring optimal performance across various domains and in large-scale systems. Traditional contrastive methods utilize feature similarity between different features extracted from multidimensional raw inputs as an indicator of anomaly severity. However, the complex objective functions and meticulously designed modules of these methods often lead to efficiency issues and a lack of interpretability. Our study introduces a structural framework called SimDetector, which is a Local–Global Multi-Scale Similarity Contrast network. Specifically, the restructured and enhanced GRU module extracts more generalized local features, including long-term cyclical trends. The multi-scale sparse attention module efficiently extracts multi-scale global features with pattern information. Additionally, we modified the KL divergence to suit the characteristics of time series anomaly detection, proposing a symmetric absolute KL divergence that focuses more on overall distribution differences. The proposed method achieves results that surpass or approach the State-of-the-Art (SOTA) on multiple real-world datasets and synthetic datasets, while also significantly reducing Multiply-Accumulate Operations (MACs) and memory usage.

识别多维序列数据中的异常情况,对于确保各领域和大规模系统的最佳性能至关重要。传统的对比方法利用从多维原始输入中提取的不同特征之间的特征相似性作为异常严重程度的指标。然而,这些方法复杂的目标函数和精心设计的模块往往会导致效率问题和缺乏可解释性。我们的研究引入了一个名为 SimDetector 的结构框架,它是一个局部-全局多尺度相似性对比网络。具体来说,经过重组和增强的 GRU 模块能提取出更具普遍性的局部特征,包括长期周期性趋势。多尺度稀疏关注模块能有效提取具有模式信息的多尺度全局特征。此外,我们还根据时间序列异常检测的特点修改了 KL 发散,提出了一种对称的绝对 KL 发散,更加关注整体分布差异。所提出的方法在多个真实世界数据集和合成数据集上取得了超越或接近最新技术水平(SOTA)的结果,同时还显著减少了乘积运算(MAC)和内存使用量。
{"title":"Learning the feature distribution similarities for online time series anomaly detection","authors":"","doi":"10.1016/j.neunet.2024.106638","DOIUrl":"10.1016/j.neunet.2024.106638","url":null,"abstract":"<div><p>Identifying anomalies in multi-dimensional sequential data is crucial for ensuring optimal performance across various domains and in large-scale systems. Traditional contrastive methods utilize feature similarity between different features extracted from multidimensional raw inputs as an indicator of anomaly severity. However, the complex objective functions and meticulously designed modules of these methods often lead to efficiency issues and a lack of interpretability. Our study introduces a structural framework called SimDetector, which is a Local–Global Multi-Scale Similarity Contrast network. Specifically, the restructured and enhanced GRU module extracts more generalized local features, including long-term cyclical trends. The multi-scale sparse attention module efficiently extracts multi-scale global features with pattern information. Additionally, we modified the KL divergence to suit the characteristics of time series anomaly detection, proposing a symmetric absolute KL divergence that focuses more on overall distribution differences. The proposed method achieves results that surpass or approach the State-of-the-Art (SOTA) on multiple real-world datasets and synthetic datasets, while also significantly reducing Multiply-Accumulate Operations (MACs) and memory usage.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SNN-BERT: Training-efficient Spiking Neural Networks for energy-efficient BERT SNN-BERT:用于高能效 BERT 的训练高效尖峰神经网络
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-20 DOI: 10.1016/j.neunet.2024.106630

Spiking Neural Networks (SNNs) are naturally suited to process sequence tasks such as NLP with low power, due to its brain-inspired spatio-temporal dynamics and spike-driven nature. Current SNNs employ ”repeat coding” that re-enter all input tokens at each timestep, which fails to fully exploit temporal relationships between the tokens and introduces memory overhead. In this work, we align the number of input tokens with the timestep and refer to this input coding as ”individual coding”. To cope with the increase in training time for individual encoded SNNs due to the dramatic increase in timesteps, we design a Bidirectional Parallel Spiking Neuron (BPSN) with following features: First, BPSN supports spike parallel computing and effectively avoids the issue of uninterrupted firing; Second, BPSN excels in handling adaptive sequence length tasks, which is a capability that existing work does not have; Third, the fusion of bidirectional information enhances the temporal information modeling capabilities of SNNs; To validate the effectiveness of our BPSN, we present the SNN-BERT, a deep direct training SNN architecture based on the BERT model in NLP. Compared to prior repeat 4-timestep coding baseline, our method achieves a 6.46× reduction in energy consumption and a significant 16.1% improvement, raising the performance upper bound of the SNN domain on the GLUE dataset to 74.4%. Additionally, our method achieves 3.5× training acceleration and 3.8× training memory optimization. Compared with artificial neural networks of similar architecture, we obtain comparable performance but up to 22.5× energy efficiency. We would provide the codes.

尖峰神经网络(SNN)具有受大脑启发的时空动态和尖峰驱动特性,因此非常适合以较低功耗处理 NLP 等序列任务。目前的神经元网络采用 "重复编码 "技术,在每个时间步重新输入所有输入标记,这种方法无法充分利用标记之间的时间关系,而且会带来内存开销。在这项研究中,我们将输入标记的数量与时间步长保持一致,并将这种输入编码称为 "单个编码"。为了应对单个编码 SNNs 因时间步长大幅增加而导致的训练时间增加,我们设计了一种具有以下特点的双向并行尖峰神经元(BPSN):首先,BPSN 支持尖峰并行计算,有效避免了不间断发射的问题;其次,BPSN 擅长处理自适应序列长度任务,这是现有工作所不具备的能力;第三,双向信息的融合增强了 SNN 的时间信息建模能力;为了验证 BPSN 的有效性,我们提出了 SNN-BERT,一种基于 NLP 中 BERT 模型的深度直接训练 SNN 架构。与之前的重复四步编码基线相比,我们的方法减少了 6.46 倍的能耗,显著提高了 16.1%,将 SNN 领域在 GLUE 数据集上的性能上限提高到了 74.4%。此外,我们的方法还实现了 3.5 倍的训练加速和 3.8 倍的训练内存优化。与类似架构的人工神经网络相比,我们的方法性能相当,但能效高达 22.5 倍。我们将提供相关代码。
{"title":"SNN-BERT: Training-efficient Spiking Neural Networks for energy-efficient BERT","authors":"","doi":"10.1016/j.neunet.2024.106630","DOIUrl":"10.1016/j.neunet.2024.106630","url":null,"abstract":"<div><p>Spiking Neural Networks (SNNs) are naturally suited to process sequence tasks such as NLP with low power, due to its brain-inspired spatio-temporal dynamics and spike-driven nature. Current SNNs employ ”repeat coding” that re-enter all input tokens at each timestep, which fails to fully exploit temporal relationships between the tokens and introduces memory overhead. In this work, we align the number of input tokens with the timestep and refer to this input coding as ”individual coding”. To cope with the increase in training time for individual encoded SNNs due to the dramatic increase in timesteps, we design a Bidirectional Parallel Spiking Neuron (BPSN) with following features: First, BPSN supports spike parallel computing and effectively avoids the issue of uninterrupted firing; Second, BPSN excels in handling adaptive sequence length tasks, which is a capability that existing work does not have; Third, the fusion of bidirectional information enhances the temporal information modeling capabilities of SNNs; To validate the effectiveness of our BPSN, we present the SNN-BERT, a deep direct training SNN architecture based on the BERT model in NLP. Compared to prior repeat 4-timestep coding baseline, our method achieves a 6.46<span><math><mo>×</mo></math></span> reduction in energy consumption and a significant 16.1% improvement, raising the performance upper bound of the SNN domain on the GLUE dataset to 74.4%. Additionally, our method achieves 3.5<span><math><mo>×</mo></math></span> training acceleration and 3.8<span><math><mo>×</mo></math></span> training memory optimization. Compared with artificial neural networks of similar architecture, we obtain comparable performance but up to 22.5<span><math><mo>×</mo></math></span> energy efficiency. We would provide the codes.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Interplay between depth and width for interpolation in neural ODEs 神经 ODE 中插值的深度和宽度之间的相互作用
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-19 DOI: 10.1016/j.neunet.2024.106640

Neural ordinary differential equations have emerged as a natural tool for supervised learning from a control perspective, yet a complete understanding of the role played by their architecture remains elusive. In this work, we examine the interplay between the width p and the number of transitions between layers L (corresponding to a depth of L+1). Specifically, we construct explicit controls interpolating either a finite dataset D, comprising N pairs of points in Rd, or two probability measures within a Wasserstein error margin ɛ>0. Our findings reveal a balancing trade-off between p and L, with L scaling as 1+O(N/p) for data interpolation, and as 1+Op1+(1+p)1ɛd for measures.

In the high-dimensional and wide setting where d,p>N, our result can be refined to achieve L=0. This naturally raises the problem of data interpolation in the autonomous regime, characterized by L=0. We adopt two alternative approaches: either controlling in a probabilistic sense, or by relaxing the target condition. In the first case, when p=N we develop an inductive control strategy based on a separability assumption whose probability increases with d. In the second one, we establish an explicit error decay rate with respect to p which results from applying a universal approximation theorem to a custom-built Lipschitz vector field interpolating D.

从控制的角度来看,神经常微分方程已成为监督学习的一种天然工具,但人们对其结构所起作用的全面了解却仍然遥不可及。在这项工作中,我们研究了宽度 p 与层间转换次数 L(对应深度 L+1)之间的相互作用。具体来说,我们构建了明确的控制方法,既可以对由 Rd 中 N 对点组成的有限数据集 D 进行插值,也可以对 Wasserstein 误差范围ɛ>0 内的两个概率度量进行插值。我们的发现揭示了 p 和 L 之间的平衡权衡,对于数据插值,L 的缩放为 1+O(N/p),而对于度量,L 的缩放为 1+Op-1+(1+p)-1ɛ-d。在 d,p>N 的高维和宽范围设置中,我们的结果可以细化到 L=0。在第一种情况下,当 p=N 时,我们基于可分性假设开发了一种归纳控制策略,其概率随 d 的增加而增加。在第二种情况下,我们建立了一个与 p 有关的显式误差衰减率,该误差衰减率是将通用近似定理应用于定制的利普斯奇茨矢量场插值 D 的结果。
{"title":"Interplay between depth and width for interpolation in neural ODEs","authors":"","doi":"10.1016/j.neunet.2024.106640","DOIUrl":"10.1016/j.neunet.2024.106640","url":null,"abstract":"<div><p>Neural ordinary differential equations have emerged as a natural tool for supervised learning from a control perspective, yet a complete understanding of the role played by their architecture remains elusive. In this work, we examine the interplay between the width <span><math><mi>p</mi></math></span> and the number of transitions between layers <span><math><mi>L</mi></math></span> (corresponding to a depth of <span><math><mrow><mi>L</mi><mo>+</mo><mn>1</mn></mrow></math></span>). Specifically, we construct explicit controls interpolating either a finite dataset <span><math><mi>D</mi></math></span>, comprising <span><math><mi>N</mi></math></span> pairs of points in <span><math><msup><mrow><mi>R</mi></mrow><mrow><mi>d</mi></mrow></msup></math></span>, or two probability measures within a Wasserstein error margin <span><math><mrow><mi>ɛ</mi><mo>&gt;</mo><mn>0</mn></mrow></math></span>. Our findings reveal a balancing trade-off between <span><math><mi>p</mi></math></span> and <span><math><mi>L</mi></math></span>, with <span><math><mi>L</mi></math></span> scaling as <span><math><mrow><mn>1</mn><mo>+</mo><mi>O</mi><mrow><mo>(</mo><mi>N</mi><mo>/</mo><mi>p</mi><mo>)</mo></mrow></mrow></math></span> for data interpolation, and as <span><math><mrow><mn>1</mn><mo>+</mo><mi>O</mi><mfenced><mrow><msup><mrow><mi>p</mi></mrow><mrow><mo>−</mo><mn>1</mn></mrow></msup><mo>+</mo><msup><mrow><mrow><mo>(</mo><mn>1</mn><mo>+</mo><mi>p</mi><mo>)</mo></mrow></mrow><mrow><mo>−</mo><mn>1</mn></mrow></msup><msup><mrow><mi>ɛ</mi></mrow><mrow><mo>−</mo><mi>d</mi></mrow></msup></mrow></mfenced></mrow></math></span> for measures.</p><p>In the high-dimensional and wide setting where <span><math><mrow><mi>d</mi><mo>,</mo><mi>p</mi><mo>&gt;</mo><mi>N</mi></mrow></math></span>, our result can be refined to achieve <span><math><mrow><mi>L</mi><mo>=</mo><mn>0</mn></mrow></math></span>. This naturally raises the problem of data interpolation in the autonomous regime, characterized by <span><math><mrow><mi>L</mi><mo>=</mo><mn>0</mn></mrow></math></span>. We adopt two alternative approaches: either controlling in a probabilistic sense, or by relaxing the target condition. In the first case, when <span><math><mrow><mi>p</mi><mo>=</mo><mi>N</mi></mrow></math></span> we develop an inductive control strategy based on a separability assumption whose probability increases with <span><math><mi>d</mi></math></span>. In the second one, we establish an explicit error decay rate with respect to <span><math><mi>p</mi></math></span> which results from applying a universal approximation theorem to a custom-built Lipschitz vector field interpolating <span><math><mi>D</mi></math></span>.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0893608024005641/pdfft?md5=480cae19d4a2c169ff78cc6025a33eae&pid=1-s2.0-S0893608024005641-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142049204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Finite-time cluster synchronization of multi-weighted fractional-order coupled neural networks with and without impulsive effects 有脉冲效应和无脉冲效应的多加权分数阶耦合神经网络的有限时间群同步。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-17 DOI: 10.1016/j.neunet.2024.106646

In this paper, finite-time cluster synchronization (FTCS) of multi-weighted fractional-order neural networks is studied. Firstly, a FTCS criterion of the considered neural networks is obtained by designing a new delayed state feedback controller. Secondly, a FTCS criterion for the considered neural networks with mixed impulsive effects is given by constructing a new piecewise controller, where both synchronizing and desynchronizing impulses are taken into account. It should be noted that it is the first time that finite-time cluster synchronization of multi-weighted neural networks has been investigated. Finally, numerical simulations are given to show the validity of the theoretical results.

本文研究了多权分阶神经网络的有限时间群同步(FTCS)。首先,通过设计一种新的延迟状态反馈控制器,获得了所考虑的神经网络的 FTCS 准则。其次,通过构建一个新的片式控制器,同时考虑同步和非同步脉冲,给出了所考虑的具有混合脉冲效应的神经网络的 FTCS 准则。需要指出的是,这是首次对多权重神经网络的有限时间群同步进行研究。最后,还给出了数值模拟,以证明理论结果的正确性。
{"title":"Finite-time cluster synchronization of multi-weighted fractional-order coupled neural networks with and without impulsive effects","authors":"","doi":"10.1016/j.neunet.2024.106646","DOIUrl":"10.1016/j.neunet.2024.106646","url":null,"abstract":"<div><p>In this paper, finite-time cluster synchronization (FTCS) of multi-weighted fractional-order neural networks is studied. Firstly, a FTCS criterion of the considered neural networks is obtained by designing a new delayed state feedback controller. Secondly, a FTCS criterion for the considered neural networks with mixed impulsive effects is given by constructing a new piecewise controller, where both synchronizing and desynchronizing impulses are taken into account. It should be noted that it is the first time that finite-time cluster synchronization of multi-weighted neural networks has been investigated. Finally, numerical simulations are given to show the validity of the theoretical results.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037534","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Face Omron Ring: Proactive defense against face forgery with identity awareness 人脸欧姆龙戒指:通过身份识别主动防御人脸伪造。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-17 DOI: 10.1016/j.neunet.2024.106639

In the era of Artificial Intelligence Generated Content (AIGC), face forgery models pose significant security threats. These models have caused widespread negative impacts through the creation of forged products targeting public figures, national leaders, and other Persons-of-interest (POI). To address this, we propose the Face Omron Ring (FOR) to proactively protect the POI from face forgery. Specifically, by introducing FOR into a target face forgery model, the model will proactively refuse to forge any face image of protected identities without compromising the forgery capability for unprotected ones. We conduct extensive experiments on 4 face forgery models, StarGAN, AGGAN, AttGAN, and HiSD on the widely used large-scale face image datasets CelebA, CelebA-HQ, and PubFig83. Our results demonstrate that the proposed method can effectively protect 5000 different identities with a 100% protection success rate, for each of which only about 100 face images are needed. Our method also shows great robustness against multiple image processing attacks, such as JPEG, cropping, noise addition, and blurring. Compared to existing proactive defense methods, our method offers identity-centric protection for any image of the protected identity without requiring any special preprocessing, resulting in improved scalability and security. We hope that this work can provide a solution for responsible AIGC companies in regulating the use of face forgery models.

在人工智能生成内容(AIGC)时代,人脸伪造模型构成了重大安全威胁。这些模型通过制造针对公众人物、国家领导人和其他利益相关者(POI)的伪造产品,造成了广泛的负面影响。针对这一问题,我们提出了人脸欧姆龙环(FOR),以主动保护利益相关者免受人脸伪造的侵害。具体来说,通过在目标人脸伪造模型中引入 FOR,该模型将主动拒绝伪造任何受保护身份的人脸图像,而不会影响未受保护身份的伪造能力。我们在广泛使用的大规模人脸图像数据集 CelebA、CelebA-HQ 和 PubFig83 上对四种人脸伪造模型 StarGAN、AGGAN、AttGAN 和 HiSD 进行了大量实验。结果表明,所提出的方法可以有效保护 5000 种不同的身份,保护成功率达到 100%,而每种身份只需要约 100 张人脸图像。我们的方法还对多种图像处理攻击(如 JPEG、裁剪、噪声添加和模糊)表现出很强的鲁棒性。与现有的主动防御方法相比,我们的方法以身份为中心,对任何受保护身份的图像都提供保护,不需要任何特殊的预处理,从而提高了可扩展性和安全性。我们希望这项工作能为负责任的 AIGC 公司在规范人脸伪造模型的使用方面提供一种解决方案。
{"title":"Face Omron Ring: Proactive defense against face forgery with identity awareness","authors":"","doi":"10.1016/j.neunet.2024.106639","DOIUrl":"10.1016/j.neunet.2024.106639","url":null,"abstract":"<div><p>In the era of Artificial Intelligence Generated Content (AIGC), face forgery models pose significant security threats. These models have caused widespread negative impacts through the creation of forged products targeting public figures, national leaders, and other Persons-of-interest (POI). To address this, we propose the <em>Face Omron Ring</em> (FOR) to proactively protect the POI from face forgery. Specifically, by introducing FOR into a target face forgery model, the model will proactively refuse to forge any face image of protected identities without compromising the forgery capability for unprotected ones. We conduct extensive experiments on 4 face forgery models, StarGAN, AGGAN, AttGAN, and HiSD on the widely used large-scale face image datasets CelebA, CelebA-HQ, and PubFig83. Our results demonstrate that the proposed method can effectively protect 5000 different identities with a 100% protection success rate, for each of which only about 100 face images are needed. Our method also shows great robustness against multiple image processing attacks, such as JPEG, cropping, noise addition, and blurring. Compared to existing proactive defense methods, our method offers identity-centric protection for any image of the protected identity without requiring any special preprocessing, resulting in improved scalability and security. We hope that this work can provide a solution for responsible AIGC companies in regulating the use of face forgery models.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive ambiguity-aware weighting for multi-label recognition with limited annotations 利用有限注释进行多标签识别的自适应模糊感知加权法
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-16 DOI: 10.1016/j.neunet.2024.106642

In multi-label recognition, effectively addressing the challenge of partial labels is crucial for reducing annotation costs and enhancing model generalization. Existing methods exhibit limitations by relying on unrealistic simulations with uniformly dropped labels, overlooking how ambiguous instances and instance-level factors impacts label ambiguity in real-world datasets. To address this deficiency, our paper introduces a realistic partial label setting grounded in instance ambiguity, complemented by Reliable Ambiguity-Aware Instance Weighting (R-AAIW)—a strategy that utilizes importance weighting to adapt dynamically to the inherent ambiguity of multi-label instances. The strategy leverages an ambiguity score to prioritize learning from clearer instances. As proficiency of the model improves, the weights are dynamically modulated to gradually shift focus towards more ambiguous instances. By employing an adaptive re-weighting method that adjusts to the complexity of each instance, our approach not only enhances the model’s capability to detect subtle variations among labels but also ensures comprehensive learning without excluding difficult instances. Extensive experimentation across various benchmarks highlights our approach’s superiority over existing methods, showcasing its ability to provide a more accurate and adaptable framework for multi-label recognition tasks.

在多标签识别中,有效解决部分标签的挑战对于降低注释成本和提高模型泛化至关重要。现有方法的局限性在于,它们依赖于不切实际的均匀丢弃标签模拟,忽略了模糊实例和实例级因素对真实世界数据集中标签模糊性的影响。为了弥补这一不足,我们的论文引入了以实例模糊性为基础的现实部分标签设置,并辅以可靠的模糊性感知实例加权(R-AAIW)--一种利用重要性加权动态适应多标签实例固有模糊性的策略。该策略利用模糊性得分来优先学习更清晰的实例。随着模型能力的提高,权重也会动态调整,逐渐将重点转移到更模糊的实例上。通过采用适应每个实例复杂性的自适应再加权方法,我们的方法不仅增强了模型检测标签间微妙变化的能力,还确保了在不排除困难实例的情况下进行全面学习。在各种基准测试中进行的广泛实验凸显了我们的方法优于现有方法,展示了它为多标签识别任务提供更准确、适应性更强的框架的能力。
{"title":"Adaptive ambiguity-aware weighting for multi-label recognition with limited annotations","authors":"","doi":"10.1016/j.neunet.2024.106642","DOIUrl":"10.1016/j.neunet.2024.106642","url":null,"abstract":"<div><p>In multi-label recognition, effectively addressing the challenge of partial labels is crucial for reducing annotation costs and enhancing model generalization. Existing methods exhibit limitations by relying on unrealistic simulations with uniformly dropped labels, overlooking how ambiguous instances and instance-level factors impacts label ambiguity in real-world datasets. To address this deficiency, our paper introduces a realistic partial label setting grounded in instance ambiguity, complemented by Reliable Ambiguity-Aware Instance Weighting (R-AAIW)—a strategy that utilizes importance weighting to adapt dynamically to the inherent ambiguity of multi-label instances. The strategy leverages an ambiguity score to prioritize learning from clearer instances. As proficiency of the model improves, the weights are dynamically modulated to gradually shift focus towards more ambiguous instances. By employing an adaptive re-weighting method that adjusts to the complexity of each instance, our approach not only enhances the model’s capability to detect subtle variations among labels but also ensures comprehensive learning without excluding difficult instances. Extensive experimentation across various benchmarks highlights our approach’s superiority over existing methods, showcasing its ability to provide a more accurate and adaptable framework for multi-label recognition tasks.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037529","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The combined Lyapunov functionals method for stability analysis of neutral Cohen–Grossberg neural networks with multiple delays 用于多延迟中性科恩-格罗斯伯格神经网络稳定性分析的组合李亚普诺夫函数法。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-16 DOI: 10.1016/j.neunet.2024.106641

This research article will employ the combined Lyapunov functionals method to deal with stability analysis of a more general type of Cohen–Grossberg neural networks which simultaneously involve constant time and neutral delay parameters. By utilizing some combinations of various Lyapunov functionals, we determine novel criteria ensuring global stability of such a model of neural systems that employ Lipschitz continuous activation functions. These proposed results are totally stated independently of delay terms and they can be completely characterized by the constants parameters involved in the neural system. By making some detailed analytical comparisons between the stability results derived in this research article and the existing corresponding stability criteria obtained in the past literature, we prove that our proposed stability results lead to establishing some sets of stability conditions and these conditions may be evaluated as different alternative results to the previously reported corresponding stability criteria. A numerical example is also presented to show the applicability of the proposed stability results.

本研究文章将采用组合李亚普诺夫函数法,对同时涉及恒定时间和中性延迟参数的更一般类型的科恩-格罗斯伯格神经网络进行稳定性分析。通过利用各种李雅普诺夫函数的一些组合,我们确定了新的标准,以确保这种采用利普齐兹连续激活函数的神经系统模型的全局稳定性。这些提出的结果完全独立于延迟项,它们可以完全由神经系统中涉及的常数参数来表征。通过对本研究文章中得出的稳定性结果和以往文献中获得的现有相应稳定性标准进行一些详细的分析比较,我们证明,我们提出的稳定性结果导致建立了一些稳定性条件集,这些条件可作为以往报告的相应稳定性标准的不同替代结果进行评估。我们还给出了一个数值示例,以说明所提出的稳定性结果的适用性。
{"title":"The combined Lyapunov functionals method for stability analysis of neutral Cohen–Grossberg neural networks with multiple delays","authors":"","doi":"10.1016/j.neunet.2024.106641","DOIUrl":"10.1016/j.neunet.2024.106641","url":null,"abstract":"<div><p>This research article will employ the combined Lyapunov functionals method to deal with stability analysis of a more general type of Cohen–Grossberg neural networks which simultaneously involve constant time and neutral delay parameters. By utilizing some combinations of various Lyapunov functionals, we determine novel criteria ensuring global stability of such a model of neural systems that employ Lipschitz continuous activation functions. These proposed results are totally stated independently of delay terms and they can be completely characterized by the constants parameters involved in the neural system. By making some detailed analytical comparisons between the stability results derived in this research article and the existing corresponding stability criteria obtained in the past literature, we prove that our proposed stability results lead to establishing some sets of stability conditions and these conditions may be evaluated as different alternative results to the previously reported corresponding stability criteria. A numerical example is also presented to show the applicability of the proposed stability results.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Manifold-based shapley explanations for high dimensional correlated features 针对高维相关特征的基于 Manifold 的 Shapley 解释
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-14 DOI: 10.1016/j.neunet.2024.106634

Explainable artificial intelligence (XAI) holds significant importance in enhancing the reliability and transparency of network decision-making. SHapley Additive exPlanations (SHAP) is a game-theoretic approach for network interpretation, attributing confidence to inputs features to measure their importance. However, SHAP often relies on a flawed assumption that the model’s features are independent, leading to incorrect results when dealing with correlated features. In this paper, we introduce a novel manifold-based Shapley explanation method, termed Latent SHAP. Latent SHAP transforms high-dimensional data into low-dimensional manifolds to capture correlations among features. We compute Shapley values on the data manifold and devise three distinct gradient-based mapping methods to transfer them back to the high-dimensional space. Our primary objectives include: (1) correcting misinterpretations by SHAP in certain samples; (2) addressing the challenge of feature correlations in high-dimensional data interpretation; and (3) reducing algorithmic complexity through Manifold SHAP for application in complex network interpretations. Code is available at https://github.com/Teriri1999/Latent-SHAP.

可解释人工智能(XAI)在提高网络决策的可靠性和透明度方面具有重要意义。SHapley Additive exPlanations(SHAP)是一种用于网络解释的博弈论方法,它将置信度赋予输入特征以衡量其重要性。然而,SHAP 通常依赖于一个错误的假设,即模型的特征是独立的,从而导致在处理相关特征时出现错误的结果。本文介绍了一种新颖的基于流形的 Shapley 解释方法,称为 Latent SHAP。Latent SHAP 将高维数据转换为低维流形,以捕捉特征之间的相关性。我们在数据流形上计算夏普利值,并设计了三种不同的基于梯度的映射方法,将它们转回高维空间。我们的主要目标包括(1) 纠正 SHAP 在某些样本中的错误解释;(2) 解决高维数据解释中特征相关性的难题;(3) 通过 Manifold SHAP 降低算法复杂性,以应用于复杂网络解释。代码见 https://github.com/Teriri1999/Latent-SHAP。
{"title":"Manifold-based shapley explanations for high dimensional correlated features","authors":"","doi":"10.1016/j.neunet.2024.106634","DOIUrl":"10.1016/j.neunet.2024.106634","url":null,"abstract":"<div><p>Explainable artificial intelligence (XAI) holds significant importance in enhancing the reliability and transparency of network decision-making. SHapley Additive exPlanations (SHAP) is a game-theoretic approach for network interpretation, attributing confidence to inputs features to measure their importance. However, SHAP often relies on a flawed assumption that the model’s features are independent, leading to incorrect results when dealing with correlated features. In this paper, we introduce a novel manifold-based Shapley explanation method, termed Latent SHAP. Latent SHAP transforms high-dimensional data into low-dimensional manifolds to capture correlations among features. We compute Shapley values on the data manifold and devise three distinct gradient-based mapping methods to transfer them back to the high-dimensional space. Our primary objectives include: (1) correcting misinterpretations by SHAP in certain samples; (2) addressing the challenge of feature correlations in high-dimensional data interpretation; and (3) reducing algorithmic complexity through Manifold SHAP for application in complex network interpretations. Code is available at <span><span>https://github.com/Teriri1999/Latent-SHAP</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142040844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalized M-sparse algorithms for constructing fault tolerant RBF networks 构建容错 RBF 网络的广义 M-稀疏算法
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-14 DOI: 10.1016/j.neunet.2024.106633

In the construction process of radial basis function (RBF) networks, two common crucial issues arise: the selection of RBF centers and the effective utilization of the given source without encountering the overfitting problem. Another important issue is the fault tolerant capability. That is, when noise or faults exist in a trained network, it is crucial that the network’s performance does not undergo significant deterioration or decrease. However, without employing a fault tolerant procedure, a trained RBF network may exhibit significantly poor performance. Unfortunately, most existing algorithms are unable to simultaneously address all of the aforementioned issues. This paper proposes fault tolerant training algorithms that can simultaneously select RBF nodes and train RBF output weights. Additionally, our algorithms can directly control the number of RBF nodes in an explicit manner, eliminating the need for a time-consuming procedure to tune the regularization parameter and achieve the target RBF network size. Based on simulation results, our algorithms demonstrate improved test set performance when more RBF nodes are used, effectively utilizing the given source without encountering the overfitting problem. This paper first defines a fault tolerant objective function, which includes a term to suppress the effects of weight faults and weight noise. This term also prevents the issue of overfitting, resulting in better test set performance when more RBF nodes are utilized. With the defined objective function, the training process is designed to solve a generalized M-sparse problem by incorporating an 0-norm constraint. The 0-norm constraint allows us to directly and explicitly control the number of RBF nodes. To address the generalized M-sparse problem, we introduce the noise-resistant iterative hard thresholding (NR-IHT) algorithm. The convergence properties of the NR-IHT algorithm are subsequently discussed theoretically. To further enhance performance, we incorporate the momentum concept into the NR-IHT algorithm, referring to the modified version as “NR-IHT-Mom”. Simulation results show that both the NR-IHT algorithm and the NR-IHT-Mom algorithm outperform several state-of-the-art comparison algorithms.

在径向基函数(RBF)网络的构建过程中,会出现两个常见的关键问题:RBF 中心的选择和有效利用给定源而不出现过拟合问题。另一个重要问题是容错能力。也就是说,当训练有素的网络中存在噪声或故障时,网络的性能不会出现明显的恶化或下降是至关重要的。然而,如果不采用容错程序,经过训练的 RBF 网络可能会表现出很差的性能。遗憾的是,大多数现有算法无法同时解决上述所有问题。本文提出的容错训练算法可以同时选择 RBF 节点和训练 RBF 输出权重。此外,我们的算法还能以明确的方式直接控制 RBF 节点的数量,从而省去了调整正则化参数和实现目标 RBF 网络大小的耗时过程。根据仿真结果,当使用更多的 RBF 节点时,我们的算法证明了测试集性能的提高,有效地利用了给定的源,而不会遇到过拟合问题。本文首先定义了一个容错目标函数,其中包含一个抑制权重故障和权重噪声影响的项。这个项还能防止过拟合问题,从而在使用更多 RBF 节点时获得更好的测试集性能。有了确定的目标函数,训练过程就可以通过加入 ℓ0-norm 约束来解决广义 M-稀疏问题。ℓ0-norm 约束条件允许我们直接、明确地控制 RBF 节点的数量。为了解决广义 M 稀疏问题,我们引入了抗噪迭代硬阈值算法(NR-IHT)。随后,我们从理论上讨论了 NR-IHT 算法的收敛特性。为了进一步提高性能,我们在 NR-IHT 算法中加入了动量概念,并将修改后的版本称为 "NR-IHT-Mom"。仿真结果表明,NR-IHT 算法和 NR-IHT-Mom 算法都优于几种最先进的比较算法。
{"title":"Generalized M-sparse algorithms for constructing fault tolerant RBF networks","authors":"","doi":"10.1016/j.neunet.2024.106633","DOIUrl":"10.1016/j.neunet.2024.106633","url":null,"abstract":"<div><p>In the construction process of radial basis function (RBF) networks, two common crucial issues arise: the selection of RBF centers and the effective utilization of the given source without encountering the overfitting problem. Another important issue is the fault tolerant capability. That is, when noise or faults exist in a trained network, it is crucial that the network’s performance does not undergo significant deterioration or decrease. However, without employing a fault tolerant procedure, a trained RBF network may exhibit significantly poor performance. Unfortunately, most existing algorithms are unable to simultaneously address all of the aforementioned issues. This paper proposes fault tolerant training algorithms that can simultaneously select RBF nodes and train RBF output weights. Additionally, our algorithms can directly control the number of RBF nodes in an explicit manner, eliminating the need for a time-consuming procedure to tune the regularization parameter and achieve the target RBF network size. Based on simulation results, our algorithms demonstrate improved test set performance when more RBF nodes are used, effectively utilizing the given source without encountering the overfitting problem. This paper first defines a fault tolerant objective function, which includes a term to suppress the effects of weight faults and weight noise. This term also prevents the issue of overfitting, resulting in better test set performance when more RBF nodes are utilized. With the defined objective function, the training process is designed to solve a generalized <span><math><mi>M</mi></math></span>-sparse problem by incorporating an <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>0</mn></mrow></msub></math></span>-norm constraint. The <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>0</mn></mrow></msub></math></span>-norm constraint allows us to directly and explicitly control the number of RBF nodes. To address the generalized <span><math><mi>M</mi></math></span>-sparse problem, we introduce the noise-resistant iterative hard thresholding (NR-IHT) algorithm. The convergence properties of the NR-IHT algorithm are subsequently discussed theoretically. To further enhance performance, we incorporate the momentum concept into the NR-IHT algorithm, referring to the modified version as “NR-IHT-Mom”. Simulation results show that both the NR-IHT algorithm and the NR-IHT-Mom algorithm outperform several state-of-the-art comparison algorithms.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142087783","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1