首页 > 最新文献

Neural Networks最新文献

英文 中文
Distributed leader-following bipartite consensus for one-sided Lipschitz multi-agent systems via dual-terminal event-triggered mechanism 通过双终端事件触发机制实现单边利普斯奇茨多代理系统的分布式领导者-追随者双方共识。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-19 DOI: 10.1016/j.neunet.2024.106808
Yanjun Zhao , Haibin Sun , Xiangyu Wang , Dong Yang , Ticao Jiao
This article analyses leader-following bipartite consensus for one-sided Lipschitz multi-agent systems by dual-terminal event-triggered output feedback control approach. A distributed observer is designed to estimate unknown system states by employing relative output information at triggering time instants, and then an event-triggered output feedback controller is proposed. Dual-terminal dynamic event-triggered mechanisms are proposed in sensor–observer channel and controller–actuator channel, which can save communication resources to a great extent, and the Zeno behavior is ruled out. A new generalized one-sided Lipschitz condition is proposed to handle the nonlinear term and achieve bipartite consensus. Some stability conditions are presented to guarantee leader-following bipartite consensus. Finally, one-link robot manipulator systems are introduced to demonstrate the availability of the designed scheme. The results demonstrate that the agents of the robot manipulators can track the reference trajectories bi-directionally, and effectively reduce communication resources by 61.22% and 68.04% at the sensor–observer and controller–actuator channels, respectively.
本文通过双终端事件触发输出反馈控制方法,分析了单边 Lipschitz 多代理系统的领导者-追随双方共识。文章设计了一个分布式观测器,利用触发时点的相对输出信息来估计未知系统状态,然后提出了一个事件触发输出反馈控制器。在传感器-观测器信道和控制器-执行器信道中提出了双端动态事件触发机制,这在很大程度上节省了通信资源,并排除了芝诺行为。提出了一种新的广义单边 Lipschitz 条件来处理非线性项并实现两方共识。还提出了一些稳定性条件,以保证领导者-跟随者的两方共识。最后,介绍了单链机器人操纵器系统,以证明所设计方案的可用性。结果表明,机器人操纵器的代理可以双向跟踪参考轨迹,并在传感器-观察者和控制器-执行器通道上分别有效减少了 61.22% 和 68.04% 的通信资源。
{"title":"Distributed leader-following bipartite consensus for one-sided Lipschitz multi-agent systems via dual-terminal event-triggered mechanism","authors":"Yanjun Zhao ,&nbsp;Haibin Sun ,&nbsp;Xiangyu Wang ,&nbsp;Dong Yang ,&nbsp;Ticao Jiao","doi":"10.1016/j.neunet.2024.106808","DOIUrl":"10.1016/j.neunet.2024.106808","url":null,"abstract":"<div><div>This article analyses leader-following bipartite consensus for one-sided Lipschitz multi-agent systems by dual-terminal event-triggered output feedback control approach. A distributed observer is designed to estimate unknown system states by employing relative output information at triggering time instants, and then an event-triggered output feedback controller is proposed. Dual-terminal dynamic event-triggered mechanisms are proposed in sensor–observer channel and controller–actuator channel, which can save communication resources to a great extent, and the Zeno behavior is ruled out. A new generalized one-sided Lipschitz condition is proposed to handle the nonlinear term and achieve bipartite consensus. Some stability conditions are presented to guarantee leader-following bipartite consensus. Finally, one-link robot manipulator systems are introduced to demonstrate the availability of the designed scheme. The results demonstrate that the agents of the robot manipulators can track the reference trajectories bi-directionally, and effectively reduce communication resources by 61.22% and 68.04% at the sensor–observer and controller–actuator channels, respectively.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106808"},"PeriodicalIF":6.0,"publicationDate":"2024-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142511781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hypergraph contrastive attention networks for hyperedge prediction with negative samples evaluation 利用负样本评估超图对比注意力网络进行超edge 预测。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-19 DOI: 10.1016/j.neunet.2024.106807
Junbo Wang , Jianrui Chen , Zhihui Wang , Maoguo Gong
Hyperedge prediction aims to predict common relations among multiple nodes that will occur in the future or remain undiscovered in the current hypergraph. It is traditionally modeled as a classification task, which performs hypergraph feature learning and classifies the target samples as either present or absent. However, these approaches involve two issues: (i) in hyperedge feature learning, they fail to measure the influence of nodes on the hyperedges that include them and the neighboring hyperedges, and (ii) in the binary classification task, the quality of the generated negative samples directly impacts the prediction results. To this end, we propose a Hypergraph Contrastive Attention Network (HCAN) model for hyperedge prediction. Inspired by the brain organization, HCAN considers the influence of hyperedges with different orders through the order propagation attention mechanism. It also utilizes the contrastive mechanism to measure the reliability of attention effectively. Furthermore, we design a negative sample generator to produce three different types of negative samples. We evaluate the impact of various negative samples on the model and analyze the problems of binary classification modeling. The effectiveness of HCAN in hyperedge prediction is validated by experimentally comparing 12 baselines on 9 datasets. Our implementations will be publicly available at https://github.com/jianruichen/HCAN.
超图预测的目的是预测多个节点之间的共同关系,这些关系将在未来出现或在当前超图中仍未被发现。传统上,它被模拟为一种分类任务,执行超图特征学习,并将目标样本分类为存在或不存在。然而,这些方法涉及两个问题:(i) 在超图特征学习中,它们无法衡量节点对包含它们的超图和相邻超图的影响;(ii) 在二元分类任务中,生成的负样本的质量直接影响预测结果。为此,我们提出了一种用于超边缘预测的超图对比注意网络(HCAN)模型。HCAN 受到大脑组织的启发,通过顺序传播注意机制考虑了不同顺序的超边缘的影响。它还利用对比机制来有效衡量注意力的可靠性。此外,我们还设计了一个负样本生成器,以生成三种不同类型的负样本。我们评估了各种负样本对模型的影响,并分析了二元分类建模的问题。通过在 9 个数据集上对 12 个基线进行实验比较,验证了 HCAN 在超边缘预测中的有效性。我们的实现将在 https://github.com/jianruichen/HCAN 上公开。
{"title":"Hypergraph contrastive attention networks for hyperedge prediction with negative samples evaluation","authors":"Junbo Wang ,&nbsp;Jianrui Chen ,&nbsp;Zhihui Wang ,&nbsp;Maoguo Gong","doi":"10.1016/j.neunet.2024.106807","DOIUrl":"10.1016/j.neunet.2024.106807","url":null,"abstract":"<div><div>Hyperedge prediction aims to predict common relations among multiple nodes that will occur in the future or remain undiscovered in the current hypergraph. It is traditionally modeled as a classification task, which performs hypergraph feature learning and classifies the target samples as either present or absent. However, these approaches involve two issues: (i) in hyperedge feature learning, they fail to measure the influence of nodes on the hyperedges that include them and the neighboring hyperedges, and (ii) in the binary classification task, the quality of the generated negative samples directly impacts the prediction results. To this end, we propose a Hypergraph Contrastive Attention Network (HCAN) model for hyperedge prediction. Inspired by the brain organization, HCAN considers the influence of hyperedges with different orders through the order propagation attention mechanism. It also utilizes the contrastive mechanism to measure the reliability of attention effectively. Furthermore, we design a negative sample generator to produce three different types of negative samples. We evaluate the impact of various negative samples on the model and analyze the problems of binary classification modeling. The effectiveness of HCAN in hyperedge prediction is validated by experimentally comparing 12 baselines on 9 datasets. Our implementations will be publicly available at <span><span>https://github.com/jianruichen/HCAN</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106807"},"PeriodicalIF":6.0,"publicationDate":"2024-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142511785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cut-and-Paste: Subject-driven video editing with attention control 剪切粘贴主题驱动的视频编辑与注意力控制
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-19 DOI: 10.1016/j.neunet.2024.106818
Zhichao Zuo , Zhao Zhang , Yan Luo , Yang Zhao , Haijun Zhang , Yi Yang , Meng Wang
This paper presents a novel framework termed Cut-and-Paste for real-word semantic video editing under the guidance of text prompt and additional reference image. While the text-driven video editing has demonstrated remarkable ability to generate highly diverse videos following given text prompts, the fine-grained semantic edits are hard to control by plain textual prompt only in terms of object details and edited region, and cumbersome long text descriptions are usually needed for the task. We therefore investigate subject-driven video editing for more precise control of both edited regions and background preservation, and fine-grained semantic generation. We achieve this goal by introducing an reference image as supplementary input to the text-driven video editing, which avoids racking your brain to come up with a cumbersome text prompt describing the detailed appearance of the object. To limit the editing area, we refer to a method of cross attention control in image editing and successfully extend it to video editing by fusing the attention map of adjacent frames, which strikes a balance between maintaining video background and spatio-temporal consistency. Compared with current methods, the whole process of our method is like “cut” the source object to be edited and then “paste” the target object provided by reference image. We demonstrate that our method performs favorably over prior arts for video editing under the guidance of text prompt and extra reference image, as measured by both quantitative and subjective evaluations.
本文提出了一个新颖的框架,称为 "剪贴"(Cut-and-Paste),用于在文本提示和附加参考图像的指导下进行实词语义视频编辑。虽然文本驱动的视频编辑已经证明了根据给定的文本提示生成高度多样化视频的卓越能力,但仅靠纯文本提示很难控制对象细节和编辑区域的细粒度语义编辑,而且通常需要繁琐的长文本描述来完成任务。因此,我们研究了主题驱动视频编辑,以便更精确地控制编辑区域和背景保存,并生成细粒度语义。为了实现这一目标,我们引入了参考图像作为文本驱动视频编辑的辅助输入,这样就可以避免绞尽脑汁想出繁琐的文本提示来描述对象的详细外观。为了限制编辑区域,我们参考了图像编辑中的交叉注意力控制方法,并通过融合相邻帧的注意力图谱成功地将其扩展到视频编辑中,从而在保持视频背景和时空一致性之间取得了平衡。与现有方法相比,我们的方法的整个过程就像 "剪切 "待编辑的源对象,然后 "粘贴 "参考图像提供的目标对象。通过定量和主观评价,我们证明了在文本提示和额外参考图像的指导下,我们的方法在视频编辑方面的表现优于现有技术。
{"title":"Cut-and-Paste: Subject-driven video editing with attention control","authors":"Zhichao Zuo ,&nbsp;Zhao Zhang ,&nbsp;Yan Luo ,&nbsp;Yang Zhao ,&nbsp;Haijun Zhang ,&nbsp;Yi Yang ,&nbsp;Meng Wang","doi":"10.1016/j.neunet.2024.106818","DOIUrl":"10.1016/j.neunet.2024.106818","url":null,"abstract":"<div><div>This paper presents a novel framework termed Cut-and-Paste for real-word semantic video editing under the guidance of text prompt and additional reference image. While the text-driven video editing has demonstrated remarkable ability to generate highly diverse videos following given text prompts, the fine-grained semantic edits are hard to control by plain textual prompt only in terms of object details and edited region, and cumbersome long text descriptions are usually needed for the task. We therefore investigate subject-driven video editing for more precise control of both edited regions and background preservation, and fine-grained semantic generation. We achieve this goal by introducing an reference image as supplementary input to the text-driven video editing, which avoids racking your brain to come up with a cumbersome text prompt describing the detailed appearance of the object. To limit the editing area, we refer to a method of cross attention control in image editing and successfully extend it to video editing by fusing the attention map of adjacent frames, which strikes a balance between maintaining video background and spatio-temporal consistency. Compared with current methods, the whole process of our method is like “cut” the source object to be edited and then “paste” the target object provided by reference image. We demonstrate that our method performs favorably over prior arts for video editing under the guidance of text prompt and extra reference image, as measured by both quantitative and subjective evaluations.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106818"},"PeriodicalIF":6.0,"publicationDate":"2024-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142540159","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quality-related fault detection for dynamic process based on quality-driven long short-term memory network and autoencoder 基于质量驱动的长短期记忆网络和自动编码器的动态过程质量相关故障检测。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-19 DOI: 10.1016/j.neunet.2024.106819
Yishun Liu , Keke Huang , Benedict Jun Ma , Ke Wei , Yuxuan Li , Chunhua Yang , Weihua Gui
Fault detection consistently plays a crucial role in industrial dynamic processes as it enables timely prevention of production losses. However, since industrial dynamic processes become increasingly coupled and complex, they introduce uneven dynamics within the collected data, posing significant challenges in effectively extracting dynamic features. In addition, it is a tricky business to distinguish whether the fault that occurs is quality-related or not, resulting in unnecessary repairing and large losses. In order to deal with these issues, this paper comes up with a novel fault detection method based on quality-driven long short-term memory and autoencoder (QLSTM-AE). Specifically, an LSTM network is initially employed to extract dynamic features, while quality variables are simultaneously incorporated in parallel to capture quality-related features. Then, a fault detection strategy based on reconstruction error statistic squared prediction error (SPE) and the quality monitoring statistic Hotelling T2 (H2) is designed, which can distinguish various types of faults to realize accurate monitoring for dynamic processes. Finally, several experiments conducted on numerical simulations and the Tennessee Eastman (TE) benchmark process demonstrate the reliability and effectiveness of the proposed QLSTM-AE method, which indicates it has higher accuracy and can separate different faults efficiently compared to some state-of-the-art methods.
故障检测在工业动态流程中一直发挥着至关重要的作用,因为它能及时防止生产损失。然而,由于工业动态流程的耦合性和复杂性越来越高,它们会在采集的数据中引入不均衡的动态,这给有效提取动态特征带来了巨大挑战。此外,如何区分发生的故障是否与质量有关也是一个棘手的问题,从而导致不必要的维修和巨大的损失。为了解决这些问题,本文提出了一种基于质量驱动长短期记忆和自动编码器(QLSTM-AE)的新型故障检测方法。具体来说,首先采用 LSTM 网络来提取动态特征,同时并行纳入质量变量来捕捉与质量相关的特征。然后,设计了一种基于重构误差统计量平方预测误差(SPE)和质量监测统计量 Hotelling T2(H2)的故障检测策略,该策略可以区分各种类型的故障,从而实现对动态过程的精确监测。最后,在数值模拟和田纳西伊士曼(Tennessee Eastman,TE)基准工艺上进行的几项实验证明了所提出的 QLSTM-AE 方法的可靠性和有效性,表明与一些最先进的方法相比,该方法具有更高的准确性,并能有效区分不同的故障。
{"title":"Quality-related fault detection for dynamic process based on quality-driven long short-term memory network and autoencoder","authors":"Yishun Liu ,&nbsp;Keke Huang ,&nbsp;Benedict Jun Ma ,&nbsp;Ke Wei ,&nbsp;Yuxuan Li ,&nbsp;Chunhua Yang ,&nbsp;Weihua Gui","doi":"10.1016/j.neunet.2024.106819","DOIUrl":"10.1016/j.neunet.2024.106819","url":null,"abstract":"<div><div>Fault detection consistently plays a crucial role in industrial dynamic processes as it enables timely prevention of production losses. However, since industrial dynamic processes become increasingly coupled and complex, they introduce uneven dynamics within the collected data, posing significant challenges in effectively extracting dynamic features. In addition, it is a tricky business to distinguish whether the fault that occurs is quality-related or not, resulting in unnecessary repairing and large losses. In order to deal with these issues, this paper comes up with a novel fault detection method based on quality-driven long short-term memory and autoencoder (QLSTM-AE). Specifically, an LSTM network is initially employed to extract dynamic features, while quality variables are simultaneously incorporated in parallel to capture quality-related features. Then, a fault detection strategy based on reconstruction error statistic squared prediction error (<span><math><mrow><mi>S</mi><mi>P</mi><mi>E</mi></mrow></math></span>) and the quality monitoring statistic Hotelling <span><math><msup><mrow><mi>T</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> (<span><math><msup><mrow><mi>H</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span>) is designed, which can distinguish various types of faults to realize accurate monitoring for dynamic processes. Finally, several experiments conducted on numerical simulations and the Tennessee Eastman (TE) benchmark process demonstrate the reliability and effectiveness of the proposed QLSTM-AE method, which indicates it has higher accuracy and can separate different faults efficiently compared to some state-of-the-art methods.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106819"},"PeriodicalIF":6.0,"publicationDate":"2024-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142511789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fractional-order stochastic gradient descent method with momentum and energy for deep neural networks 用于深度神经网络的带动量和能量的分数阶随机梯度下降法
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-19 DOI: 10.1016/j.neunet.2024.106810
Xingwen Zhou , Zhenghao You , Weiguo Sun , Dongdong Zhao , Shi Yan
In this paper, a novel fractional-order stochastic gradient descent with momentum and energy (FOSGDME) approach is proposed. Specifically, to address the challenge of converging to a real extreme point encountered by the existing fractional gradient algorithms, a novel fractional-order stochastic gradient descent (FOSGD) method is presented by modifying the definition of the Caputo fractional-order derivative. A FOSGD with moment (FOSGDM) is established by incorporating momentum information to accelerate the convergence speed and accuracy further. In addition, to improve the robustness and accuracy, a FOSGD with moment and energy is established by further introducing energy formation. The extensive experimental results on the image classification CIFAR-10 dataset obtained with ResNet and DenseNet demonstrate that the proposed FOSGD, FOSGDM and FOSGDME algorithms are superior to the integer order optimization algorithms, and achieve state-of-the-art performance.
本文提出了一种新颖的带动量和能量的分数阶随机梯度下降(FOSGDME)方法。具体来说,为了解决现有分数梯度算法遇到的收敛到真实极值点的难题,本文通过修改 Caputo 分数阶导数的定义,提出了一种新型分数阶随机梯度下降(FOSGD)方法。通过结合动量信息,建立了带矩量的 FOSGD(FOSGDM),进一步加快了收敛速度和精度。此外,为了提高鲁棒性和准确性,还进一步引入了能量形成,建立了带矩和能量的 FOSGD。利用 ResNet 和 DenseNet 在图像分类 CIFAR-10 数据集上获得的大量实验结果表明,所提出的 FOSGD、FOSGDM 和 FOSGDME 算法优于整数阶优化算法,达到了最先进的性能。
{"title":"Fractional-order stochastic gradient descent method with momentum and energy for deep neural networks","authors":"Xingwen Zhou ,&nbsp;Zhenghao You ,&nbsp;Weiguo Sun ,&nbsp;Dongdong Zhao ,&nbsp;Shi Yan","doi":"10.1016/j.neunet.2024.106810","DOIUrl":"10.1016/j.neunet.2024.106810","url":null,"abstract":"<div><div>In this paper, a novel fractional-order stochastic gradient descent with momentum and energy (FOSGDME) approach is proposed. Specifically, to address the challenge of converging to a real extreme point encountered by the existing fractional gradient algorithms, a novel fractional-order stochastic gradient descent (FOSGD) method is presented by modifying the definition of the Caputo fractional-order derivative. A FOSGD with moment (FOSGDM) is established by incorporating momentum information to accelerate the convergence speed and accuracy further. In addition, to improve the robustness and accuracy, a FOSGD with moment and energy is established by further introducing energy formation. The extensive experimental results on the image classification CIFAR-10 dataset obtained with ResNet and DenseNet demonstrate that the proposed FOSGD, FOSGDM and FOSGDME algorithms are superior to the integer order optimization algorithms, and achieve state-of-the-art performance.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106810"},"PeriodicalIF":6.0,"publicationDate":"2024-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142511784","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Reducing semantic ambiguity in domain adaptive semantic segmentation via probabilistic prototypical pixel contrast 通过概率原型像素对比,减少领域自适应语义分割中的语义模糊性。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1016/j.neunet.2024.106806
Xiaoke Hao, Shiyu Liu, Chuanbo Feng, Ye Zhu
Domain adaptation aims to reduce the model degradation on the target domain caused by the domain shift between the source and target domains. Although encouraging performance has been achieved by combining contrastive learning with the self-training paradigm, they suffer from ambiguous scenarios caused by scale, illumination, or overlapping when deploying deterministic embedding. To address these issues, we propose probabilistic prototypical pixel contrast (PPPC), a universal adaptation framework that models each pixel embedding as a probability via multivariate Gaussian distribution to fully exploit the uncertainty within them, eventually improving the representation quality of the model. In addition, we derive prototypes from probability estimation posterior probability estimation which helps to push the decision boundary away from the ambiguity points. Moreover, we employ an efficient method to compute similarity between distributions, eliminating the need for sampling and reparameterization, thereby significantly reducing computational overhead. Further, we dynamically select the ambiguous crops at the image level to enlarge the number of boundary points involved in contrastive learning, which benefits the establishment of precise distributions for each category. Extensive experimentation demonstrates that PPPC not only helps to address ambiguity at the pixel level, yielding discriminative representations but also achieves significant improvements in both synthetic-to-real and day-to-night adaptation tasks. It surpasses the previous state-of-the-art (SOTA) by +5.2% mIoU in the most challenging daytime-to-nighttime adaptation scenario, exhibiting stronger generalization on other unseen datasets. The code and models are available at https://github.com/DarlingInTheSV/Probabilistic-Prototypical-Pixel-Contrast.
域适应旨在减少源域和目标域之间的域转移造成的目标域模型退化。虽然通过将对比学习与自我训练范式相结合,已经取得了令人鼓舞的性能,但在部署确定性嵌入时,它们会受到尺度、光照或重叠造成的模糊场景的影响。为了解决这些问题,我们提出了概率原型像素对比度(PPPC),这是一种通用的适应框架,通过多元高斯分布将每个像素嵌入建模为一种概率,以充分利用其中的不确定性,最终提高模型的表示质量。此外,我们还从概率估计的后验概率估计中推导出原型,这有助于将决策边界推离模糊点。此外,我们还采用了一种高效的方法来计算分布之间的相似性,无需进行采样和重新参数化,从而大大减少了计算开销。此外,我们在图像层面动态选择模糊作物,以扩大对比学习中涉及的边界点数量,这有利于为每个类别建立精确的分布。广泛的实验证明,PPPC 不仅有助于解决像素级的模糊性问题,产生具有区分性的表征,而且在合成到真实和白天到黑夜的适应任务中都取得了显著的改进。在最具挑战性的白天到黑夜的适应场景中,它的 mIoU 超过了之前的最先进水平(SOTA)+5.2%,并在其他未见数据集上表现出更强的泛化能力。代码和模型可在 https://github.com/DarlingInTheSV/Probabilistic-Prototypical-Pixel-Contrast 上获取。
{"title":"Reducing semantic ambiguity in domain adaptive semantic segmentation via probabilistic prototypical pixel contrast","authors":"Xiaoke Hao,&nbsp;Shiyu Liu,&nbsp;Chuanbo Feng,&nbsp;Ye Zhu","doi":"10.1016/j.neunet.2024.106806","DOIUrl":"10.1016/j.neunet.2024.106806","url":null,"abstract":"<div><div>Domain adaptation aims to reduce the model degradation on the target domain caused by the domain shift between the source and target domains. Although encouraging performance has been achieved by combining contrastive learning with the self-training paradigm, they suffer from ambiguous scenarios caused by scale, illumination, or overlapping when deploying deterministic embedding. To address these issues, we propose probabilistic prototypical pixel contrast (PPPC), a universal adaptation framework that models each pixel embedding as a probability via multivariate Gaussian distribution to fully exploit the uncertainty within them, eventually improving the representation quality of the model. In addition, we derive prototypes from probability estimation posterior probability estimation which helps to push the decision boundary away from the ambiguity points. Moreover, we employ an efficient method to compute similarity between distributions, eliminating the need for sampling and reparameterization, thereby significantly reducing computational overhead. Further, we dynamically select the ambiguous crops at the image level to enlarge the number of boundary points involved in contrastive learning, which benefits the establishment of precise distributions for each category. Extensive experimentation demonstrates that PPPC not only helps to address ambiguity at the pixel level, yielding discriminative representations but also achieves significant improvements in both synthetic-to-real and day-to-night adaptation tasks. It surpasses the previous state-of-the-art (SOTA) by +5.2% mIoU in the most challenging daytime-to-nighttime adaptation scenario, exhibiting stronger generalization on other unseen datasets. The code and models are available at <span><span>https://github.com/DarlingInTheSV/Probabilistic-Prototypical-Pixel-Contrast</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106806"},"PeriodicalIF":6.0,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142511790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GO-MAE: Self-supervised pre-training via masked autoencoder for OCT image classification of gynecology GO-MAE:通过掩码自动编码器进行自监督预训练,用于妇科 OCT 图像分类
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1016/j.neunet.2024.106817
Haoran Wang, Xinyu Guo, Kaiwen Song, Mingyang Sun, Yanbin Shao, Songfeng Xue, Hongwei Zhang, Tianyu Zhang
Genitourinary syndrome of menopause (GSM) is a physiological disorder caused by reduced levels of oestrogen in menopausal women. Gradually, its symptoms worsen with age and prolonged menopausal status, which gravely impacts the quality of life as well as the physical and mental health of the patients. In this regard, optical coherence tomography (OCT) system effectively reduces the patient’s burden in clinical diagnosis with its noncontact, noninvasive tomographic imaging process. Consequently, supervised computer vision models applied on OCT images have yielded excellent results for disease diagnosis. However, manual labeling on an extensive number of medical images is expensive and time-consuming. To this end, this paper proposes GO-MAE, a pretraining framework for self-supervised learning of GSM OCT images based on Masked Autoencoder (MAE). To the best of our knowledge, this is the first study that applies self-supervised learning methods on the field of GSM disease screening. Focusing on the semantic complexity and feature sparsity of GSM OCT images, the objective of this study is two-pronged: first, a dynamic masking strategy is introduced for OCT characteristics in downstream tasks. This method can reduce the interference of invalid features on the model and shorten the training time. In the encoder design of MAE, we propose a convolutional neural network and transformer parallel network architecture (C&T), which aims to fuse the local and global representations of the relevant lesions in an interactive manner such that the model can still learn the richer differences between the feature information without labels. Thereafter, a series of experimental results on the acquired GSM-OCT dataset revealed that GO-MAE yields significant improvements over existing state-of-the-art techniques. Furthermore, the superiority of the model in terms of robustness and interpretability was verified through a series of comparative experiments and visualization operations, which consequently demonstrated its great potential for screening GSM symptoms.
更年期泌尿生殖系统综合征(GSM)是由更年期妇女体内雌激素水平降低引起的一种生理紊乱。随着年龄的增长和绝经期的延长,其症状会逐渐加重,严重影响患者的生活质量和身心健康。在这方面,光学相干断层扫描(OCT)系统以其非接触、非侵入性的断层成像过程,有效地减轻了患者的临床诊断负担。因此,应用于 OCT 图像的计算机视觉监督模型在疾病诊断方面取得了卓越的成果。然而,对大量医学图像进行人工标注既昂贵又耗时。为此,本文提出了 GO-MAE,一种基于掩码自动编码器(MAE)的 GSM OCT 图像自监督学习预训练框架。据我们所知,这是第一项将自监督学习方法应用于 GSM 疾病筛查领域的研究。针对 GSM OCT 图像的语义复杂性和特征稀疏性,本研究的目标是双管齐下的:首先,针对下游任务中的 OCT 特征引入动态掩蔽策略。这种方法可以减少无效特征对模型的干扰,缩短训练时间。在 MAE 的编码器设计中,我们提出了卷积神经网络和变压器并行网络架构(C&T),旨在以交互的方式融合相关病变的局部和全局表征,使模型在没有标签的情况下仍能学习到更丰富的差异特征信息。此后,在获取的 GSM-OCT 数据集上进行的一系列实验结果表明,GO-MAE 比现有的最先进技术有显著改进。此外,通过一系列对比实验和可视化操作,验证了该模型在鲁棒性和可解释性方面的优越性,从而证明了其在筛查 GSM 症状方面的巨大潜力。
{"title":"GO-MAE: Self-supervised pre-training via masked autoencoder for OCT image classification of gynecology","authors":"Haoran Wang,&nbsp;Xinyu Guo,&nbsp;Kaiwen Song,&nbsp;Mingyang Sun,&nbsp;Yanbin Shao,&nbsp;Songfeng Xue,&nbsp;Hongwei Zhang,&nbsp;Tianyu Zhang","doi":"10.1016/j.neunet.2024.106817","DOIUrl":"10.1016/j.neunet.2024.106817","url":null,"abstract":"<div><div>Genitourinary syndrome of menopause (GSM) is a physiological disorder caused by reduced levels of oestrogen in menopausal women. Gradually, its symptoms worsen with age and prolonged menopausal status, which gravely impacts the quality of life as well as the physical and mental health of the patients. In this regard, optical coherence tomography (OCT) system effectively reduces the patient’s burden in clinical diagnosis with its noncontact, noninvasive tomographic imaging process. Consequently, supervised computer vision models applied on OCT images have yielded excellent results for disease diagnosis. However, manual labeling on an extensive number of medical images is expensive and time-consuming. To this end, this paper proposes GO-MAE, a pretraining framework for self-supervised learning of GSM OCT images based on Masked Autoencoder (MAE). To the best of our knowledge, this is the first study that applies self-supervised learning methods on the field of GSM disease screening. Focusing on the semantic complexity and feature sparsity of GSM OCT images, the objective of this study is two-pronged: first, a dynamic masking strategy is introduced for OCT characteristics in downstream tasks. This method can reduce the interference of invalid features on the model and shorten the training time. In the encoder design of MAE, we propose a convolutional neural network and transformer parallel network architecture (C&amp;T), which aims to fuse the local and global representations of the relevant lesions in an interactive manner such that the model can still learn the richer differences between the feature information without labels. Thereafter, a series of experimental results on the acquired GSM-OCT dataset revealed that GO-MAE yields significant improvements over existing state-of-the-art techniques. Furthermore, the superiority of the model in terms of robustness and interpretability was verified through a series of comparative experiments and visualization operations, which consequently demonstrated its great potential for screening GSM symptoms.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106817"},"PeriodicalIF":6.0,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142578522","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deployable mixed-precision quantization with co-learning and one-time search 采用共同学习和一次性搜索的可部署混合精度量化技术
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1016/j.neunet.2024.106812
Shiguang Wang , Zhongyu Zhang , Guo Ai , Jian Cheng
Mixed-precision quantization plays a pivotal role in deploying deep neural networks in resource-constrained environments. However, the task of finding the optimal bit-width configurations for different layers under deployable mixed-precision quantization has barely been explored and remains a challenge. In this work, we present Cobits, an efficient and effective deployable mixed-precision quantization framework based on the relationship between the range of real-valued input and the range of quantized real-valued. It assigns a higher bit-width to the quantizer with a narrower quantized real-valued range and a lower bit-width to the quantizer with a wider quantized real-valued range. Cobits employs a co-learning approach to entangle and learn quantization parameters across various bit-widths, distinguishing between shared and specific parts. The shared part collaborates, while the specific part isolates precision conflicts. Additionally, we upgrade the normal quantizer to dynamic quantizer to mitigate statistical issues in the deployable mixed-precision supernet. Over the trained mixed-precision supernet, we utilize the quantized real-valued ranges to derive quantized-bit-sensitivity, which can serve as importance indicators for efficiently determining bit-width configurations, eliminating the need for iterative validation dataset evaluations. Extensive experiments show that Cobits outperforms previous state-of-the-art quantization methods on the ImageNet and COCO datasets while retaining superior efficiency. We show this approach dynamically adapts to varying bit-width and can generalize to various deployable backends. The code will be made public in https://github.com/sunnyxiaohu/cobits.
在资源受限的环境中部署深度神经网络时,混合精度量化起着举足轻重的作用。然而,如何在可部署混合精度量化条件下为不同层找到最佳位宽配置,这一任务几乎没有被探索过,仍然是一个挑战。在这项工作中,我们基于实值输入范围和量化实值范围之间的关系,提出了一种高效、有效的可部署混合精度量化框架 Cobits。它为量化实值范围较窄的量化器分配较高的位宽,为量化实值范围较宽的量化器分配较低的位宽。Cobits 采用共同学习的方法来纠缠和学习不同位宽的量化参数,并区分共享部分和特定部分。共享部分进行协作,而特定部分则隔离精度冲突。此外,我们还将普通量化器升级为动态量化器,以缓解可部署混合精度超级网络中的统计问题。在训练有素的混合精度超网上,我们利用量化的实值范围推导出量化比特灵敏度,它可以作为有效确定比特宽度配置的重要性指标,从而消除了迭代验证数据集评估的需要。大量实验表明,Cobits 在 ImageNet 和 COCO 数据集上的表现优于之前最先进的量化方法,同时保持了卓越的效率。我们表明,这种方法能动态适应不同的位宽,并能推广到各种可部署的后端。代码将在 https://github.com/sunnyxiaohu/cobits 上公开。
{"title":"Deployable mixed-precision quantization with co-learning and one-time search","authors":"Shiguang Wang ,&nbsp;Zhongyu Zhang ,&nbsp;Guo Ai ,&nbsp;Jian Cheng","doi":"10.1016/j.neunet.2024.106812","DOIUrl":"10.1016/j.neunet.2024.106812","url":null,"abstract":"<div><div>Mixed-precision quantization plays a pivotal role in deploying deep neural networks in resource-constrained environments. However, the task of finding the optimal bit-width configurations for different layers under <strong>deployable mixed-precision quantization</strong> has barely been explored and remains a challenge. In this work, we present Cobits, an efficient and effective deployable mixed-precision quantization framework based on the relationship between the range of real-valued input and the range of quantized real-valued. It assigns a higher bit-width to the quantizer with a narrower quantized real-valued range and a lower bit-width to the quantizer with a wider quantized real-valued range. Cobits employs a co-learning approach to entangle and learn quantization parameters across various bit-widths, distinguishing between shared and specific parts. The shared part collaborates, while the specific part isolates precision conflicts. Additionally, we upgrade the normal quantizer to dynamic quantizer to mitigate statistical issues in the deployable mixed-precision supernet. Over the trained mixed-precision supernet, we utilize the quantized real-valued ranges to derive <em>quantized-bit-sensitivity</em>, which can serve as importance indicators for efficiently determining bit-width configurations, eliminating the need for iterative validation dataset evaluations. Extensive experiments show that Cobits outperforms previous state-of-the-art quantization methods on the ImageNet and COCO datasets while retaining superior efficiency. We show this approach dynamically adapts to varying bit-width and can generalize to various deployable backends. The code will be made public in <span><span>https://github.com/sunnyxiaohu/cobits</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106812"},"PeriodicalIF":6.0,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142551933","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rectangling and enhancing underwater stitched image via content-aware warping and perception balancing 通过内容感知扭曲和感知平衡对水下拼接图像进行矩形化和增强
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1016/j.neunet.2024.106809
Laibin Chang , Yunke Wang , Bo Du , Chang Xu
Single underwater images often face limitations in field-of-view and visual perception due to scattering and absorption. Numerous image stitching techniques have attempted to provide a wider viewing range, but the resulting stitched images may exhibit unsightly irregular boundaries. Unlike natural landscapes, the absence of reliable high-fidelity references in water complicates the replicability of these deep learning-based methods, leading to unpredictable distortions in cross-domain applications. To address these challenges, we propose an Underwater Wide-field Image Rectangling and Enhancement (UWIRE) framework that incorporates two procedures, i.e., the R-procedure and E-procedure, both of which employ self-coordinated modes, requiring only a single underwater stitched image as input. The R-procedure rectangles the irregular boundaries in stitched images by employing the initial shape resizing and mesh-based image preservation warping. Instead of local linear constraints, we use complementary optimization of boundary–structure–content to ensure a natural appearance with minimal distortion. The E-procedure enhances the rectangled image by employing parameter-adaptive correction to balance information distribution across channels. We further propose an attentive weight-guided fusion method to balance the perception of color restoration, contrast enhancement, and texture sharpening in a complementary manner. Comprehensive experiments demonstrate the superior performance of our UWIRE framework over state-of-the-art image rectangling and enhancement methods, both in quantitative and qualitative evaluation.
由于散射和吸收的原因,单幅水下图像在视野和视觉感知方面往往受到限制。许多图像拼接技术都试图提供更宽的视角范围,但拼接后的图像可能会出现难看的不规则边界。与自然景观不同,水中缺乏可靠的高保真参照物,这使得这些基于深度学习的方法的可复制性变得更加复杂,从而导致跨领域应用中出现不可预测的失真。为了应对这些挑战,我们提出了水下宽视场图像矩形化和增强(UWIRE)框架,该框架包含两个程序,即 R 程序和 E 程序,这两个程序都采用自协调模式,只需要一个水下拼接图像作为输入。R 程序通过调整初始形状大小和基于网格的图像保存扭曲,对拼接图像中的不规则边界进行矩形化处理。我们使用边界-结构-内容的互补优化来代替局部线性约束,以确保外观自然,失真最小。E 程序通过采用参数自适应校正来平衡各通道的信息分布,从而增强矩形图像。我们进一步提出了一种贴心的权重引导融合方法,以互补的方式平衡色彩还原、对比度增强和纹理锐化的感知。综合实验证明,我们的 UWIRE 框架在定量和定性评估方面都优于最先进的图像纠偏和增强方法。
{"title":"Rectangling and enhancing underwater stitched image via content-aware warping and perception balancing","authors":"Laibin Chang ,&nbsp;Yunke Wang ,&nbsp;Bo Du ,&nbsp;Chang Xu","doi":"10.1016/j.neunet.2024.106809","DOIUrl":"10.1016/j.neunet.2024.106809","url":null,"abstract":"<div><div>Single underwater images often face limitations in field-of-view and visual perception due to scattering and absorption. Numerous image stitching techniques have attempted to provide a wider viewing range, but the resulting stitched images may exhibit unsightly irregular boundaries. Unlike natural landscapes, the absence of reliable high-fidelity references in water complicates the replicability of these deep learning-based methods, leading to unpredictable distortions in cross-domain applications. To address these challenges, we propose an Underwater Wide-field Image Rectangling and Enhancement (UWIRE) framework that incorporates two procedures, <em>i.e.</em>, the R-procedure and E-procedure, both of which employ self-coordinated modes, requiring only a single underwater stitched image as input. The R-procedure rectangles the irregular boundaries in stitched images by employing the initial shape resizing and mesh-based image preservation warping. Instead of local linear constraints, we use complementary optimization of boundary–structure–content to ensure a natural appearance with minimal distortion. The E-procedure enhances the rectangled image by employing parameter-adaptive correction to balance information distribution across channels. We further propose an attentive weight-guided fusion method to balance the perception of color restoration, contrast enhancement, and texture sharpening in a complementary manner. Comprehensive experiments demonstrate the superior performance of our UWIRE framework over state-of-the-art image rectangling and enhancement methods, both in quantitative and qualitative evaluation.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106809"},"PeriodicalIF":6.0,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142551934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust generalized PCA for enhancing discriminability and recoverability 增强可辨别性和可恢复性的健壮通用 PCA。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-10-18 DOI: 10.1016/j.neunet.2024.106814
Zhenlei Dai , Liangchen Hu , Huaijiang Sun
The dependency of low-dimensional embedding to principal component space seriously limits the effectiveness of existing robust principal component analysis (PCA) algorithms. Simply projecting the original sample coordinates onto orthogonal principal component directions may not effectively address various noise-corrupted scenarios, impairing both discriminability and recoverability. Our method addresses this issue through a generalized PCA (GPCA), which optimizes regression bias rather than sample mean, leading to more adaptable properties. And, we propose a robust GPCA model with joint loss and regularization based on the 2,μ norm and 2,ν norms, respectively. This approach not only mitigates sensitivity to outliers but also enhances feature extraction and selection flexibility. Additionally, we introduce a truncated and reweighted loss strategy, where truncation eliminates severely deviated outliers, and reweighting prioritizes the remaining samples. These innovations collectively improve the GPCA model’s performance. To solve the proposed model, we propose a non-greedy iterative algorithm and theoretically guarantee the convergence. Experimental results demonstrate that the proposed GPCA model outperforms the previous robust PCA models in both recoverability and discrimination.
低维嵌入对主成分空间的依赖性严重限制了现有稳健主成分分析(PCA)算法的有效性。简单地将原始样本坐标投影到正交主成分方向上,可能无法有效解决各种噪声干扰情况,从而影响了可辨别性和可恢复性。我们的方法通过广义 PCA(GPCA)解决了这一问题,GPCA 优化的是回归偏差而不是样本平均值,因此具有更强的适应性。此外,我们还提出了一种稳健的 GPCA 模型,该模型具有联合损失和正则化,分别基于 ℓ2,μ 规范和 ℓ2,ν 规范。这种方法不仅能降低对异常值的敏感性,还能提高特征提取和选择的灵活性。此外,我们还引入了截断和重新加权损失策略,其中截断消除了严重偏离的异常值,而重新加权则优先考虑其余样本。这些创新共同提高了 GPCA 模型的性能。为了求解所提出的模型,我们提出了一种非贪心迭代算法,并从理论上保证了算法的收敛性。实验结果表明,所提出的 GPCA 模型在可恢复性和区分度方面都优于之前的鲁棒 PCA 模型。
{"title":"Robust generalized PCA for enhancing discriminability and recoverability","authors":"Zhenlei Dai ,&nbsp;Liangchen Hu ,&nbsp;Huaijiang Sun","doi":"10.1016/j.neunet.2024.106814","DOIUrl":"10.1016/j.neunet.2024.106814","url":null,"abstract":"<div><div>The dependency of low-dimensional embedding to principal component space seriously limits the effectiveness of existing robust principal component analysis (PCA) algorithms. Simply projecting the original sample coordinates onto orthogonal principal component directions may not effectively address various noise-corrupted scenarios, impairing both discriminability and recoverability. Our method addresses this issue through a generalized PCA (GPCA), which optimizes regression bias rather than sample mean, leading to more adaptable properties. And, we propose a robust GPCA model with joint loss and regularization based on the <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>2</mn><mo>,</mo><mi>μ</mi></mrow></msub></math></span> norm and <span><math><msub><mrow><mi>ℓ</mi></mrow><mrow><mn>2</mn><mo>,</mo><mi>ν</mi></mrow></msub></math></span> norms, respectively. This approach not only mitigates sensitivity to outliers but also enhances feature extraction and selection flexibility. Additionally, we introduce a truncated and reweighted loss strategy, where truncation eliminates severely deviated outliers, and reweighting prioritizes the remaining samples. These innovations collectively improve the GPCA model’s performance. To solve the proposed model, we propose a non-greedy iterative algorithm and theoretically guarantee the convergence. Experimental results demonstrate that the proposed GPCA model outperforms the previous robust PCA models in both recoverability and discrimination.</div></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":"181 ","pages":"Article 106814"},"PeriodicalIF":6.0,"publicationDate":"2024-10-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142511796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1