首页 > 最新文献

Neural Networks最新文献

英文 中文
Finding core labels for maximizing generalization of graph neural networks. 寻找核心标签,实现图神经网络泛化最大化
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-14 DOI: 10.1016/j.neunet.2024.106635
Sichao Fu, Xueqi Ma, Yibing Zhan, Fanyu You, Qinmu Peng, Tongliang Liu, James Bailey, Danilo Mandic

Graph neural networks (GNNs) have become a popular approach for semi-supervised graph representation learning. GNNs research has generally focused on improving methodological details, whereas less attention has been paid to exploring the importance of labeling the data. However, for semi-supervised learning, the quality of training data is vital. In this paper, we first introduce and elaborate on the problem of training data selection for GNNs. More specifically, focusing on node classification, we aim to select representative nodes from a graph used to train GNNs to achieve the best performance. To solve this problem, we are inspired by the popular lottery ticket hypothesis, typically used for sparse architectures, and we propose the following subset hypothesis for graph data: "There exists a core subset when selecting a fixed-size dataset from the dense training dataset, that can represent the properties of the dataset, and GNNs trained on this core subset can achieve a better graph representation". Equipped with this subset hypothesis, we present an efficient algorithm to identify the core data in the graph for GNNs. Extensive experiments demonstrate that the selected data (as a training set) can obtain performance improvements across various datasets and GNNs architectures.

图神经网络(GNN)已成为半监督图表示学习的一种流行方法。图神经网络的研究通常侧重于改进方法细节,而较少关注探索标记数据的重要性。然而,对于半监督学习来说,训练数据的质量至关重要。在本文中,我们首先介绍并阐述了 GNN 的训练数据选择问题。更具体地说,以节点分类为重点,我们的目标是从用于训练 GNN 的图中选择具有代表性的节点,以实现最佳性能。为了解决这个问题,我们受到通常用于稀疏架构的流行彩票假设的启发,提出了以下针对图数据的子集假设:"从密集的训练数据集中选择一个固定大小的数据集时,存在一个核心子集,它可以代表数据集的属性,在这个核心子集上训练的 GNN 可以实现更好的图表示"。有了这个子集假设,我们提出了一种高效算法来为 GNNs 识别图中的核心数据。广泛的实验证明,所选数据(作为训练集)可以在各种数据集和 GNN 架构中提高性能。
{"title":"Finding core labels for maximizing generalization of graph neural networks.","authors":"Sichao Fu, Xueqi Ma, Yibing Zhan, Fanyu You, Qinmu Peng, Tongliang Liu, James Bailey, Danilo Mandic","doi":"10.1016/j.neunet.2024.106635","DOIUrl":"https://doi.org/10.1016/j.neunet.2024.106635","url":null,"abstract":"<p><p>Graph neural networks (GNNs) have become a popular approach for semi-supervised graph representation learning. GNNs research has generally focused on improving methodological details, whereas less attention has been paid to exploring the importance of labeling the data. However, for semi-supervised learning, the quality of training data is vital. In this paper, we first introduce and elaborate on the problem of training data selection for GNNs. More specifically, focusing on node classification, we aim to select representative nodes from a graph used to train GNNs to achieve the best performance. To solve this problem, we are inspired by the popular lottery ticket hypothesis, typically used for sparse architectures, and we propose the following subset hypothesis for graph data: \"There exists a core subset when selecting a fixed-size dataset from the dense training dataset, that can represent the properties of the dataset, and GNNs trained on this core subset can achieve a better graph representation\". Equipped with this subset hypothesis, we present an efficient algorithm to identify the core data in the graph for GNNs. Extensive experiments demonstrate that the selected data (as a training set) can obtain performance improvements across various datasets and GNNs architectures.</p>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037533","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GazeForensics: DeepFake detection via gaze-guided spatial inconsistency learning GazeForensics:通过凝视引导的空间不一致性学习进行深度防伪检测。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-14 DOI: 10.1016/j.neunet.2024.106636

DeepFake detection is pivotal in personal privacy and public safety. With the iterative advancement of DeepFake techniques, high-quality forged videos and images are becoming increasingly deceptive. Prior research has seen numerous attempts by scholars to incorporate biometric features into the field of DeepFake detection. However, traditional biometric-based approaches tend to segregate biometric features from general ones and freeze the biometric feature extractor. These approaches resulted in the exclusion of valuable general features, potentially leading to a performance decline and, consequently, a failure to fully exploit the potential of biometric information in assisting DeepFake detection. Moreover, insufficient attention has been dedicated to scrutinizing gaze authenticity within the realm of DeepFake detection in recent years. In this paper, we introduce GazeForensics, an innovative DeepFake detection method that utilizes gaze representation obtained from a 3D gaze estimation model to regularize the corresponding representation within our DeepFake detection model, while concurrently integrating general features to further enhance the performance of our model. Experimental results demonstrate that our proposed GazeForensics method performs admirably in terms of performance and exhibits excellent interpretability.

深度防伪检测对个人隐私和公共安全至关重要。随着 DeepFake 技术的不断进步,高质量的伪造视频和图像越来越具有欺骗性。在之前的研究中,学者们曾多次尝试将生物特征纳入 DeepFake 检测领域。然而,传统的基于生物特征的方法倾向于将生物特征与一般特征分离,并冻结生物特征提取器。这些方法排除了有价值的一般特征,可能导致性能下降,从而无法充分利用生物识别信息在协助深度防伪检测方面的潜力。此外,近年来在 DeepFake 检测领域,人们对仔细检查凝视的真实性关注不够。在本文中,我们介绍了一种创新的 DeepFake 检测方法--GazeForensics,该方法利用从三维注视估计模型中获得的注视表征来正则化 DeepFake 检测模型中的相应表征,同时整合一般特征以进一步提高模型的性能。实验结果表明,我们提出的 GazeForensics 方法在性能方面表现出色,并具有出色的可解释性。
{"title":"GazeForensics: DeepFake detection via gaze-guided spatial inconsistency learning","authors":"","doi":"10.1016/j.neunet.2024.106636","DOIUrl":"10.1016/j.neunet.2024.106636","url":null,"abstract":"<div><p>DeepFake detection is pivotal in personal privacy and public safety. With the iterative advancement of DeepFake techniques, high-quality forged videos and images are becoming increasingly deceptive. Prior research has seen numerous attempts by scholars to incorporate biometric features into the field of DeepFake detection. However, traditional biometric-based approaches tend to segregate biometric features from general ones and freeze the biometric feature extractor. These approaches resulted in the exclusion of valuable general features, potentially leading to a performance decline and, consequently, a failure to fully exploit the potential of biometric information in assisting DeepFake detection. Moreover, insufficient attention has been dedicated to scrutinizing gaze authenticity within the realm of DeepFake detection in recent years. In this paper, we introduce <em>GazeForensics</em>, an innovative DeepFake detection method that utilizes gaze representation obtained from a 3D gaze estimation model to regularize the corresponding representation within our DeepFake detection model, while concurrently integrating general features to further enhance the performance of our model. Experimental results demonstrate that our proposed <em>GazeForensics</em> method performs admirably in terms of performance and exhibits excellent interpretability.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037535","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FLAT: Fusing layer representations for more efficient transfer learning in NLP FLAT:融合层表征,提高 NLP 迁移学习效率
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1016/j.neunet.2024.106631

Parameter efficient transfer learning (PETL) methods provide an efficient alternative for fine-tuning. However, typical PETL methods inject the same structures to all Pre-trained Language Model (PLM) layers and only use the final hidden states for downstream tasks, regardless of the knowledge diversity across PLM layers. Additionally, the backpropagation path of existing PETL methods still passes through the frozen PLM during training, which is computational and memory inefficient. In this paper, we propose FLAT, a generic PETL method that explicitly and individually combines knowledge across all PLM layers based on the tokens to perform a better transferring. FLAT considers the backbone PLM as a feature extractor and combines the features in a side-network, hence the backpropagation does not involve the PLM, which results in much less memory requirement than previous methods. The results on the GLUE benchmark show that FLAT outperforms other tuning techniques in the low-resource scenarios and achieves on-par performance in the high-resource scenarios with only 0.53% trainable parameters per task and 3.2× less GPU memory usagewith BERTbase. Besides, further ablation study is conducted to reveal that the proposed fusion layer effectively combines knowledge from PLM and helps the classifier to exploit the PLM knowledge to downstream tasks. We will release our code for better reproducibility.

参数高效迁移学习(PETL)方法为微调提供了一种高效的替代方法。然而,典型的 PETL 方法会向所有预训练语言模型(PLM)层注入相同的结构,并且只将最终隐藏状态用于下游任务,而不考虑 PLM 层间的知识多样性。此外,现有 PETL 方法的反向传播路径在训练过程中仍会经过冻结的 PLM,这在计算和内存方面都是低效的。在本文中,我们提出了一种通用的 PETL 方法--FLAT,该方法基于令牌明确、单独地将所有 PLM 层的知识结合起来,以实现更好的转移。FLAT 将骨干 PLM 视为特征提取器,并在侧网络中组合特征,因此反向传播不涉及 PLM,从而比以前的方法所需内存少得多。GLUE 基准测试结果表明,FLAT 在低资源场景下的性能优于其他调优技术,在高资源场景下的性能与其他调优技术相当,每个任务的可训练参数仅为 0.53%,GPU 内存使用量比 BERTbase 少 3.2 倍。此外,我们还进行了进一步的消融研究,发现所提出的融合层有效地结合了 PLM 知识,并帮助分类器利用 PLM 知识完成下游任务。我们将发布我们的代码,以提高可重复性。
{"title":"FLAT: Fusing layer representations for more efficient transfer learning in NLP","authors":"","doi":"10.1016/j.neunet.2024.106631","DOIUrl":"10.1016/j.neunet.2024.106631","url":null,"abstract":"<div><p>Parameter efficient transfer learning (PETL) methods provide an efficient alternative for fine-tuning. However, typical PETL methods inject the same structures to all Pre-trained Language Model (PLM) layers and only use the final hidden states for downstream tasks, regardless of the knowledge diversity across PLM layers. Additionally, the backpropagation path of existing PETL methods still passes through the frozen PLM during training, which is computational and memory inefficient. In this paper, we propose FLAT, a generic PETL method that explicitly and individually combines knowledge across all PLM layers based on the tokens to perform a better transferring. FLAT considers the backbone PLM as a feature extractor and combines the features in a side-network, hence the backpropagation does not involve the PLM, which results in much less memory requirement than previous methods. The results on the GLUE benchmark show that FLAT outperforms other tuning techniques in the low-resource scenarios and achieves on-par performance in the high-resource scenarios with only 0.53% trainable parameters per task and <span><math><mrow><mn>3</mn><mo>.</mo><mn>2</mn><mo>×</mo></mrow></math></span> less GPU memory usagewith BERT<span><math><msub><mrow></mrow><mrow><mtext>base</mtext></mrow></msub></math></span>. Besides, further ablation study is conducted to reveal that the proposed fusion layer effectively combines knowledge from PLM and helps the classifier to exploit the PLM knowledge to downstream tasks. We will release our code for better reproducibility.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142002181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Emotion recognition using hierarchical spatial–temporal learning transformer from regional to global brain 利用从区域到全球大脑的分层时空学习转换器进行情绪识别
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1016/j.neunet.2024.106624

Emotion recognition is an essential but challenging task in human–computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial–temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial–temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial–temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the long-range spatial–temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.

在人机交互系统中,情绪识别是一项必不可少但又极具挑战性的任务,因为每种情绪都具有独特的空间结构和动态时间依赖性。然而,目前的方法无法准确捕捉不同脑区的脑电图(EEG)信号对情绪识别的复杂影响。因此,本文设计了一种基于变压器的方法(以 R2G-STLT 表示),该方法依赖于具有区域到全局分层学习功能的时空变压器编码器,可学习从电极级到脑区级的代表性时空特征。区域时空变换器(RST-Trans)编码器旨在获取电极层面的空间信息和上下文相关性,从而学习区域时空特征。然后,利用全局时空变换器(GST-Trans)编码器提取可靠的全局时空特征,以反映不同脑区对情绪识别任务的影响。此外,在 GST-Trans 编码器中加入了多头注意力机制,使其能够捕捉大脑区域之间的长程时空信息。最后,在 DEAP、SEED 和 SEED-IV 数据集的每个频段上进行了与受试者无关的实验,以评估所提出模型的性能。结果表明,R2G-STLT 模型超越了几种最先进的方法。
{"title":"Emotion recognition using hierarchical spatial–temporal learning transformer from regional to global brain","authors":"","doi":"10.1016/j.neunet.2024.106624","DOIUrl":"10.1016/j.neunet.2024.106624","url":null,"abstract":"<div><p>Emotion recognition is an essential but challenging task in human–computer interaction systems due to the distinctive spatial structures and dynamic temporal dependencies associated with each emotion. However, current approaches fail to accurately capture the intricate effects of electroencephalogram (EEG) signals across different brain regions on emotion recognition. Therefore, this paper designs a transformer-based method, denoted by R2G-STLT, which relies on a spatial–temporal transformer encoder with regional to global hierarchical learning that learns the representative spatiotemporal features from the electrode level to the brain-region level. The regional spatial–temporal transformer (RST-Trans) encoder is designed to obtain spatial information and context dependence at the electrode level aiming to learn the regional spatiotemporal features. Then, the global spatial–temporal transformer (GST-Trans) encoder is utilized to extract reliable global spatiotemporal features, reflecting the impact of various brain regions on emotion recognition tasks. Moreover, the multi-head attention mechanism is placed into the GST-Trans encoder to empower it to capture the long-range spatial–temporal information among the brain regions. Finally, subject-independent experiments are conducted on each frequency band of the DEAP, SEED, and SEED-IV datasets to assess the performance of the proposed model. Results indicate that the R2G-STLT model surpasses several state-of-the-art approaches.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142006990","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Universal approximation theorem for vector- and hypercomplex-valued neural networks 向量和超复值神经网络的通用近似定理
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1016/j.neunet.2024.106632

The universal approximation theorem states that a neural network with one hidden layer can approximate continuous functions on compact sets with any desired precision. This theorem supports using neural networks for various applications, including regression and classification tasks. Furthermore, it is valid for real-valued neural networks and some hypercomplex-valued neural networks such as complex-, quaternion-, tessarine-, and Clifford-valued neural networks. However, hypercomplex-valued neural networks are a type of vector-valued neural network defined on an algebra with additional algebraic or geometric properties. This paper extends the universal approximation theorem for a wide range of vector-valued neural networks, including hypercomplex-valued models as particular instances. Precisely, we introduce the concept of non-degenerate algebra and state the universal approximation theorem for neural networks defined on such algebras.

通用逼近定理指出,具有一个隐藏层的神经网络可以以任意所需的精度逼近紧凑集合上的连续函数。该定理支持将神经网络用于各种应用,包括回归和分类任务。此外,它还适用于实值神经网络和一些超复数值神经网络,如复数、四元、魔方和克利福德值神经网络。然而,超复值神经网络是一种在代数上定义的矢量值神经网络,具有额外的代数或几何特性。本文扩展了普遍逼近定理,适用于各种矢量值神经网络,包括作为特殊实例的超复值模型。确切地说,我们引入了非退化代数的概念,并阐述了定义在此类代数上的神经网络的通用近似定理。
{"title":"Universal approximation theorem for vector- and hypercomplex-valued neural networks","authors":"","doi":"10.1016/j.neunet.2024.106632","DOIUrl":"10.1016/j.neunet.2024.106632","url":null,"abstract":"<div><p>The universal approximation theorem states that a neural network with one hidden layer can approximate continuous functions on compact sets with any desired precision. This theorem supports using neural networks for various applications, including regression and classification tasks. Furthermore, it is valid for real-valued neural networks and some hypercomplex-valued neural networks such as complex-, quaternion-, tessarine-, and Clifford-valued neural networks. However, hypercomplex-valued neural networks are a type of vector-valued neural network defined on an algebra with additional algebraic or geometric properties. This paper extends the universal approximation theorem for a wide range of vector-valued neural networks, including hypercomplex-valued models as particular instances. Precisely, we introduce the concept of non-degenerate algebra and state the universal approximation theorem for neural networks defined on such algebras.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037538","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian learning of feature spaces for multitask regression 多任务回归的特征空间贝叶斯学习
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-13 DOI: 10.1016/j.neunet.2024.106619

This paper introduces a novel approach to learn multi-task regression models with constrained architecture complexity. The proposed model, named RFF-BLR, consists of a randomised feedforward neural network with two fundamental characteristics: a single hidden layer whose units implement the random Fourier features that approximate an RBF kernel, and a Bayesian formulation that optimises the weights connecting the hidden and output layers. The RFF-based hidden layer inherits the robustness of kernel methods. The Bayesian formulation enables promoting multioutput sparsity: all tasks interplay during the optimisation to select a compact subset of the hidden layer units that serve as common non-linear mapping for every tasks. The experimental results show that the RFF-BLR framework can lead to significant performance improvements compared to the state-of-the-art methods in multitask nonlinear regression, especially in small-sized training dataset scenarios.

本文介绍了一种学习多任务回归模型的新方法,其架构复杂度受到限制。该模型被命名为 RFF-BLR,由随机化前馈神经网络组成,具有两个基本特征:一是单隐层,其单元实现了近似 RBF 核的随机傅立叶特征;二是贝叶斯公式,优化了连接隐层和输出层的权重。基于 RFF 的隐藏层继承了核方法的鲁棒性。贝叶斯公式能够促进多输出稀疏性:所有任务在优化过程中相互影响,以选择一个紧凑的隐藏层单元子集,作为每个任务的共同非线性映射。实验结果表明,与最先进的多任务非线性回归方法相比,RFF-BLR 框架能显著提高性能,尤其是在小规模训练数据集的情况下。
{"title":"Bayesian learning of feature spaces for multitask regression","authors":"","doi":"10.1016/j.neunet.2024.106619","DOIUrl":"10.1016/j.neunet.2024.106619","url":null,"abstract":"<div><p>This paper introduces a novel approach to learn multi-task regression models with constrained architecture complexity. The proposed model, named RFF-BLR, consists of a randomised feedforward neural network with two fundamental characteristics: a single hidden layer whose units implement the random Fourier features that approximate an RBF kernel, and a Bayesian formulation that optimises the weights connecting the hidden and output layers. The RFF-based hidden layer inherits the robustness of kernel methods. The Bayesian formulation enables promoting multioutput sparsity: all tasks interplay during the optimisation to select a compact subset of the hidden layer units that serve as common non-linear mapping for every tasks. The experimental results show that the RFF-BLR framework can lead to significant performance improvements compared to the state-of-the-art methods in multitask nonlinear regression, especially in small-sized training dataset scenarios.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0893608024005434/pdfft?md5=b5ec56e8de25c78b0d2793417101b953&pid=1-s2.0-S0893608024005434-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142006992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deformation depth decoupling network for point cloud domain adaptation 用于点云域适应的变形深度解耦网络。
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1016/j.neunet.2024.106626

Recently, point cloud domain adaptation (DA) practices have been implemented to improve the generalization ability of deep learning models on point cloud data. However, variations across domains often result in decreased performance of models trained on different distributed data sources. Previous studies have focused on output-level domain alignment to address this challenge. But this approach may increase the amount of errors experienced when aligning different domains, particularly for targets that would otherwise be predicted incorrectly. Therefore, in this study, we propose an input-level discretization-based matching to enhance the generalization ability of DA. Specifically, an efficient geometric deformation depth decoupling network (3DeNet) is implemented to learn the knowledge from the source domain and embed it into an implicit feature space, which facilitates the effective constraint of unsupervised predictions for downstream tasks. Secondly, we demonstrate that the sparsity within the implicit feature space varies between domains, rendering domain differences difficult to support. Consequently, we match sets of neighboring points with different densities and biases by differentiating the adaptive densities. Finally, inter-domain differences are aligned by constraining the loss originating from and between the target domains. We conduct experiments on point cloud DA datasets PointDA-10 and PointSegDA, achieving advanced results (over 1.2% and 1% on average).

最近,为了提高深度学习模型在点云数据上的泛化能力,人们实施了点云领域适应(DA)实践。然而,不同领域之间的差异往往会导致在不同分布式数据源上训练的模型性能下降。以往的研究主要通过输出级域对齐来解决这一难题。但这种方法可能会增加不同域对齐时的误差,特别是对于那些本来会被错误预测的目标。因此,在本研究中,我们提出了一种基于输入级离散化的匹配方法,以增强 DA 的泛化能力。具体来说,我们采用了一种高效的几何形变深度解耦网络(3DeNet)来学习源域的知识,并将其嵌入到隐式特征空间中,从而为下游任务的无监督预测提供了有效的约束。其次,我们证明了隐式特征空间内的稀疏性因领域而异,导致难以支持领域差异。因此,我们通过区分自适应密度来匹配具有不同密度和偏差的相邻点集。最后,通过限制来自目标域和目标域之间的损失来调整域间差异。我们在点云数据集 PointDA-10 和 PointSegDA 上进行了实验,取得了先进的结果(平均超过 1.2% 和 1%)。
{"title":"Deformation depth decoupling network for point cloud domain adaptation","authors":"","doi":"10.1016/j.neunet.2024.106626","DOIUrl":"10.1016/j.neunet.2024.106626","url":null,"abstract":"<div><p>Recently, point cloud domain adaptation (DA) practices have been implemented to improve the generalization ability of deep learning models on point cloud data. However, variations across domains often result in decreased performance of models trained on different distributed data sources. Previous studies have focused on output-level domain alignment to address this challenge. But this approach may increase the amount of errors experienced when aligning different domains, particularly for targets that would otherwise be predicted incorrectly. Therefore, in this study, we propose an input-level discretization-based matching to enhance the generalization ability of DA. Specifically, an efficient geometric deformation depth decoupling network (3DeNet) is implemented to learn the knowledge from the source domain and embed it into an implicit feature space, which facilitates the effective constraint of unsupervised predictions for downstream tasks. Secondly, we demonstrate that the sparsity within the implicit feature space varies between domains, rendering domain differences difficult to support. Consequently, we match sets of neighboring points with different densities and biases by differentiating the adaptive densities. Finally, inter-domain differences are aligned by constraining the loss originating from and between the target domains. We conduct experiments on point cloud DA datasets PointDA-10 and PointSegDA, achieving advanced results (over 1.2% and 1% on average).</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142037530","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mask-Shift-Inference: A novel paradigm for domain generalization 掩码转换推理:领域泛化的新范式
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1016/j.neunet.2024.106629

Domain Generalization (DG) focuses on the Out-Of-Distribution (OOD) generalization, which is able to learn a robust model that generalizes the knowledge acquired from the source domain to the unseen target domain. However, due to the existence of the domain shift, domain-invariant representation learning is challenging. Guided by fine-grained knowledge, we propose a novel paradigm Mask-Shift-Inference (MSI) for DG based on the architecture of Convolutional Neural Networks (CNN). Different from relying on a series of constraints and assumptions for model optimization, this paradigm novelly shifts the focus to feature channels in the latent space for domain-invariant representation learning. We put forward a two-branch working mode of a main module and multiple domain-specific sub-modules. The latter can only achieve good prediction performance in its own specific domain but poor predictions in other source domains, which provides the main module with the fine-grained knowledge guidance and contributes to the improvement of the cognitive ability of MSI. Firstly, during the forward propagation of the main module, the proposed MSI accurately discards unstable channels based on spurious classifications varying across domains, which have domain-specific prediction limitations and are not conducive to generalization. In this process, a progressive scheme is adopted to adaptively increase the masking ratio according to the training progress to further reduce the risk of overfitting. Subsequently, our paradigm enters the compatible shifting stage before the formal prediction. Based on maximizing semantic retention, we implement the domain style matching and shifting through the simple transformation through Fourier transform, which can explicitly and safely shift the target domain back to the source domain whose style is closest to it, requiring no additional model updates and reducing the domain gap. Eventually, the paradigm MSI enters the formal inference stage. The updated target domain is predicted in the main module trained in the previous stage with the benefit of familiar knowledge from the nearest source domain masking scheme. Our paradigm is logically progressive, which can intuitively exclude the confounding influence of domain-specific spurious information along with mitigating domain shifts and implicitly perform semantically invariant representation learning, achieving robust OOD generalization. Extensive experimental results on PACS, VLCS, Office-Home and DomainNet datasets verify the superiority and effectiveness of the proposed method.

领域泛化(Domain Generalization,DG)侧重于分布外泛化(Out-Of-Distribution,OOD),它能够学习一个稳健的模型,将从源领域获得的知识泛化到未见过的目标领域。然而,由于领域偏移的存在,与领域无关的表征学习具有挑战性。在细粒度知识的指导下,我们提出了一种基于卷积神经网络(CNN)架构的用于 DG 的新型掩码偏移推理(MSI)范例。与依赖一系列约束和假设进行模型优化不同,这种范式新颖地将重点转移到了潜空间的特征通道上,以实现领域不变的表征学习。我们提出了一个主模块和多个特定领域子模块的双分支工作模式。子模块只能在自己的特定领域获得良好的预测性能,而在其他源领域的预测性能较差,这为主模块提供了细粒度的知识指导,有助于提高 MSI 的认知能力。首先,在主模块的前向传播过程中,所提出的 MSI 会根据不同领域的虚假分类准确摒弃不稳定的信道,因为这些信道具有特定领域的预测局限性,不利于泛化。在这一过程中,我们采用了渐进式方案,根据训练进度自适应地增加屏蔽率,以进一步降低过拟合风险。随后,我们的范式进入正式预测前的兼容转换阶段。在语义保留最大化的基础上,我们通过傅立叶变换的简单变换来实现域风格的匹配和转移,这样就可以明确而安全地将目标域转移回风格最接近的源域,无需额外的模型更新,减少了域差距。最终,范式 MSI 进入正式推理阶段。更新后的目标域将在前一阶段训练好的主模块中进行预测,并从最近的源域屏蔽方案中获得熟悉的知识。我们的范式在逻辑上是渐进的,可以直观地排除特定领域虚假信息的干扰影响,同时减轻领域偏移,并隐式地执行语义不变表征学习,实现稳健的 OOD 泛化。在 PACS、VLCS、Office-Home 和 DomainNet 数据集上的大量实验结果验证了所提方法的优越性和有效性。
{"title":"Mask-Shift-Inference: A novel paradigm for domain generalization","authors":"","doi":"10.1016/j.neunet.2024.106629","DOIUrl":"10.1016/j.neunet.2024.106629","url":null,"abstract":"<div><p>Domain Generalization (DG) focuses on the Out-Of-Distribution (OOD) generalization, which is able to learn a robust model that generalizes the knowledge acquired from the source domain to the unseen target domain. However, due to the existence of the domain shift, domain-invariant representation learning is challenging. Guided by fine-grained knowledge, we propose a novel paradigm Mask-Shift-Inference (MSI) for DG based on the architecture of Convolutional Neural Networks (CNN). Different from relying on a series of constraints and assumptions for model optimization, this paradigm novelly shifts the focus to feature channels in the latent space for domain-invariant representation learning. We put forward a two-branch working mode of a main module and multiple domain-specific sub-modules. The latter can only achieve good prediction performance in its own specific domain but poor predictions in other source domains, which provides the main module with the fine-grained knowledge guidance and contributes to the improvement of the cognitive ability of MSI. Firstly, during the forward propagation of the main module, the proposed MSI accurately discards unstable channels based on spurious classifications varying across domains, which have domain-specific prediction limitations and are not conducive to generalization. In this process, a progressive scheme is adopted to adaptively increase the masking ratio according to the training progress to further reduce the risk of overfitting. Subsequently, our paradigm enters the compatible shifting stage before the formal prediction. Based on maximizing semantic retention, we implement the domain style matching and shifting through the simple transformation through Fourier transform, which can explicitly and safely shift the target domain back to the source domain whose style is closest to it, requiring no additional model updates and reducing the domain gap. Eventually, the paradigm MSI enters the formal inference stage. The updated target domain is predicted in the main module trained in the previous stage with the benefit of familiar knowledge from the nearest source domain masking scheme. Our paradigm is logically progressive, which can intuitively exclude the confounding influence of domain-specific spurious information along with mitigating domain shifts and implicitly perform semantically invariant representation learning, achieving robust OOD generalization. Extensive experimental results on PACS, VLCS, Office-Home and DomainNet datasets verify the superiority and effectiveness of the proposed method.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141993422","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A smoothing approximation-based adaptive neurodynamic approach for nonsmooth resource allocation problem 基于平滑近似的非平滑资源分配问题自适应神经动力学方法
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-12 DOI: 10.1016/j.neunet.2024.106625

In this paper, a smoothing approximation-based adaptive neurodynamic approach is proposed for a nonsmooth resource allocation problem (NRAP) with multiple constraints. The smoothing approximation method is combined with multi-agent systems to avoid the introduction of set-valued subgradient terms, thereby facilitating the practical implementation of the neurodynamic approach. In addition, using the adaptive penalty technique, private inequality constraints are processed, which eliminates the need for additional quantitative estimation of penalty parameters and significantly reduces the computational cost. Moreover, to reduce the impact of smoothing approximation on the convergence of the neurodynamic approach, time-varying control parameters are introduced. Due to the parallel computing characteristics of multi-agent systems, the neurodynamic approach proposed in this paper is completely distributed. Theoretical proof shows that the state solution of the neurodynamic approach converges to the optimal solution of NRAP. Finally, two application examples are used to validate the feasibility of the neurodynamic approach.

本文针对具有多重约束条件的非光滑资源分配问题(NRAP),提出了一种基于平滑近似的自适应神经动力学方法。平滑近似方法与多代理系统相结合,避免了引入集值子梯度项,从而促进了神经动力学方法的实际应用。此外,利用自适应惩罚技术处理私有不等式约束,无需额外对惩罚参数进行定量估计,大大降低了计算成本。此外,为了减少平滑近似对神经动力学方法收敛性的影响,还引入了时变控制参数。由于多代理系统的并行计算特性,本文提出的神经动力学方法是完全分布式的。理论证明表明,神经动力学方法的状态解收敛于 NRAP 的最优解。最后,两个应用实例验证了神经动力学方法的可行性。
{"title":"A smoothing approximation-based adaptive neurodynamic approach for nonsmooth resource allocation problem","authors":"","doi":"10.1016/j.neunet.2024.106625","DOIUrl":"10.1016/j.neunet.2024.106625","url":null,"abstract":"<div><p>In this paper, a smoothing approximation-based adaptive neurodynamic approach is proposed for a nonsmooth resource allocation problem (NRAP) with multiple constraints. The smoothing approximation method is combined with multi-agent systems to avoid the introduction of set-valued subgradient terms, thereby facilitating the practical implementation of the neurodynamic approach. In addition, using the adaptive penalty technique, private inequality constraints are processed, which eliminates the need for additional quantitative estimation of penalty parameters and significantly reduces the computational cost. Moreover, to reduce the impact of smoothing approximation on the convergence of the neurodynamic approach, time-varying control parameters are introduced. Due to the parallel computing characteristics of multi-agent systems, the neurodynamic approach proposed in this paper is completely distributed. Theoretical proof shows that the state solution of the neurodynamic approach converges to the optimal solution of NRAP. Finally, two application examples are used to validate the feasibility of the neurodynamic approach.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142012446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Randomized algorithms for large-scale dictionary learning 大规模词典学习的随机算法
IF 6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pub Date : 2024-08-10 DOI: 10.1016/j.neunet.2024.106628

Dictionary learning is an important sparse representation algorithm which has been widely used in machine learning and artificial intelligence. However, for massive data in the big data era, classical dictionary learning algorithms are computationally expensive and even can be infeasible. To overcome this difficulty, we propose new dictionary learning methods based on randomized algorithms. The contributions of this work are as follows. First, we find that dictionary matrix is often numerically low-rank. Based on this property, we apply randomized singular value decomposition (RSVD) to the dictionary matrix, and propose a randomized algorithm for linear dictionary learning. Compared with the classical K-SVD algorithm, an advantage is that one can update all the elements of the dictionary matrix simultaneously. Second, to the best of our knowledge, there are few theoretical results on why one can solve the involved matrix computation problems inexactly in dictionary learning. To fill-in this gap, we show the rationality of this randomized algorithm with inexact solving, from a matrix perturbation analysis point of view. Third, based on the numerically low-rank property and Nyström approximation of the kernel matrix, we propose a randomized kernel dictionary learning algorithm, and establish the distance between the exact solution and the computed solution, to show the effectiveness of the proposed randomized kernel dictionary learning algorithm. Fourth, we propose an efficient scheme for the testing stage in kernel dictionary learning. By using this strategy, there is no need to form nor store kernel matrices explicitly both in the training and the testing stages. Comprehensive numerical experiments are performed on some real-world data sets. Numerical results demonstrate the rationality of our strategies, and show that the proposed algorithms are much efficient than some state-of-the-art dictionary learning algorithms. The MATLAB codes of the proposed algorithms are publicly available from https://github.com/Jiali-yang/RALDL_RAKDL.

字典学习是一种重要的稀疏表示算法,已被广泛应用于机器学习和人工智能领域。然而,对于大数据时代的海量数据,经典的字典学习算法计算成本高昂,甚至不可行。为了克服这一困难,我们提出了基于随机算法的新字典学习方法。这项工作的贡献如下。首先,我们发现字典矩阵在数值上通常是低秩的。基于这一特性,我们对字典矩阵进行了随机奇异值分解(RSVD),并提出了一种线性字典学习的随机算法。与经典的 K-SVD 算法相比,该算法的优势在于可以同时更新字典矩阵的所有元素。其次,据我们所知,关于在字典学习中为什么可以不精确地解决所涉及的矩阵计算问题的理论成果很少。为了填补这一空白,我们从矩阵扰动分析的角度展示了这种非精确求解随机算法的合理性。第三,基于核矩阵的数值低阶特性和 Nyström 近似,我们提出了一种随机核词典学习算法,并建立了精确解与计算解之间的距离,以证明所提出的随机核词典学习算法的有效性。第四,我们提出了内核字典学习中测试阶段的高效方案。通过使用这种策略,在训练和测试阶段都不需要明确地形成或存储内核矩阵。我们在一些真实世界的数据集上进行了全面的数值实验。数值结果证明了我们策略的合理性,并表明所提出的算法比一些最先进的字典学习算法要高效得多。建议算法的 MATLAB 代码可从 https://github.com/Jiali-yang/RALDL_RAKDL 公开获取。
{"title":"Randomized algorithms for large-scale dictionary learning","authors":"","doi":"10.1016/j.neunet.2024.106628","DOIUrl":"10.1016/j.neunet.2024.106628","url":null,"abstract":"<div><p>Dictionary learning is an important sparse representation algorithm which has been widely used in machine learning and artificial intelligence. However, for massive data in the big data era, classical dictionary learning algorithms are computationally expensive and even can be infeasible. To overcome this difficulty, we propose new dictionary learning methods based on randomized algorithms. The contributions of this work are as follows. First, we find that dictionary matrix is often numerically low-rank. Based on this property, we apply randomized singular value decomposition (RSVD) to the dictionary matrix, and propose a randomized algorithm for linear dictionary learning. Compared with the classical K-SVD algorithm, an advantage is that one can update all the elements of the dictionary matrix simultaneously. Second, to the best of our knowledge, there are few theoretical results on why one can solve the involved matrix computation problems <em>inexactly</em> in dictionary learning. To fill-in this gap, we show the rationality of this randomized algorithm with inexact solving, from a matrix perturbation analysis point of view. Third, based on the numerically low-rank property and Nyström approximation of the kernel matrix, we propose a randomized kernel dictionary learning algorithm, and establish the distance between the exact solution and the computed solution, to show the effectiveness of the proposed randomized kernel dictionary learning algorithm. Fourth, we propose an efficient scheme for the testing stage in kernel dictionary learning. By using this strategy, there is no need to form nor store kernel matrices explicitly both in the training and the testing stages. Comprehensive numerical experiments are performed on some real-world data sets. Numerical results demonstrate the rationality of our strategies, and show that the proposed algorithms are much efficient than some state-of-the-art dictionary learning algorithms. The MATLAB codes of the proposed algorithms are publicly available from <span><span>https://github.com/Jiali-yang/RALDL_RAKDL</span><svg><path></path></svg></span>.</p></div>","PeriodicalId":49763,"journal":{"name":"Neural Networks","volume":null,"pages":null},"PeriodicalIF":6.0,"publicationDate":"2024-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142012445","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Neural Networks
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1