Pub Date : 2024-10-28DOI: 10.1016/j.neucom.2024.128761
Haoran Lu , Lei Wang , Xiaoliang Ma , Jun Cheng , Mengchu Zhou
Graph Neural Networks (GNNs) have emerged as a powerful tool for analyzing and modeling graph-structured data. In recent years, GNNs have gained significant attention in various domains. This review paper aims to provide an overview of the state-of-the-art graph neural network techniques and their industrial applications. First, we introduce the fundamental concepts and architectures of GNNs, highlighting their ability to capture complex relationships and dependencies in graph data. We then delve into the variants and evolution of graphs, including directed graphs, heterogeneous graphs, dynamic graphs, and hypergraphs. Next, we discuss the interpretability of GNN, and GNN theory including graph augmentation, expressivity, and over-smoothing. Finally, we introduce the specific use cases of GNNs in industrial settings, including finance, biology, knowledge graphs, recommendation systems, Internet of Things (IoT), and knowledge distillation. This review paper highlights the immense potential of GNNs in solving real-world problems, while also addressing the challenges and opportunities for further advancement in this field.
{"title":"A survey of graph neural networks and their industrial applications","authors":"Haoran Lu , Lei Wang , Xiaoliang Ma , Jun Cheng , Mengchu Zhou","doi":"10.1016/j.neucom.2024.128761","DOIUrl":"10.1016/j.neucom.2024.128761","url":null,"abstract":"<div><div>Graph Neural Networks (GNNs) have emerged as a powerful tool for analyzing and modeling graph-structured data. In recent years, GNNs have gained significant attention in various domains. This review paper aims to provide an overview of the state-of-the-art graph neural network techniques and their industrial applications. First, we introduce the fundamental concepts and architectures of GNNs, highlighting their ability to capture complex relationships and dependencies in graph data. We then delve into the variants and evolution of graphs, including directed graphs, heterogeneous graphs, dynamic graphs, and hypergraphs. Next, we discuss the interpretability of GNN, and GNN theory including graph augmentation, expressivity, and over-smoothing. Finally, we introduce the specific use cases of GNNs in industrial settings, including finance, biology, knowledge graphs, recommendation systems, Internet of Things (IoT), and knowledge distillation. This review paper highlights the immense potential of GNNs in solving real-world problems, while also addressing the challenges and opportunities for further advancement in this field.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128761"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142586922","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-28DOI: 10.1016/j.neucom.2024.128785
Xinglong Sun , Haijiang Sun , Shan Jiang , Jiacheng Wang , Xilai Wei , Zhonghe Hu
Classification-regression prediction networks have realized impressive success in several modern deep trackers. However, there is an inherent difference between classification and regression tasks, so they have diverse even opposite demands for feature matching. Existed models always ignore the key issue and only employ a unified matching block in two task branches, decaying the decision quality. Besides, these models also struggle with decision misalignment situation. In this paper, we propose a multi-attention associate prediction network (MAPNet) to tackle the above problems. Concretely, two novel matchers, i.e., category-aware matcher and spatial-aware matcher, are first designed for feature comparison by integrating self, cross, channel or spatial attentions organically. They are capable of fully capturing the category-related semantics for classification and the local spatial contexts for regression, respectively. Then, we present a dual alignment module to enhance the correspondences between two branches, which is useful to find the optimal tracking solution. Finally, we describe a Siamese tracker built upon the proposed prediction network, which achieves the leading performance on five tracking benchmarks, consisting of LaSOT, TrackingNet, GOT-10k, TNL2k and UAV123, and surpasses other state-of-the-art approaches.
{"title":"Multi-attention associate prediction network for visual tracking","authors":"Xinglong Sun , Haijiang Sun , Shan Jiang , Jiacheng Wang , Xilai Wei , Zhonghe Hu","doi":"10.1016/j.neucom.2024.128785","DOIUrl":"10.1016/j.neucom.2024.128785","url":null,"abstract":"<div><div>Classification-regression prediction networks have realized impressive success in several modern deep trackers. However, there is an inherent difference between classification and regression tasks, so they have diverse even opposite demands for feature matching. Existed models always ignore the key issue and only employ a unified matching block in two task branches, decaying the decision quality. Besides, these models also struggle with decision misalignment situation. In this paper, we propose a multi-attention associate prediction network (MAPNet) to tackle the above problems. Concretely, two novel matchers, i.e., category-aware matcher and spatial-aware matcher, are first designed for feature comparison by integrating self, cross, channel or spatial attentions organically. They are capable of fully capturing the category-related semantics for classification and the local spatial contexts for regression, respectively. Then, we present a dual alignment module to enhance the correspondences between two branches, which is useful to find the optimal tracking solution. Finally, we describe a Siamese tracker built upon the proposed prediction network, which achieves the leading performance on five tracking benchmarks, consisting of LaSOT, TrackingNet, GOT-10k, TNL2k and UAV123, and surpasses other state-of-the-art approaches.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128785"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142573298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The spiking neural networks (SNN) that are inspired by the human brain offers wider scope for application in the growth of neuromorphic computing systems due to their brain level computational capabilities, reduced power consumption, and minimal data movement cost, among other advantages. Spike-based neurons and synapses are the essential building blocks of SNN, and their efficient implementation is vital to their performance enhancement. In this regard, the design and implementation of spiking neurons have been the major focus among the researchers. In this paper, functioning of different leaky integrate fire (LIF)-based spiking neuron circuits like frequency adaptable CMOS-based LIF, resistor-capacitor-based (RC) LIF, and volatile memristor-based LIF are subjected to comparison. The work mainly focuses on revealing analysis of spike duration and amplitude, number of spikes produced during excitation period, threshold operation, field of application, and various other significant parameters of aforementioned neuron circuits. Extensive simulations of these circuits are carried out utilizing the Cadence Virtuoso simulation environment in order to validate their behavior. Further, a brief comparative analysis is executed considering into account the attributes like circuit complexity, supply voltage, firing rate, membrane capacitance, nature of input/output, refractory mechanism, and energy consumption per spike. This work seeks to assist researchers in selecting an appropriate LIF model to efficiently construct memristors and/or non-memristors based SNN for certain application.
{"title":"Simulation-based effective comparative analysis of neuron circuits for neuromorphic computation systems","authors":"Deepthi M.S. , Shashidhara H.R. , Jayaramu Raghu , Rudraswamy S.B.","doi":"10.1016/j.neucom.2024.128758","DOIUrl":"10.1016/j.neucom.2024.128758","url":null,"abstract":"<div><div>The spiking neural networks (SNN) that are inspired by the human brain offers wider scope for application in the growth of neuromorphic computing systems due to their brain level computational capabilities, reduced power consumption, and minimal data movement cost, among other advantages. Spike-based neurons and synapses are the essential building blocks of SNN, and their efficient implementation is vital to their performance enhancement. In this regard, the design and implementation of spiking neurons have been the major focus among the researchers. In this paper, functioning of different leaky integrate fire (LIF)-based spiking neuron circuits like frequency adaptable CMOS-based LIF, resistor-capacitor-based (RC) LIF, and volatile memristor-based LIF are subjected to comparison. The work mainly focuses on revealing analysis of spike duration and amplitude, number of spikes produced during excitation period, threshold operation, field of application, and various other significant parameters of aforementioned neuron circuits. Extensive simulations of these circuits are carried out utilizing the Cadence Virtuoso simulation environment in order to validate their behavior. Further, a brief comparative analysis is executed considering into account the attributes like circuit complexity, supply voltage, firing rate, membrane capacitance, nature of input/output, refractory mechanism, and energy consumption per spike. This work seeks to assist researchers in selecting an appropriate LIF model to efficiently construct memristors and/or non-memristors based SNN for certain application.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128758"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142573219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-28DOI: 10.1016/j.neucom.2024.128791
Kailai Sun , Xinwei Wang , Xi Miao , Qianchuan Zhao
Artificial Intelligence of Things (AIoT) which integrates artificial intelligence (AI) and the Internet of Things (IoT), has attracted increasing attention recently. With the remarkable development of AI, convolutional neural networks (CNN) have achieved great success from research to deployment in many applications. However, deploying complex and state-of-the-art (SOTA) AI models on edge applications is increasingly a big challenge. This paper investigates literature that deploys lightweight CNNs on AI edge devices in practice. We provide a comprehensive analysis of them and many practical suggestions for researchers: how to obtain/design lightweight CNNs, select suitable AI edge devices, and compress and deploy them in practice. Finally, future trends and opportunities are presented, including the deployment of large language models, trustworthy AI and robust deployment.
{"title":"A review of AI edge devices and lightweight CNN and LLM deployment","authors":"Kailai Sun , Xinwei Wang , Xi Miao , Qianchuan Zhao","doi":"10.1016/j.neucom.2024.128791","DOIUrl":"10.1016/j.neucom.2024.128791","url":null,"abstract":"<div><div>Artificial Intelligence of Things (AIoT) which integrates artificial intelligence (AI) and the Internet of Things (IoT), has attracted increasing attention recently. With the remarkable development of AI, convolutional neural networks (CNN) have achieved great success from research to deployment in many applications. However, deploying complex and state-of-the-art (SOTA) AI models on edge applications is increasingly a big challenge. This paper investigates literature that deploys lightweight CNNs on AI edge devices in practice. We provide a comprehensive analysis of them and many practical suggestions for researchers: how to obtain/design lightweight CNNs, select suitable AI edge devices, and compress and deploy them in practice. Finally, future trends and opportunities are presented, including the deployment of large language models, trustworthy AI and robust deployment.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128791"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142573288","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Existing IoU-guided trackers use IoU score to weight the classification score only in testing phase, this model mismatch between training and testing phases leads to poor tracking performance especially when facing background distractors. In this paper, we propose an IoU-guided Siamese network with High-confidence template fusion (SiamIH) for visual tracking. An IoU-guided distractor suppression network is proposed, which uses IoU information to guide classification in training phase and testing phase, and makes the tracking model to suppress background distractors. To cope with appearance variations, we design a high-confidence template fusion network that fuses APCE-based high-confidence template and the initial template to generate more reliable template. Experimental results on OTB2013, OTB2015, UAV123, LaSOT, and GOT10k demonstrate that the proposed SiamIH achieves state-of-the-art tracking performance.
{"title":"IoU-guided Siamese network with high-confidence template fusion for visual tracking","authors":"Zhigang Liu , Hao Huang , Hongyu Dong , Fuyuan Xing","doi":"10.1016/j.neucom.2024.128774","DOIUrl":"10.1016/j.neucom.2024.128774","url":null,"abstract":"<div><div>Existing IoU-guided trackers use IoU score to weight the classification score only in testing phase, this model mismatch between training and testing phases leads to poor tracking performance especially when facing background distractors. In this paper, we propose an IoU-guided Siamese network with High-confidence template fusion (SiamIH) for visual tracking. An IoU-guided distractor suppression network is proposed, which uses IoU information to guide classification in training phase and testing phase, and makes the tracking model to suppress background distractors. To cope with appearance variations, we design a high-confidence template fusion network that fuses APCE-based high-confidence template and the initial template to generate more reliable template. Experimental results on OTB2013, OTB2015, UAV123, LaSOT, and GOT10k demonstrate that the proposed SiamIH achieves state-of-the-art tracking performance.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128774"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142573290","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In recent years, neural network models are used in various tasks. To eliminate privacy concern, differential privacy (DP) is introduced to the training phase of neural network models. However, introducing DP into neural network models is very subtle and error-prone, resulting in that some differentially private neural network models may not achieve privacy guarantee claimed. In this paper, we propose a method, which can audit privacy budget of differentially private neural network models. The proposed method is general and can be used to audit some other AI models. We elaborate on how to audit privacy budget of basic DP mechanisms and neural network models by the proposed method first. Then, we run experiments to verify our method. Experiment results indicate that the proposed method is better than the advanced method and the auditing precise is high when the privacy budget is small. In particular, when auditing privacy budget of ResNet-18 over CIFAR-10 protected by the differentially private mechanism with theoretical privacy budget 0.2, the accuracy of our method is about 17 times that of the state-of-the-art method. For the simpler dataset FMNIST, the accuracy of our method is about 32 times that of the state-of-the-art method when theoretical privacy budget is 0.2.
{"title":"Auditing privacy budget of differentially private neural network models","authors":"Wen Huang , Zhishuo Zhang , Weixin Zhao , Jian Peng , Wenzheng Xu , Yongjian Liao , Shijie Zhou , Ziming Wang","doi":"10.1016/j.neucom.2024.128756","DOIUrl":"10.1016/j.neucom.2024.128756","url":null,"abstract":"<div><div>In recent years, neural network models are used in various tasks. To eliminate privacy concern, differential privacy (DP) is introduced to the training phase of neural network models. However, introducing DP into neural network models is very subtle and error-prone, resulting in that some differentially private neural network models may not achieve privacy guarantee claimed. In this paper, we propose a method, which can audit privacy budget of differentially private neural network models. The proposed method is general and can be used to audit some other AI models. We elaborate on how to audit privacy budget of basic DP mechanisms and neural network models by the proposed method first. Then, we run experiments to verify our method. Experiment results indicate that the proposed method is better than the advanced method and the auditing precise is high when the privacy budget is small. In particular, when auditing privacy budget of ResNet-18 over CIFAR-10 protected by the differentially private mechanism with theoretical privacy budget 0.2, the accuracy of our method is about 17 times that of the state-of-the-art method. For the simpler dataset FMNIST, the accuracy of our method is about 32 times that of the state-of-the-art method when theoretical privacy budget is 0.2.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128756"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142578802","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-28DOI: 10.1016/j.neucom.2024.128778
Dan Luo , Kangfeng Zheng , Chunhua Wu , Xiujuan Wang , Jvjie Wang
Despite their potential, the industrial deployment of large language models (LLMs) is constrained by traditional fine-tuning procedures that are both resource-intensive and time-consuming. Low-Rank Adaptation (LoRA) has emerged as a pioneering methodology for addressing these challenges. By integrating low-rank decomposition matrices into network weights to reduce trainable parameters, LoRA effectively accelerates the adaptation process. While research on LoRA primarily focuses on adjusting low-rank matrices, DyLoRA optimizes the rank-setting mechanism to avoid extensive effort in rank size training and searching. However, DyLoRA rank configuration mechanism has its own limitation. First, DyLoRA sets the same rank size for all the low-rank adaptation layers at each time step. Given that layers with different depth contain distinct information, they should have varying rank values to accurately capture their unique characteristics. Second, the truncated phase selected for ordering representation based on nested dropout regulation is only half dynamic, continuously dropping tail units, thereby limiting its ability to access information. In this work, we propose a novel technique, enhanced range adaptation in time and depth aware dynamic LoRA (ERAT-DLoRA) to address these problems. The ERAT-DLoRA method introduces a dynamic range to the truncated phase that makes the truncated phase fully dynamic. Additionally, we design a time and layer-aware dynamic rank to ensure appropriate adjustments at different time steps and layer levels. We evaluate our solution on natural languages understanding and language generation tasks. Extensive evaluation results demonstrate the effectiveness of the proposed method.
{"title":"ERAT-DLoRA: Parameter-efficient tuning with enhanced range adaptation in time and depth aware dynamic LoRA","authors":"Dan Luo , Kangfeng Zheng , Chunhua Wu , Xiujuan Wang , Jvjie Wang","doi":"10.1016/j.neucom.2024.128778","DOIUrl":"10.1016/j.neucom.2024.128778","url":null,"abstract":"<div><div>Despite their potential, the industrial deployment of large language models (LLMs) is constrained by traditional fine-tuning procedures that are both resource-intensive and time-consuming. Low-Rank Adaptation (LoRA) has emerged as a pioneering methodology for addressing these challenges. By integrating low-rank decomposition matrices into network weights to reduce trainable parameters, LoRA effectively accelerates the adaptation process. While research on LoRA primarily focuses on adjusting low-rank matrices, DyLoRA optimizes the rank-setting mechanism to avoid extensive effort in rank size training and searching. However, DyLoRA rank configuration mechanism has its own limitation. First, DyLoRA sets the same rank size for all the low-rank adaptation layers at each time step. Given that layers with different depth contain distinct information, they should have varying rank values to accurately capture their unique characteristics. Second, the truncated phase selected for ordering representation based on nested dropout regulation is only half dynamic, continuously dropping tail units, thereby limiting its ability to access information. In this work, we propose a novel technique, enhanced range adaptation in time and depth aware dynamic LoRA (ERAT-DLoRA) to address these problems. The ERAT-DLoRA method introduces a dynamic range to the truncated phase that makes the truncated phase fully dynamic. Additionally, we design a time and layer-aware dynamic rank to ensure appropriate adjustments at different time steps and layer levels. We evaluate our solution on natural languages understanding and language generation tasks. Extensive evaluation results demonstrate the effectiveness of the proposed method.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128778"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142586926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-28DOI: 10.1016/j.neucom.2024.128759
Xinkai Sun, Sanguo Zhang
A prominent effect of label noise on neural networks is the disruption of the consistency of predictions. While prior efforts primarily focused on predictions’ consistency at the individual instance level, they often fell short of fully harnessing the consistency across multiple instances. This paper introduces subclass consistency regularization (SCCR) to maximize the potential of this collective consistency of predictions. SCCR mitigates the impact of label noise on neural networks by imposing constraints on the consistency of predictions within each subclass. However, constructing high-quality subclasses poses a formidable challenge, which we formulate as a special clustering problem. To efficiently establish these subclasses, we incorporate a clustering-based contrastive learning framework. Additionally, we introduce the -enhancing algorithm to tailor the contrastive learning framework, ensuring alignment with subclass construction. We conducted comprehensive experiments using benchmark datasets and real datasets to evaluate the effectiveness of our proposed method under various scenarios with differing noise rates. The results unequivocally demonstrate the enhancement in classification accuracy, especially in challenging high-noise settings. Moreover, the refined contrastive learning framework significantly elevates the quality of subclasses even in the presence of noise. Furthermore, we delve into the compatibility of contrastive learning and learning with noisy labels, using the projection head as an illustrative example. This investigation sheds light on an aspect that has hitherto been overlooked in prior research efforts.
{"title":"Subclass consistency regularization for learning with noisy labels based on contrastive learning","authors":"Xinkai Sun, Sanguo Zhang","doi":"10.1016/j.neucom.2024.128759","DOIUrl":"10.1016/j.neucom.2024.128759","url":null,"abstract":"<div><div>A prominent effect of label noise on neural networks is the disruption of the consistency of predictions. While prior efforts primarily focused on predictions’ consistency at the individual instance level, they often fell short of fully harnessing the consistency across multiple instances. This paper introduces subclass consistency regularization (SCCR) to maximize the potential of this collective consistency of predictions. SCCR mitigates the impact of label noise on neural networks by imposing constraints on the consistency of predictions within each subclass. However, constructing high-quality subclasses poses a formidable challenge, which we formulate as a special clustering problem. To efficiently establish these subclasses, we incorporate a clustering-based contrastive learning framework. Additionally, we introduce the <span><math><mi>Q</mi></math></span>-enhancing algorithm to tailor the contrastive learning framework, ensuring alignment with subclass construction. We conducted comprehensive experiments using benchmark datasets and real datasets to evaluate the effectiveness of our proposed method under various scenarios with differing noise rates. The results unequivocally demonstrate the enhancement in classification accuracy, especially in challenging high-noise settings. Moreover, the refined contrastive learning framework significantly elevates the quality of subclasses even in the presence of noise. Furthermore, we delve into the compatibility of contrastive learning and learning with noisy labels, using the projection head as an illustrative example. This investigation sheds light on an aspect that has hitherto been overlooked in prior research efforts.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128759"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142592861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
With the wide application of knowledge graphs, knowledge graph completion has garnered increasing attention in recent years. However, we find that the long tail relation is more common in the KG. These relations typically do not have a large number of triples for training and are referred to as few-shot relations. The knowledge graph completion in the few-shot scenario is a major challenge currently. The current mainstream knowledge graph completion algorithms have the following drawbacks. The metric-based methods lack interpretability of results, while the algorithms based on path interaction are not suitable for few-shot scenarios and the availability of the model is limited in sparse knowledge graphs. In this paper, we propose a one-shot relational learning of knowledge graphs completion based on multi-hop information enhancement(MHEC). Firstly, MHEC extracts entity concepts from multi-hop paths to obtain task related entity concepts and filters out noisy neighbor attributes. Then, MHEC combines multi-hop path information between head and tail to represent entity pairs. Compared to previous completion methods that only consider structural features of entities, MHEC considers the reasoning logic between entity pairs, which not only includes structural features but also possesses rich semantic features. Next, MHEC introduces a reasoning process in the completion task to address the issues of lack of interpretability in the one-shot scenario. In addition, to improve completion and reasoning quality in sparse knowledge graphs, MHEC utilizes contrastive learning to enhance pre-training vector representations of entities and relations and proposes a matching processor that leverages the semantic information of pre-training vectors to assist the reasoning model in expanding the multi-hop paths. Experiments demonstrate that MHEC outperforms the state-of-the-art completion techniques on real-world datasets NELL-One and FB15k237-One.
{"title":"MHEC: One-shot relational learning of knowledge graphs completion based on multi-hop information enhancement","authors":"Ruixin Ma, Buyun Gao, Weihe Wang, Longfei Wang, Xiaoru Wang, Liang Zhao","doi":"10.1016/j.neucom.2024.128760","DOIUrl":"10.1016/j.neucom.2024.128760","url":null,"abstract":"<div><div>With the wide application of knowledge graphs, knowledge graph completion has garnered increasing attention in recent years. However, we find that the long tail relation is more common in the KG. These relations typically do not have a large number of triples for training and are referred to as few-shot relations. The knowledge graph completion in the few-shot scenario is a major challenge currently. The current mainstream knowledge graph completion algorithms have the following drawbacks. The metric-based methods lack interpretability of results, while the algorithms based on path interaction are not suitable for few-shot scenarios and the availability of the model is limited in sparse knowledge graphs. In this paper, we propose a one-shot relational learning of knowledge graphs completion based on multi-hop information enhancement(MHEC). Firstly, MHEC extracts entity concepts from multi-hop paths to obtain task related entity concepts and filters out noisy neighbor attributes. Then, MHEC combines multi-hop path information between head and tail to represent entity pairs. Compared to previous completion methods that only consider structural features of entities, MHEC considers the reasoning logic between entity pairs, which not only includes structural features but also possesses rich semantic features. Next, MHEC introduces a reasoning process in the completion task to address the issues of lack of interpretability in the one-shot scenario. In addition, to improve completion and reasoning quality in sparse knowledge graphs, MHEC utilizes contrastive learning to enhance pre-training vector representations of entities and relations and proposes a matching processor that leverages the semantic information of pre-training vectors to assist the reasoning model in expanding the multi-hop paths. Experiments demonstrate that MHEC outperforms the state-of-the-art completion techniques on real-world datasets NELL-One and FB15k237-One.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128760"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142573297","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-10-28DOI: 10.1016/j.neucom.2024.128770
Tiechao Wang, Hongyang Zhang, Shuai Sui
This paper proposes an adaptive neural network event-triggered and quantized output feedback control scheme for quarter vehicle active suspensions with actuator saturation. The scheme uses neural networks to approximate the unknown parts of the active suspension. When the system states of the suspension are not entirely available, a state observer is designed to estimate the unknown states. The measurable system states, partially estimated observer states, neural network weights, and a filtered virtual control are sequentially event-triggered, quantified, and transmitted to the controller via in-vehicle networks. The problem of non-differentiable virtual control is solved using dynamic surface control technology in the backstepping quantized control design. Integrating a Gaussian error function and a first-order auxiliary subsystem compensates for the nonlinearity caused by asymmetric saturation. Theoretical analysis proves that all error signals of the closed-loop active suspension system are semi-globally uniformly ultimately bounded, and the Zeno phenomenon can be ruled out. Simulation results validate the effectiveness of the proposed control method.
{"title":"Observer-based adaptive neural network event-triggered quantized control for active suspensions with actuator saturation","authors":"Tiechao Wang, Hongyang Zhang, Shuai Sui","doi":"10.1016/j.neucom.2024.128770","DOIUrl":"10.1016/j.neucom.2024.128770","url":null,"abstract":"<div><div>This paper proposes an adaptive neural network event-triggered and quantized output feedback control scheme for quarter vehicle active suspensions with actuator saturation. The scheme uses neural networks to approximate the unknown parts of the active suspension. When the system states of the suspension are not entirely available, a state observer is designed to estimate the unknown states. The measurable system states, partially estimated observer states, neural network weights, and a filtered virtual control are sequentially event-triggered, quantified, and transmitted to the controller via in-vehicle networks. The problem of non-differentiable virtual control is solved using dynamic surface control technology in the backstepping quantized control design. Integrating a Gaussian error function and a first-order auxiliary subsystem compensates for the nonlinearity caused by asymmetric saturation. Theoretical analysis proves that all error signals of the closed-loop active suspension system are semi-globally uniformly ultimately bounded, and the Zeno phenomenon can be ruled out. Simulation results validate the effectiveness of the proposed control method.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"614 ","pages":"Article 128770"},"PeriodicalIF":5.5,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142573193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}