Chuxiong Sun , Jie Hu , Hongming Gu , Jinpeng Chen , Wei Liang , Mingchuan Yang
{"title":"Scalable and Adaptive Graph Neural Networks with Self-Label-Enhanced Training","authors":"Chuxiong Sun , Jie Hu , Hongming Gu , Jinpeng Chen , Wei Liang , Mingchuan Yang","doi":"10.1016/j.patcog.2024.111210","DOIUrl":null,"url":null,"abstract":"<div><div>Although GNNs have achieved success in semi-supervised graph learning tasks, common GNNs suffer from expensive message passing during each epoch and the exponentially growing receptive field occupying too much memory, especially on large graphs. Neighbor sampling techniques can reduce GNNs’ memory footprints, but they encounter either redundant computation or incomplete edges. Some simplified GNNs decouple graph convolutions and feature transformations to reduce computation in training. However, only a part of them can scale to large graphs without neighbor sampling techniques, which can be concluded as decoupled GNNs. Nevertheless, they either only utilize the last convolution output or simply add multi-hop features with uniform weights, which limits their expressiveness. In this paper, we refine the pipeline of decoupled GNNs and propose Scalable and Adaptive Graph Neural Networks (SAGN), which effectively leverages multi-hop information with a scalable attention mechanism. Moreover, we generalize the input of decoupled GNNs to view another classical technique, label propagation, as a special case of decoupled GNNs and propose decoupled label trick (DecLT) to incorporate label information into decoupled GNNs. Furthermore, by incorporating self-training technique, we further propose the Self-Label-Enhanced (SLE) training framework, leveraging pseudo labels to simultaneously augment the training set and improve label propagation. Extensive experiments show that SAGN outperforms other baselines, and that DecLT and SLE can consistently and significantly improve all types of models on semi-supervised node classification tasks. Many top-ranked models on Open Graph Benchmark (OGB) leaderboard adopt our methods as the main backbone.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"160 ","pages":"Article 111210"},"PeriodicalIF":7.5000,"publicationDate":"2024-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0031320324009610","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Although GNNs have achieved success in semi-supervised graph learning tasks, common GNNs suffer from expensive message passing during each epoch and the exponentially growing receptive field occupying too much memory, especially on large graphs. Neighbor sampling techniques can reduce GNNs’ memory footprints, but they encounter either redundant computation or incomplete edges. Some simplified GNNs decouple graph convolutions and feature transformations to reduce computation in training. However, only a part of them can scale to large graphs without neighbor sampling techniques, which can be concluded as decoupled GNNs. Nevertheless, they either only utilize the last convolution output or simply add multi-hop features with uniform weights, which limits their expressiveness. In this paper, we refine the pipeline of decoupled GNNs and propose Scalable and Adaptive Graph Neural Networks (SAGN), which effectively leverages multi-hop information with a scalable attention mechanism. Moreover, we generalize the input of decoupled GNNs to view another classical technique, label propagation, as a special case of decoupled GNNs and propose decoupled label trick (DecLT) to incorporate label information into decoupled GNNs. Furthermore, by incorporating self-training technique, we further propose the Self-Label-Enhanced (SLE) training framework, leveraging pseudo labels to simultaneously augment the training set and improve label propagation. Extensive experiments show that SAGN outperforms other baselines, and that DecLT and SLE can consistently and significantly improve all types of models on semi-supervised node classification tasks. Many top-ranked models on Open Graph Benchmark (OGB) leaderboard adopt our methods as the main backbone.
期刊介绍:
The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.