AAGCN: An adaptive data augmentation for graph contrastive learning

IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE Pattern Recognition Pub Date : 2025-02-22 DOI:10.1016/j.patcog.2025.111471
Peng Qin , Yaochun Lu , Weifu Chen , Defang Li , Guocan Feng
{"title":"AAGCN: An adaptive data augmentation for graph contrastive learning","authors":"Peng Qin ,&nbsp;Yaochun Lu ,&nbsp;Weifu Chen ,&nbsp;Defang Li ,&nbsp;Guocan Feng","doi":"10.1016/j.patcog.2025.111471","DOIUrl":null,"url":null,"abstract":"<div><div>Contrastive learning has achieved great success in many applications. A key step in contrastive learning is to find a positive sample and negative samples. Traditional methods find the positive sample by choosing the most similar sample. A more popular approach is to do data augmentation where the original data and the augmented data are naturally treated as positive pairs. It is easy for grid data to do the augmentation, for example, we can rotate, crop or color an image to get augmented images. But it is challenging to do augmentation for graph data, due to non-Euclidean nature of graphs. Current graph augmentation methods mainly focus on masking nodes, dropping edges, or extracting subgraphs. Such methods lack of flexibility and require intensive manual settings. In this work, we propose a model called <em>Adaptive Augmentation Graph Convolutional Network (AAGCN)</em> for semi-supervised node classification, based on adaptive graph augmentation. Rather than choose a probability distribution, for example, Bernoulli distribution, to drop some of the nodes or edges in Dropout, the proposed model learns the mask matrices for nodes or edges adaptively. Experiments on citation networks such as Cora, CiteSeer and Cora-ML show that AAGCN achieved state-of-the-art performance compared with other popular graph neural networks. The proposed model was also tested on a more challenging and large-scale graph dataset, OGBN-Arxiv, which has 169,343 nodes and 1,166,243 edges. The proposed model could still achieve competitive prediction results.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"163 ","pages":"Article 111471"},"PeriodicalIF":7.6000,"publicationDate":"2025-02-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0031320325001311","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Contrastive learning has achieved great success in many applications. A key step in contrastive learning is to find a positive sample and negative samples. Traditional methods find the positive sample by choosing the most similar sample. A more popular approach is to do data augmentation where the original data and the augmented data are naturally treated as positive pairs. It is easy for grid data to do the augmentation, for example, we can rotate, crop or color an image to get augmented images. But it is challenging to do augmentation for graph data, due to non-Euclidean nature of graphs. Current graph augmentation methods mainly focus on masking nodes, dropping edges, or extracting subgraphs. Such methods lack of flexibility and require intensive manual settings. In this work, we propose a model called Adaptive Augmentation Graph Convolutional Network (AAGCN) for semi-supervised node classification, based on adaptive graph augmentation. Rather than choose a probability distribution, for example, Bernoulli distribution, to drop some of the nodes or edges in Dropout, the proposed model learns the mask matrices for nodes or edges adaptively. Experiments on citation networks such as Cora, CiteSeer and Cora-ML show that AAGCN achieved state-of-the-art performance compared with other popular graph neural networks. The proposed model was also tested on a more challenging and large-scale graph dataset, OGBN-Arxiv, which has 169,343 nodes and 1,166,243 edges. The proposed model could still achieve competitive prediction results.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
图对比学习的自适应数据增强
对比学习在许多应用中取得了巨大的成功。对比学习的关键步骤是找到正样本和负样本。传统的方法是选择最相似的样本来寻找阳性样本。一种更流行的方法是进行数据扩充,其中原始数据和扩充后的数据自然地被视为正对。对网格数据进行增强很容易,例如,我们可以旋转、裁剪或着色图像来获得增强图像。但由于图的非欧几里得性质,对图数据进行增广是一个挑战。目前的图增强方法主要集中在屏蔽节点、掉边或提取子图。这种方法缺乏灵活性,需要大量的手动设置。在这项工作中,我们提出了一种基于自适应图增强的半监督节点分类模型,称为自适应增强图卷积网络(AAGCN)。该模型不是选择一个概率分布(如伯努利分布)来删除Dropout中的一些节点或边,而是自适应地学习节点或边的掩码矩阵。在引用网络(如Cora、CiteSeer和Cora- ml)上的实验表明,与其他流行的图神经网络相比,AAGCN取得了最先进的性能。该模型还在一个更具挑战性和更大规模的图数据集OGBN-Arxiv上进行了测试,该数据集有169,343个节点和1,166,243条边。所提出的模型仍然可以获得具有竞争力的预测结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Pattern Recognition
Pattern Recognition 工程技术-工程:电子与电气
CiteScore
14.40
自引率
16.20%
发文量
683
审稿时长
5.6 months
期刊介绍: The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.
期刊最新文献
IrisMAE: Structure-aware masked image modeling for iris recognition Minimizing the pretraining gap: Domain-aligned text-based person retrieval Stealthy backdoor attack method targeting group fairness in self-supervised learning Single-domain generalization for fastener detection via sample reconstruction and class-wise domain contrast EdgeFusionNet: Edge information-guided small object detection for remote sensing images
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1