Meng Qin, Chaorui Zhang, Yu Gao, Yibin Ding, Weipeng Jiang, Weixi Zhang, Wei Han, Bo Bai
{"title":"Towards Faster Graph Partitioning via Pre-training and Inductive Inference","authors":"Meng Qin, Chaorui Zhang, Yu Gao, Yibin Ding, Weipeng Jiang, Weixi Zhang, Wei Han, Bo Bai","doi":"arxiv-2409.00670","DOIUrl":null,"url":null,"abstract":"Graph partitioning (GP) is a classic problem that divides the node set of a\ngraph into densely-connected blocks. Following the IEEE HPEC Graph Challenge\nand recent advances in pre-training techniques (e.g., large-language models),\nwe propose PR-GPT (Pre-trained & Refined Graph ParTitioning) based on a novel\npre-training & refinement paradigm. We first conduct the offline pre-training\nof a deep graph learning (DGL) model on small synthetic graphs with various\ntopology properties. By using the inductive inference of DGL, one can directly\ngeneralize the pre-trained model (with frozen model parameters) to large graphs\nand derive feasible GP results. We also use the derived partition as a good\ninitialization of an efficient GP method (e.g., InfoMap) to further refine the\nquality of partitioning. In this setting, the online generalization and\nrefinement of PR-GPT can not only benefit from the transfer ability regarding\nquality but also ensure high inference efficiency without re-training. Based on\na mechanism of reducing the scale of a graph to be processed by the refinement\nmethod, PR-GPT also has the potential to support streaming GP. Experiments on\nthe Graph Challenge benchmark demonstrate that PR-GPT can ensure faster GP on\nlarge-scale graphs without significant quality degradation, compared with\nrunning a refinement method from scratch. We will make our code public at\nhttps://github.com/KuroginQin/PRGPT.","PeriodicalId":501032,"journal":{"name":"arXiv - CS - Social and Information Networks","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Social and Information Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.00670","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Graph partitioning (GP) is a classic problem that divides the node set of a
graph into densely-connected blocks. Following the IEEE HPEC Graph Challenge
and recent advances in pre-training techniques (e.g., large-language models),
we propose PR-GPT (Pre-trained & Refined Graph ParTitioning) based on a novel
pre-training & refinement paradigm. We first conduct the offline pre-training
of a deep graph learning (DGL) model on small synthetic graphs with various
topology properties. By using the inductive inference of DGL, one can directly
generalize the pre-trained model (with frozen model parameters) to large graphs
and derive feasible GP results. We also use the derived partition as a good
initialization of an efficient GP method (e.g., InfoMap) to further refine the
quality of partitioning. In this setting, the online generalization and
refinement of PR-GPT can not only benefit from the transfer ability regarding
quality but also ensure high inference efficiency without re-training. Based on
a mechanism of reducing the scale of a graph to be processed by the refinement
method, PR-GPT also has the potential to support streaming GP. Experiments on
the Graph Challenge benchmark demonstrate that PR-GPT can ensure faster GP on
large-scale graphs without significant quality degradation, compared with
running a refinement method from scratch. We will make our code public at
https://github.com/KuroginQin/PRGPT.