Tiesunlong Shen, Erik Cambria, Jin Wang, Yi Cai, Xuejie Zhang
{"title":"Insight at the right spot: Provide decisive subgraph information to Graph LLM with reinforcement learning","authors":"Tiesunlong Shen, Erik Cambria, Jin Wang, Yi Cai, Xuejie Zhang","doi":"10.1016/j.inffus.2024.102860","DOIUrl":null,"url":null,"abstract":"Large language models (LLMs) cannot see or understand graphs. The current Graph LLM method transform graph structures into a format LLMs understands, utilizing LLM as a predictor to perform graph-learning task. However, these approaches have underperformed in graph-learning tasks. The issues arise because these methods typically rely on a fixed neighbor hop count for the target node set by expert experience, limiting the LLM’s access to only a certain range of neighbor information. Due to the black-box nature of LLM, it is challenging to determine which specific pieces of neighborhood information can effectively assist LLMs in making accurate inferences, which prevents LLMs from generating correct inferences. This study proposes to assist LLM in gaining insight at the right <ce:bold><ce:italic>s</ce:italic></ce:bold>pot by <ce:bold><ce:italic>p</ce:italic></ce:bold>rov<ce:bold><ce:italic>i</ce:italic></ce:bold>ding <ce:bold><ce:italic>de</ce:italic></ce:bold>cisive subgraph information to Graph LLM with <ce:bold><ce:italic>r</ce:italic></ce:bold>einforcement learning (<ce:bold><ce:italic>Spider</ce:italic></ce:bold>). A reinforcement subgraph detection module was designed to search for essential neighborhoods that influence LLM’s predictions. A decisive node-guided network was then applied to guide the reinforcement subgraph network, allowing LLMs to rely more on crucial nodes within the essential neighborhood for predictions. Essential neighborhood and decisive node information are provided to LLM in text form without the requirement of retraining. Experiments on five graph learning datasets demonstrate the effectiveness of the proposed model against all baselines, including GNN and LLM methods.","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"50 5 1","pages":""},"PeriodicalIF":14.7000,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1016/j.inffus.2024.102860","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Large language models (LLMs) cannot see or understand graphs. The current Graph LLM method transform graph structures into a format LLMs understands, utilizing LLM as a predictor to perform graph-learning task. However, these approaches have underperformed in graph-learning tasks. The issues arise because these methods typically rely on a fixed neighbor hop count for the target node set by expert experience, limiting the LLM’s access to only a certain range of neighbor information. Due to the black-box nature of LLM, it is challenging to determine which specific pieces of neighborhood information can effectively assist LLMs in making accurate inferences, which prevents LLMs from generating correct inferences. This study proposes to assist LLM in gaining insight at the right spot by providing decisive subgraph information to Graph LLM with reinforcement learning (Spider). A reinforcement subgraph detection module was designed to search for essential neighborhoods that influence LLM’s predictions. A decisive node-guided network was then applied to guide the reinforcement subgraph network, allowing LLMs to rely more on crucial nodes within the essential neighborhood for predictions. Essential neighborhood and decisive node information are provided to LLM in text form without the requirement of retraining. Experiments on five graph learning datasets demonstrate the effectiveness of the proposed model against all baselines, including GNN and LLM methods.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.