MIGP: Metapath Integrated Graph Prompt Neural Network

Graph neural networks (GNNs) leveraging metapaths have garnered extensive utilization. Nevertheless, the escalating parameters and data corpus within graph pre-training models incur mounting training costs. Consequently, GNN models encounter hurdles including diminished generalization capacity and c...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Neural networks 2024-11, Vol.179, p.106595, Article 106595
Hauptverfasser: Lai, Pei-Yuan, Dai, Qing-Yun, Lu, Yi-Hong, Wang, Zeng-Hui, Chen, Man-Sheng, Wang, Chang-Dong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Graph neural networks (GNNs) leveraging metapaths have garnered extensive utilization. Nevertheless, the escalating parameters and data corpus within graph pre-training models incur mounting training costs. Consequently, GNN models encounter hurdles including diminished generalization capacity and compromised performance amidst small sample datasets. Drawing inspiration from the efficacy demonstrated by self-supervised learning methodologies in natural language processing, we embark on an exploration. We endeavor to imbue graph data with augmentable, learnable prompt vectors targeting node representation enhancement to foster superior adaptability to downstream tasks. This paper proposes a novel approach, the Metapath Integrated Graph Prompt Neural Network (MIGP), which leverages learnable prompt vectors to enhance node representations within a pretrained model framework. By leveraging learnable prompt vectors, MIGP aims to address the limitations posed by mall sample datasets and improve GNNs’ model generalization. In the pretraining stage, we split symmetric metapaths in heterogeneous graphs into short metapaths and explicitly propagate information along the metapaths to update node representations. In the prompt-tuning stage, the parameters of the pretrained model are fixed, a set of independent basis vectors is introduced, and an attention mechanism is employed to generate task-specific learnable prompt vectors for each node. Another notable contribution of our work is the introduction of three patent datasets, which is a pioneering application in related fields. We will make these three patent datasets publicly available to facilitate further research on large-scale patent data analysis. Through comprehensive experiments conducted on three patent datasets and three other public datasets, i.e., ACM, IMDB, and DBLP, we demonstrate the superior performance of the MIGP model in enhancing model applicability and performance across a variety of downstream datasets. The source code and datasets are available in the website.11https://github.com/hzw-ai/MIGP, with password: MIGPNN2024.
ISSN:0893-6080
1879-2782
1879-2782
DOI:10.1016/j.neunet.2024.106595