DHSEGATs: distance and hop-wise structures encoding enhanced graph attention networks
Numerous works prove that existing neighbor-averag-ing graph neural networks (GNNs) cannot efficiently catch struc-ture features,and many works show that injecting structure,dis-tance,position,or spatial features can significantly improve the performance of GNNs,however,injecting high-level structur...
Gespeichert in:
Veröffentlicht in: | Journal of systems engineering and electronics 2023-04, Vol.34 (2), p.350-359 |
---|---|
1. Verfasser: | |
Format: | Artikel |
Sprache: | eng |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Numerous works prove that existing neighbor-averag-ing graph neural networks (GNNs) cannot efficiently catch struc-ture features,and many works show that injecting structure,dis-tance,position,or spatial features can significantly improve the performance of GNNs,however,injecting high-level structure and distance into GNNs is an intuitive but untouched idea. This work sheds light on this issue and proposes a scheme to enhance graph attention networks (GATs) by encoding distance and hop-wise structure statistics. Firstly,the hop-wise structure and distributional distance information are extracted based on several hop-wise ego-nets of every target node. Secondly,the derived structure information,distance information,and intrinsic features are encoded into the same vector space and then added together to get initial embedding vectors. Thirdly,the derived embedding vectors are fed into GATs,such as GAT and adaptive graph diffusion network (AGDN) to get the soft labels. Fourthly,the soft labels are fed into correct and smooth (C&S) to conduct label propagation and get final predictions. Experi-ments show that the distance and hop-wise structures encoding enhanced graph attention networks (DHSEGATs) achieve a com-petitive result. |
---|---|
ISSN: | 1004-4132 1004-4132 |
DOI: | 10.23919/JSEE.2023.000057 |