Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities

Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Neural networks 2022-01, Vol.145, p.356-373
Hauptverfasser: Itoh, Takeshi D., Kubo, Takatomi, Ikeda, Kazushi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 373
container_issue
container_start_page 356
container_title Neural networks
container_volume 145
creator Itoh, Takeshi D.
Kubo, Takatomi
Ikeda, Kazushi
description Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.
doi_str_mv 10.1016/j.neunet.2021.11.001
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2601480583</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608021004299</els_id><sourcerecordid>2601480583</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</originalsourceid><addsrcrecordid>eNp9kDFv1TAUhS0Eoo_CP0DII0vCtZM4DgMSqoBWKmKhs-XYN60ffnGwnVb99zhKYWS6g79zjvwR8pZBzYCJD8d6xnXGXHPgrGasBmDPyIHJfqh4L_lzcgA5NJUACWfkVUpHABCybV6Ss6aVIDvZH8jyffXZVR7v0VOdM87ZhZkuIXg339IpRHob9XJHy1jUvpz8EOKv9JHezG563Jj9PeISMZW43goSfXD5jp628sUj9cFo77LD9Jq8mLRP-ObpnpObr19-XlxW1z--XV18vq5MCzJXZkTbtTBKMfR6AGRWAvIBOY6MCyts36PoRGfattXY6AlggIYLZodxsgDNOXm_9y4x_F4xZXVyyaD3esawJsUFsGKhk01B2x01MaQUcVJLdCcdHxUDtblWR7W7VptrxZgqrkvs3dPCOp7Q_gv9lVuATzuA5Z_3DqNKxuFs0LqIJisb3P8X_gDtmJTE</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2601480583</pqid></control><display><type>article</type><title>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</title><source>MEDLINE</source><source>Elsevier ScienceDirect Journals</source><creator>Itoh, Takeshi D. ; Kubo, Takatomi ; Ikeda, Kazushi</creator><creatorcontrib>Itoh, Takeshi D. ; Kubo, Takatomi ; Ikeda, Kazushi</creatorcontrib><description>Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/j.neunet.2021.11.001</identifier><identifier>PMID: 34808587</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Attention ; Graph neural network (GNN) ; Graph representation learning (GRL) ; Learning ; Multi-level attention pooling (MLAP) ; Multi-level locality ; Neural Networks, Computer</subject><ispartof>Neural networks, 2022-01, Vol.145, p.356-373</ispartof><rights>2021 The Author(s)</rights><rights>Copyright © 2021 The Author(s). Published by Elsevier Ltd.. All rights reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</citedby><cites>FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</cites><orcidid>0000-0002-4021-1481 ; 0000-0003-0192-7177</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0893608021004299$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/34808587$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Itoh, Takeshi D.</creatorcontrib><creatorcontrib>Kubo, Takatomi</creatorcontrib><creatorcontrib>Ikeda, Kazushi</creatorcontrib><title>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.</description><subject>Attention</subject><subject>Graph neural network (GNN)</subject><subject>Graph representation learning (GRL)</subject><subject>Learning</subject><subject>Multi-level attention pooling (MLAP)</subject><subject>Multi-level locality</subject><subject>Neural Networks, Computer</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNp9kDFv1TAUhS0Eoo_CP0DII0vCtZM4DgMSqoBWKmKhs-XYN60ffnGwnVb99zhKYWS6g79zjvwR8pZBzYCJD8d6xnXGXHPgrGasBmDPyIHJfqh4L_lzcgA5NJUACWfkVUpHABCybV6Ss6aVIDvZH8jyffXZVR7v0VOdM87ZhZkuIXg339IpRHob9XJHy1jUvpz8EOKv9JHezG563Jj9PeISMZW43goSfXD5jp628sUj9cFo77LD9Jq8mLRP-ObpnpObr19-XlxW1z--XV18vq5MCzJXZkTbtTBKMfR6AGRWAvIBOY6MCyts36PoRGfattXY6AlggIYLZodxsgDNOXm_9y4x_F4xZXVyyaD3esawJsUFsGKhk01B2x01MaQUcVJLdCcdHxUDtblWR7W7VptrxZgqrkvs3dPCOp7Q_gv9lVuATzuA5Z_3DqNKxuFs0LqIJisb3P8X_gDtmJTE</recordid><startdate>202201</startdate><enddate>202201</enddate><creator>Itoh, Takeshi D.</creator><creator>Kubo, Takatomi</creator><creator>Ikeda, Kazushi</creator><general>Elsevier Ltd</general><scope>6I.</scope><scope>AAFTH</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-4021-1481</orcidid><orcidid>https://orcid.org/0000-0003-0192-7177</orcidid></search><sort><creationdate>202201</creationdate><title>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</title><author>Itoh, Takeshi D. ; Kubo, Takatomi ; Ikeda, Kazushi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Attention</topic><topic>Graph neural network (GNN)</topic><topic>Graph representation learning (GRL)</topic><topic>Learning</topic><topic>Multi-level attention pooling (MLAP)</topic><topic>Multi-level locality</topic><topic>Neural Networks, Computer</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Itoh, Takeshi D.</creatorcontrib><creatorcontrib>Kubo, Takatomi</creatorcontrib><creatorcontrib>Ikeda, Kazushi</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Itoh, Takeshi D.</au><au>Kubo, Takatomi</au><au>Ikeda, Kazushi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2022-01</date><risdate>2022</risdate><volume>145</volume><spage>356</spage><epage>373</epage><pages>356-373</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>34808587</pmid><doi>10.1016/j.neunet.2021.11.001</doi><tpages>18</tpages><orcidid>https://orcid.org/0000-0002-4021-1481</orcidid><orcidid>https://orcid.org/0000-0003-0192-7177</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0893-6080
ispartof Neural networks, 2022-01, Vol.145, p.356-373
issn 0893-6080
1879-2782
language eng
recordid cdi_proquest_miscellaneous_2601480583
source MEDLINE; Elsevier ScienceDirect Journals
subjects Attention
Graph neural network (GNN)
Graph representation learning (GRL)
Learning
Multi-level attention pooling (MLAP)
Multi-level locality
Neural Networks, Computer
title Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T22%3A52%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multi-level%20attention%20pooling%20for%20graph%20neural%20networks:%20Unifying%20graph%20representations%20with%20multiple%20localities&rft.jtitle=Neural%20networks&rft.au=Itoh,%20Takeshi%20D.&rft.date=2022-01&rft.volume=145&rft.spage=356&rft.epage=373&rft.pages=356-373&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/j.neunet.2021.11.001&rft_dat=%3Cproquest_cross%3E2601480583%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2601480583&rft_id=info:pmid/34808587&rft_els_id=S0893608021004299&rfr_iscdi=true