Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities
Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this...
Gespeichert in:
Veröffentlicht in: | Neural networks 2022-01, Vol.145, p.356-373 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 373 |
---|---|
container_issue | |
container_start_page | 356 |
container_title | Neural networks |
container_volume | 145 |
creator | Itoh, Takeshi D. Kubo, Takatomi Ikeda, Kazushi |
description | Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations. |
doi_str_mv | 10.1016/j.neunet.2021.11.001 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2601480583</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0893608021004299</els_id><sourcerecordid>2601480583</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</originalsourceid><addsrcrecordid>eNp9kDFv1TAUhS0Eoo_CP0DII0vCtZM4DgMSqoBWKmKhs-XYN60ffnGwnVb99zhKYWS6g79zjvwR8pZBzYCJD8d6xnXGXHPgrGasBmDPyIHJfqh4L_lzcgA5NJUACWfkVUpHABCybV6Ss6aVIDvZH8jyffXZVR7v0VOdM87ZhZkuIXg339IpRHob9XJHy1jUvpz8EOKv9JHezG563Jj9PeISMZW43goSfXD5jp628sUj9cFo77LD9Jq8mLRP-ObpnpObr19-XlxW1z--XV18vq5MCzJXZkTbtTBKMfR6AGRWAvIBOY6MCyts36PoRGfattXY6AlggIYLZodxsgDNOXm_9y4x_F4xZXVyyaD3esawJsUFsGKhk01B2x01MaQUcVJLdCcdHxUDtblWR7W7VptrxZgqrkvs3dPCOp7Q_gv9lVuATzuA5Z_3DqNKxuFs0LqIJisb3P8X_gDtmJTE</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2601480583</pqid></control><display><type>article</type><title>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</title><source>MEDLINE</source><source>Elsevier ScienceDirect Journals</source><creator>Itoh, Takeshi D. ; Kubo, Takatomi ; Ikeda, Kazushi</creator><creatorcontrib>Itoh, Takeshi D. ; Kubo, Takatomi ; Ikeda, Kazushi</creatorcontrib><description>Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.</description><identifier>ISSN: 0893-6080</identifier><identifier>EISSN: 1879-2782</identifier><identifier>DOI: 10.1016/j.neunet.2021.11.001</identifier><identifier>PMID: 34808587</identifier><language>eng</language><publisher>United States: Elsevier Ltd</publisher><subject>Attention ; Graph neural network (GNN) ; Graph representation learning (GRL) ; Learning ; Multi-level attention pooling (MLAP) ; Multi-level locality ; Neural Networks, Computer</subject><ispartof>Neural networks, 2022-01, Vol.145, p.356-373</ispartof><rights>2021 The Author(s)</rights><rights>Copyright © 2021 The Author(s). Published by Elsevier Ltd.. All rights reserved.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</citedby><cites>FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</cites><orcidid>0000-0002-4021-1481 ; 0000-0003-0192-7177</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S0893608021004299$$EHTML$$P50$$Gelsevier$$Hfree_for_read</linktohtml><link.rule.ids>314,776,780,3537,27901,27902,65306</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/34808587$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Itoh, Takeshi D.</creatorcontrib><creatorcontrib>Kubo, Takatomi</creatorcontrib><creatorcontrib>Ikeda, Kazushi</creatorcontrib><title>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</title><title>Neural networks</title><addtitle>Neural Netw</addtitle><description>Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.</description><subject>Attention</subject><subject>Graph neural network (GNN)</subject><subject>Graph representation learning (GRL)</subject><subject>Learning</subject><subject>Multi-level attention pooling (MLAP)</subject><subject>Multi-level locality</subject><subject>Neural Networks, Computer</subject><issn>0893-6080</issn><issn>1879-2782</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNp9kDFv1TAUhS0Eoo_CP0DII0vCtZM4DgMSqoBWKmKhs-XYN60ffnGwnVb99zhKYWS6g79zjvwR8pZBzYCJD8d6xnXGXHPgrGasBmDPyIHJfqh4L_lzcgA5NJUACWfkVUpHABCybV6Ss6aVIDvZH8jyffXZVR7v0VOdM87ZhZkuIXg339IpRHob9XJHy1jUvpz8EOKv9JHezG563Jj9PeISMZW43goSfXD5jp628sUj9cFo77LD9Jq8mLRP-ObpnpObr19-XlxW1z--XV18vq5MCzJXZkTbtTBKMfR6AGRWAvIBOY6MCyts36PoRGfattXY6AlggIYLZodxsgDNOXm_9y4x_F4xZXVyyaD3esawJsUFsGKhk01B2x01MaQUcVJLdCcdHxUDtblWR7W7VptrxZgqrkvs3dPCOp7Q_gv9lVuATzuA5Z_3DqNKxuFs0LqIJisb3P8X_gDtmJTE</recordid><startdate>202201</startdate><enddate>202201</enddate><creator>Itoh, Takeshi D.</creator><creator>Kubo, Takatomi</creator><creator>Ikeda, Kazushi</creator><general>Elsevier Ltd</general><scope>6I.</scope><scope>AAFTH</scope><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-4021-1481</orcidid><orcidid>https://orcid.org/0000-0003-0192-7177</orcidid></search><sort><creationdate>202201</creationdate><title>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</title><author>Itoh, Takeshi D. ; Kubo, Takatomi ; Ikeda, Kazushi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-cbed540b8697a90e1d80e29e2eb126d6d77e6565c444ae3af00903261d9bfd003</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Attention</topic><topic>Graph neural network (GNN)</topic><topic>Graph representation learning (GRL)</topic><topic>Learning</topic><topic>Multi-level attention pooling (MLAP)</topic><topic>Multi-level locality</topic><topic>Neural Networks, Computer</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Itoh, Takeshi D.</creatorcontrib><creatorcontrib>Kubo, Takatomi</creatorcontrib><creatorcontrib>Ikeda, Kazushi</creatorcontrib><collection>ScienceDirect Open Access Titles</collection><collection>Elsevier:ScienceDirect:Open Access</collection><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Neural networks</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Itoh, Takeshi D.</au><au>Kubo, Takatomi</au><au>Ikeda, Kazushi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities</atitle><jtitle>Neural networks</jtitle><addtitle>Neural Netw</addtitle><date>2022-01</date><risdate>2022</risdate><volume>145</volume><spage>356</spage><epage>373</epage><pages>356-373</pages><issn>0893-6080</issn><eissn>1879-2782</eissn><abstract>Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data and achieved better task performance than conventional methods. The foundation of GNNs is the message passing procedure, which propagates the information in a node to its neighbors. Since this procedure proceeds one step per layer, the range of the information propagation among nodes is small in the lower layers, and it expands toward the higher layers. Therefore, a GNN model has to be deep enough to capture global structural information in a graph. On the other hand, it is known that deep GNN models suffer from performance degradation because they lose nodes’ local information, which would be essential for good model performance, through many message passing steps. In this study, we propose multi-level attention pooling (MLAP) for graph-level classification tasks, which can adapt to both local and global structural information in a graph. It has an attention pooling layer for each message passing step and computes the final graph representation by unifying the layer-wise graph representations. The MLAP architecture allows models to utilize the structural information of graphs with multiple levels of localities because it preserves layer-wise information before losing them due to oversmoothing. Results of our experiments show that the MLAP architecture improves the graph classification performance compared to the baseline architectures. In addition, analyses on the layer-wise graph representations suggest that aggregating information from multiple levels of localities indeed has the potential to improve the discriminability of learned graph representations.</abstract><cop>United States</cop><pub>Elsevier Ltd</pub><pmid>34808587</pmid><doi>10.1016/j.neunet.2021.11.001</doi><tpages>18</tpages><orcidid>https://orcid.org/0000-0002-4021-1481</orcidid><orcidid>https://orcid.org/0000-0003-0192-7177</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0893-6080 |
ispartof | Neural networks, 2022-01, Vol.145, p.356-373 |
issn | 0893-6080 1879-2782 |
language | eng |
recordid | cdi_proquest_miscellaneous_2601480583 |
source | MEDLINE; Elsevier ScienceDirect Journals |
subjects | Attention Graph neural network (GNN) Graph representation learning (GRL) Learning Multi-level attention pooling (MLAP) Multi-level locality Neural Networks, Computer |
title | Multi-level attention pooling for graph neural networks: Unifying graph representations with multiple localities |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T22%3A52%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multi-level%20attention%20pooling%20for%20graph%20neural%20networks:%20Unifying%20graph%20representations%20with%20multiple%20localities&rft.jtitle=Neural%20networks&rft.au=Itoh,%20Takeshi%20D.&rft.date=2022-01&rft.volume=145&rft.spage=356&rft.epage=373&rft.pages=356-373&rft.issn=0893-6080&rft.eissn=1879-2782&rft_id=info:doi/10.1016/j.neunet.2021.11.001&rft_dat=%3Cproquest_cross%3E2601480583%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2601480583&rft_id=info:pmid/34808587&rft_els_id=S0893608021004299&rfr_iscdi=true |