Efficient Scaling of Dynamic Graph Neural Networks

We present distributed algorithms for training dynamic Graph Neural Networks (GNN) on large scale graphs spanning multi-node, multi-GPU systems. To the best of our knowledge, this is the first scaling study on dynamic GNN. We devise mechanisms for reducing the GPU memory usage and identify two execu...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2021-09
Hauptverfasser: Chakaravarthy, Venkatesan T, Pandian, Shivmaran S, Raje, Saurabh, Sabharwal, Yogish, Suzumura, Toyotaro, Ubaru, Shashanka
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Chakaravarthy, Venkatesan T
Pandian, Shivmaran S
Raje, Saurabh
Sabharwal, Yogish
Suzumura, Toyotaro
Ubaru, Shashanka
description We present distributed algorithms for training dynamic Graph Neural Networks (GNN) on large scale graphs spanning multi-node, multi-GPU systems. To the best of our knowledge, this is the first scaling study on dynamic GNN. We devise mechanisms for reducing the GPU memory usage and identify two execution time bottlenecks: CPU-GPU data transfer; and communication volume. Exploiting properties of dynamic graphs, we design a graph difference-based strategy to significantly reduce the transfer time. We develop a simple, but effective data distribution technique under which the communication volume remains fixed and linear in the input size, for any number of GPUs. Our experiments using billion-size graphs on a system of 128 GPUs shows that: (i) the distribution scheme achieves up to 30x speedup on 128 GPUs; (ii) the graph-difference technique reduces the transfer time by a factor of up to 4.1x and the overall execution time by up to 40%
doi_str_mv 10.48550/arxiv.2109.07893
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2109_07893</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2573634325</sourcerecordid><originalsourceid>FETCH-LOGICAL-a525-4ee305a8f414669c8ac3690e4f5b6573b5e48f222bcd9a0ca01e39580362c2053</originalsourceid><addsrcrecordid>eNotj8FOwzAQRC0kJKrSD-BEJM4pm12v4xxRKS1SBQd6jxxjQ0qaBCcB-veEltNcnmbmCXGVwFxqZrg14af8mmMC2RxSndGZmCBREmuJeCFmXbcDAFQpMtNE4NL70pau7qMXa6qyfosaH90farMvbbQKpn2PntwQTDVG_92Ej-5SnHtTdW72n1OxfVhuF-t487x6XNxtYsPIsXSOgI32MpFKZVYbSyoDJz0XilMq2EntEbGwr5kBayBxlLEGUmgRmKbi-lR7FMrbUO5NOOR_YvlRbCRuTkQbms_BdX2-a4ZQj59yHBcUSUKmX_eKTm8</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2573634325</pqid></control><display><type>article</type><title>Efficient Scaling of Dynamic Graph Neural Networks</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Chakaravarthy, Venkatesan T ; Pandian, Shivmaran S ; Raje, Saurabh ; Sabharwal, Yogish ; Suzumura, Toyotaro ; Ubaru, Shashanka</creator><creatorcontrib>Chakaravarthy, Venkatesan T ; Pandian, Shivmaran S ; Raje, Saurabh ; Sabharwal, Yogish ; Suzumura, Toyotaro ; Ubaru, Shashanka</creatorcontrib><description>We present distributed algorithms for training dynamic Graph Neural Networks (GNN) on large scale graphs spanning multi-node, multi-GPU systems. To the best of our knowledge, this is the first scaling study on dynamic GNN. We devise mechanisms for reducing the GPU memory usage and identify two execution time bottlenecks: CPU-GPU data transfer; and communication volume. Exploiting properties of dynamic graphs, we design a graph difference-based strategy to significantly reduce the transfer time. We develop a simple, but effective data distribution technique under which the communication volume remains fixed and linear in the input size, for any number of GPUs. Our experiments using billion-size graphs on a system of 128 GPUs shows that: (i) the distribution scheme achieves up to 30x speedup on 128 GPUs; (ii) the graph-difference technique reduces the transfer time by a factor of up to 4.1x and the overall execution time by up to 40%</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2109.07893</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning ; Data transfer (computers) ; Graph neural networks ; Graphs ; Neural networks</subject><ispartof>arXiv.org, 2021-09</ispartof><rights>2021. This work is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,784,885,27923</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2109.07893$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1145/3458817.3480858$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Chakaravarthy, Venkatesan T</creatorcontrib><creatorcontrib>Pandian, Shivmaran S</creatorcontrib><creatorcontrib>Raje, Saurabh</creatorcontrib><creatorcontrib>Sabharwal, Yogish</creatorcontrib><creatorcontrib>Suzumura, Toyotaro</creatorcontrib><creatorcontrib>Ubaru, Shashanka</creatorcontrib><title>Efficient Scaling of Dynamic Graph Neural Networks</title><title>arXiv.org</title><description>We present distributed algorithms for training dynamic Graph Neural Networks (GNN) on large scale graphs spanning multi-node, multi-GPU systems. To the best of our knowledge, this is the first scaling study on dynamic GNN. We devise mechanisms for reducing the GPU memory usage and identify two execution time bottlenecks: CPU-GPU data transfer; and communication volume. Exploiting properties of dynamic graphs, we design a graph difference-based strategy to significantly reduce the transfer time. We develop a simple, but effective data distribution technique under which the communication volume remains fixed and linear in the input size, for any number of GPUs. Our experiments using billion-size graphs on a system of 128 GPUs shows that: (i) the distribution scheme achieves up to 30x speedup on 128 GPUs; (ii) the graph-difference technique reduces the transfer time by a factor of up to 4.1x and the overall execution time by up to 40%</description><subject>Algorithms</subject><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><subject>Data transfer (computers)</subject><subject>Graph neural networks</subject><subject>Graphs</subject><subject>Neural networks</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GOX</sourceid><recordid>eNotj8FOwzAQRC0kJKrSD-BEJM4pm12v4xxRKS1SBQd6jxxjQ0qaBCcB-veEltNcnmbmCXGVwFxqZrg14af8mmMC2RxSndGZmCBREmuJeCFmXbcDAFQpMtNE4NL70pau7qMXa6qyfosaH90farMvbbQKpn2PntwQTDVG_92Ej-5SnHtTdW72n1OxfVhuF-t487x6XNxtYsPIsXSOgI32MpFKZVYbSyoDJz0XilMq2EntEbGwr5kBayBxlLEGUmgRmKbi-lR7FMrbUO5NOOR_YvlRbCRuTkQbms_BdX2-a4ZQj59yHBcUSUKmX_eKTm8</recordid><startdate>20210916</startdate><enddate>20210916</enddate><creator>Chakaravarthy, Venkatesan T</creator><creator>Pandian, Shivmaran S</creator><creator>Raje, Saurabh</creator><creator>Sabharwal, Yogish</creator><creator>Suzumura, Toyotaro</creator><creator>Ubaru, Shashanka</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210916</creationdate><title>Efficient Scaling of Dynamic Graph Neural Networks</title><author>Chakaravarthy, Venkatesan T ; Pandian, Shivmaran S ; Raje, Saurabh ; Sabharwal, Yogish ; Suzumura, Toyotaro ; Ubaru, Shashanka</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a525-4ee305a8f414669c8ac3690e4f5b6573b5e48f222bcd9a0ca01e39580362c2053</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Algorithms</topic><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><topic>Data transfer (computers)</topic><topic>Graph neural networks</topic><topic>Graphs</topic><topic>Neural networks</topic><toplevel>online_resources</toplevel><creatorcontrib>Chakaravarthy, Venkatesan T</creatorcontrib><creatorcontrib>Pandian, Shivmaran S</creatorcontrib><creatorcontrib>Raje, Saurabh</creatorcontrib><creatorcontrib>Sabharwal, Yogish</creatorcontrib><creatorcontrib>Suzumura, Toyotaro</creatorcontrib><creatorcontrib>Ubaru, Shashanka</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chakaravarthy, Venkatesan T</au><au>Pandian, Shivmaran S</au><au>Raje, Saurabh</au><au>Sabharwal, Yogish</au><au>Suzumura, Toyotaro</au><au>Ubaru, Shashanka</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Efficient Scaling of Dynamic Graph Neural Networks</atitle><jtitle>arXiv.org</jtitle><date>2021-09-16</date><risdate>2021</risdate><eissn>2331-8422</eissn><abstract>We present distributed algorithms for training dynamic Graph Neural Networks (GNN) on large scale graphs spanning multi-node, multi-GPU systems. To the best of our knowledge, this is the first scaling study on dynamic GNN. We devise mechanisms for reducing the GPU memory usage and identify two execution time bottlenecks: CPU-GPU data transfer; and communication volume. Exploiting properties of dynamic graphs, we design a graph difference-based strategy to significantly reduce the transfer time. We develop a simple, but effective data distribution technique under which the communication volume remains fixed and linear in the input size, for any number of GPUs. Our experiments using billion-size graphs on a system of 128 GPUs shows that: (i) the distribution scheme achieves up to 30x speedup on 128 GPUs; (ii) the graph-difference technique reduces the transfer time by a factor of up to 4.1x and the overall execution time by up to 40%</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2109.07893</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2021-09
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2109_07893
source arXiv.org; Free E- Journals
subjects Algorithms
Computer Science - Distributed, Parallel, and Cluster Computing
Computer Science - Learning
Data transfer (computers)
Graph neural networks
Graphs
Neural networks
title Efficient Scaling of Dynamic Graph Neural Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T03%3A04%3A36IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Efficient%20Scaling%20of%20Dynamic%20Graph%20Neural%20Networks&rft.jtitle=arXiv.org&rft.au=Chakaravarthy,%20Venkatesan%20T&rft.date=2021-09-16&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2109.07893&rft_dat=%3Cproquest_arxiv%3E2573634325%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2573634325&rft_id=info:pmid/&rfr_iscdi=true