Efficient and Local Parallel Random Walks

Random walks are a fundamental primitive used in many machine learning algorithms with several applications in clustering and semi-supervised learning. Despite their relevance, the first efficient parallel algorithm to compute random walks has been introduced very recently (Lacki et al.). Unfortunat...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Kapralov, Michael, Lattanzi, Silvio, Nouri, Navid, Tardos, Jakab
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Kapralov, Michael
Lattanzi, Silvio
Nouri, Navid
Tardos, Jakab
description Random walks are a fundamental primitive used in many machine learning algorithms with several applications in clustering and semi-supervised learning. Despite their relevance, the first efficient parallel algorithm to compute random walks has been introduced very recently (Lacki et al.). Unfortunately their method has a fundamental shortcoming: their algorithm is non-local in that it heavily relies on computing random walks out of all nodes in the input graph, even though in many practical applications one is interested in computing random walks only from a small subset of nodes in the graph. In this paper, we present a new algorithm that overcomes this limitation by building random walk efficiently and locally at the same time. We show that our technique is both memory and round efficient, and in particular yields an efficient parallel local clustering algorithm. Finally, we complement our theoretical analysis with experimental results showing that our algorithm is significantly more scalable than previous approaches.
doi_str_mv 10.48550/arxiv.2112.00655
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2112_00655</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2112_00655</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-ecb5d059039b912ed6a2e8838c960ca4255350b45256f8b197fe79755ef3b3d33</originalsourceid><addsrcrecordid>eNotzjsLwjAYheEsDqL-ACezOrTm0i9NRhFvUFCk4Fi-pAkU44Uqov_e63TgHQ4PIUPO0kwDsAm2j-aeCs5FypgC6JLxPITGNf50o3iqaXF2GOkWW4zRR7p7t_OR7jEern3SCRivfvDfHikX83K2SorNcj2bFgmqHBLvLNQMDJPGGi58rVB4raV2RjGHmQCQwGwGAlTQlps8-NzkAD5IK2spe2T0u_1aq0vbHLF9Vh9z9TXLF5f_OgA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Efficient and Local Parallel Random Walks</title><source>arXiv.org</source><creator>Kapralov, Michael ; Lattanzi, Silvio ; Nouri, Navid ; Tardos, Jakab</creator><creatorcontrib>Kapralov, Michael ; Lattanzi, Silvio ; Nouri, Navid ; Tardos, Jakab</creatorcontrib><description>Random walks are a fundamental primitive used in many machine learning algorithms with several applications in clustering and semi-supervised learning. Despite their relevance, the first efficient parallel algorithm to compute random walks has been introduced very recently (Lacki et al.). Unfortunately their method has a fundamental shortcoming: their algorithm is non-local in that it heavily relies on computing random walks out of all nodes in the input graph, even though in many practical applications one is interested in computing random walks only from a small subset of nodes in the graph. In this paper, we present a new algorithm that overcomes this limitation by building random walk efficiently and locally at the same time. We show that our technique is both memory and round efficient, and in particular yields an efficient parallel local clustering algorithm. Finally, we complement our theoretical analysis with experimental results showing that our algorithm is significantly more scalable than previous approaches.</description><identifier>DOI: 10.48550/arxiv.2112.00655</identifier><language>eng</language><subject>Computer Science - Data Structures and Algorithms ; Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning</subject><creationdate>2021-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2112.00655$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2112.00655$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Kapralov, Michael</creatorcontrib><creatorcontrib>Lattanzi, Silvio</creatorcontrib><creatorcontrib>Nouri, Navid</creatorcontrib><creatorcontrib>Tardos, Jakab</creatorcontrib><title>Efficient and Local Parallel Random Walks</title><description>Random walks are a fundamental primitive used in many machine learning algorithms with several applications in clustering and semi-supervised learning. Despite their relevance, the first efficient parallel algorithm to compute random walks has been introduced very recently (Lacki et al.). Unfortunately their method has a fundamental shortcoming: their algorithm is non-local in that it heavily relies on computing random walks out of all nodes in the input graph, even though in many practical applications one is interested in computing random walks only from a small subset of nodes in the graph. In this paper, we present a new algorithm that overcomes this limitation by building random walk efficiently and locally at the same time. We show that our technique is both memory and round efficient, and in particular yields an efficient parallel local clustering algorithm. Finally, we complement our theoretical analysis with experimental results showing that our algorithm is significantly more scalable than previous approaches.</description><subject>Computer Science - Data Structures and Algorithms</subject><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzjsLwjAYheEsDqL-ACezOrTm0i9NRhFvUFCk4Fi-pAkU44Uqov_e63TgHQ4PIUPO0kwDsAm2j-aeCs5FypgC6JLxPITGNf50o3iqaXF2GOkWW4zRR7p7t_OR7jEern3SCRivfvDfHikX83K2SorNcj2bFgmqHBLvLNQMDJPGGi58rVB4raV2RjGHmQCQwGwGAlTQlps8-NzkAD5IK2spe2T0u_1aq0vbHLF9Vh9z9TXLF5f_OgA</recordid><startdate>20211201</startdate><enddate>20211201</enddate><creator>Kapralov, Michael</creator><creator>Lattanzi, Silvio</creator><creator>Nouri, Navid</creator><creator>Tardos, Jakab</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20211201</creationdate><title>Efficient and Local Parallel Random Walks</title><author>Kapralov, Michael ; Lattanzi, Silvio ; Nouri, Navid ; Tardos, Jakab</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-ecb5d059039b912ed6a2e8838c960ca4255350b45256f8b197fe79755ef3b3d33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Data Structures and Algorithms</topic><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Kapralov, Michael</creatorcontrib><creatorcontrib>Lattanzi, Silvio</creatorcontrib><creatorcontrib>Nouri, Navid</creatorcontrib><creatorcontrib>Tardos, Jakab</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kapralov, Michael</au><au>Lattanzi, Silvio</au><au>Nouri, Navid</au><au>Tardos, Jakab</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Efficient and Local Parallel Random Walks</atitle><date>2021-12-01</date><risdate>2021</risdate><abstract>Random walks are a fundamental primitive used in many machine learning algorithms with several applications in clustering and semi-supervised learning. Despite their relevance, the first efficient parallel algorithm to compute random walks has been introduced very recently (Lacki et al.). Unfortunately their method has a fundamental shortcoming: their algorithm is non-local in that it heavily relies on computing random walks out of all nodes in the input graph, even though in many practical applications one is interested in computing random walks only from a small subset of nodes in the graph. In this paper, we present a new algorithm that overcomes this limitation by building random walk efficiently and locally at the same time. We show that our technique is both memory and round efficient, and in particular yields an efficient parallel local clustering algorithm. Finally, we complement our theoretical analysis with experimental results showing that our algorithm is significantly more scalable than previous approaches.</abstract><doi>10.48550/arxiv.2112.00655</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2112.00655
ispartof
issn
language eng
recordid cdi_arxiv_primary_2112_00655
source arXiv.org
subjects Computer Science - Data Structures and Algorithms
Computer Science - Distributed, Parallel, and Cluster Computing
Computer Science - Learning
title Efficient and Local Parallel Random Walks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-31T00%3A32%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Efficient%20and%20Local%20Parallel%20Random%20Walks&rft.au=Kapralov,%20Michael&rft.date=2021-12-01&rft_id=info:doi/10.48550/arxiv.2112.00655&rft_dat=%3Carxiv_GOX%3E2112_00655%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true