Efficient and Local Parallel Random Walks
Random walks are a fundamental primitive used in many machine learning algorithms with several applications in clustering and semi-supervised learning. Despite their relevance, the first efficient parallel algorithm to compute random walks has been introduced very recently (Lacki et al.). Unfortunat...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Kapralov, Michael Lattanzi, Silvio Nouri, Navid Tardos, Jakab |
description | Random walks are a fundamental primitive used in many machine learning
algorithms with several applications in clustering and semi-supervised
learning. Despite their relevance, the first efficient parallel algorithm to
compute random walks has been introduced very recently (Lacki et al.).
Unfortunately their method has a fundamental shortcoming: their algorithm is
non-local in that it heavily relies on computing random walks out of all nodes
in the input graph, even though in many practical applications one is
interested in computing random walks only from a small subset of nodes in the
graph. In this paper, we present a new algorithm that overcomes this limitation
by building random walk efficiently and locally at the same time. We show that
our technique is both memory and round efficient, and in particular yields an
efficient parallel local clustering algorithm. Finally, we complement our
theoretical analysis with experimental results showing that our algorithm is
significantly more scalable than previous approaches. |
doi_str_mv | 10.48550/arxiv.2112.00655 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2112_00655</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2112_00655</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-ecb5d059039b912ed6a2e8838c960ca4255350b45256f8b197fe79755ef3b3d33</originalsourceid><addsrcrecordid>eNotzjsLwjAYheEsDqL-ACezOrTm0i9NRhFvUFCk4Fi-pAkU44Uqov_e63TgHQ4PIUPO0kwDsAm2j-aeCs5FypgC6JLxPITGNf50o3iqaXF2GOkWW4zRR7p7t_OR7jEern3SCRivfvDfHikX83K2SorNcj2bFgmqHBLvLNQMDJPGGi58rVB4raV2RjGHmQCQwGwGAlTQlps8-NzkAD5IK2spe2T0u_1aq0vbHLF9Vh9z9TXLF5f_OgA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Efficient and Local Parallel Random Walks</title><source>arXiv.org</source><creator>Kapralov, Michael ; Lattanzi, Silvio ; Nouri, Navid ; Tardos, Jakab</creator><creatorcontrib>Kapralov, Michael ; Lattanzi, Silvio ; Nouri, Navid ; Tardos, Jakab</creatorcontrib><description>Random walks are a fundamental primitive used in many machine learning
algorithms with several applications in clustering and semi-supervised
learning. Despite their relevance, the first efficient parallel algorithm to
compute random walks has been introduced very recently (Lacki et al.).
Unfortunately their method has a fundamental shortcoming: their algorithm is
non-local in that it heavily relies on computing random walks out of all nodes
in the input graph, even though in many practical applications one is
interested in computing random walks only from a small subset of nodes in the
graph. In this paper, we present a new algorithm that overcomes this limitation
by building random walk efficiently and locally at the same time. We show that
our technique is both memory and round efficient, and in particular yields an
efficient parallel local clustering algorithm. Finally, we complement our
theoretical analysis with experimental results showing that our algorithm is
significantly more scalable than previous approaches.</description><identifier>DOI: 10.48550/arxiv.2112.00655</identifier><language>eng</language><subject>Computer Science - Data Structures and Algorithms ; Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning</subject><creationdate>2021-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2112.00655$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2112.00655$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Kapralov, Michael</creatorcontrib><creatorcontrib>Lattanzi, Silvio</creatorcontrib><creatorcontrib>Nouri, Navid</creatorcontrib><creatorcontrib>Tardos, Jakab</creatorcontrib><title>Efficient and Local Parallel Random Walks</title><description>Random walks are a fundamental primitive used in many machine learning
algorithms with several applications in clustering and semi-supervised
learning. Despite their relevance, the first efficient parallel algorithm to
compute random walks has been introduced very recently (Lacki et al.).
Unfortunately their method has a fundamental shortcoming: their algorithm is
non-local in that it heavily relies on computing random walks out of all nodes
in the input graph, even though in many practical applications one is
interested in computing random walks only from a small subset of nodes in the
graph. In this paper, we present a new algorithm that overcomes this limitation
by building random walk efficiently and locally at the same time. We show that
our technique is both memory and round efficient, and in particular yields an
efficient parallel local clustering algorithm. Finally, we complement our
theoretical analysis with experimental results showing that our algorithm is
significantly more scalable than previous approaches.</description><subject>Computer Science - Data Structures and Algorithms</subject><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzjsLwjAYheEsDqL-ACezOrTm0i9NRhFvUFCk4Fi-pAkU44Uqov_e63TgHQ4PIUPO0kwDsAm2j-aeCs5FypgC6JLxPITGNf50o3iqaXF2GOkWW4zRR7p7t_OR7jEern3SCRivfvDfHikX83K2SorNcj2bFgmqHBLvLNQMDJPGGi58rVB4raV2RjGHmQCQwGwGAlTQlps8-NzkAD5IK2spe2T0u_1aq0vbHLF9Vh9z9TXLF5f_OgA</recordid><startdate>20211201</startdate><enddate>20211201</enddate><creator>Kapralov, Michael</creator><creator>Lattanzi, Silvio</creator><creator>Nouri, Navid</creator><creator>Tardos, Jakab</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20211201</creationdate><title>Efficient and Local Parallel Random Walks</title><author>Kapralov, Michael ; Lattanzi, Silvio ; Nouri, Navid ; Tardos, Jakab</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-ecb5d059039b912ed6a2e8838c960ca4255350b45256f8b197fe79755ef3b3d33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Data Structures and Algorithms</topic><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Kapralov, Michael</creatorcontrib><creatorcontrib>Lattanzi, Silvio</creatorcontrib><creatorcontrib>Nouri, Navid</creatorcontrib><creatorcontrib>Tardos, Jakab</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kapralov, Michael</au><au>Lattanzi, Silvio</au><au>Nouri, Navid</au><au>Tardos, Jakab</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Efficient and Local Parallel Random Walks</atitle><date>2021-12-01</date><risdate>2021</risdate><abstract>Random walks are a fundamental primitive used in many machine learning
algorithms with several applications in clustering and semi-supervised
learning. Despite their relevance, the first efficient parallel algorithm to
compute random walks has been introduced very recently (Lacki et al.).
Unfortunately their method has a fundamental shortcoming: their algorithm is
non-local in that it heavily relies on computing random walks out of all nodes
in the input graph, even though in many practical applications one is
interested in computing random walks only from a small subset of nodes in the
graph. In this paper, we present a new algorithm that overcomes this limitation
by building random walk efficiently and locally at the same time. We show that
our technique is both memory and round efficient, and in particular yields an
efficient parallel local clustering algorithm. Finally, we complement our
theoretical analysis with experimental results showing that our algorithm is
significantly more scalable than previous approaches.</abstract><doi>10.48550/arxiv.2112.00655</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2112.00655 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2112_00655 |
source | arXiv.org |
subjects | Computer Science - Data Structures and Algorithms Computer Science - Distributed, Parallel, and Cluster Computing Computer Science - Learning |
title | Efficient and Local Parallel Random Walks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-31T00%3A32%3A07IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Efficient%20and%20Local%20Parallel%20Random%20Walks&rft.au=Kapralov,%20Michael&rft.date=2021-12-01&rft_id=info:doi/10.48550/arxiv.2112.00655&rft_dat=%3Carxiv_GOX%3E2112_00655%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |