A note on quickly sampling a sparse matrix with low rank expectation
Given matrices \(X,Y \in R^{n \times K}\) and \(S \in R^{K \times K}\) with positive elements, this paper proposes an algorithm fastRG to sample a sparse matrix \(A\) with low rank expectation \(E(A) = XSY^T\) and independent Poisson elements. This allows for quickly sampling from a broad class of s...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2017-03 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Given matrices \(X,Y \in R^{n \times K}\) and \(S \in R^{K \times K}\) with positive elements, this paper proposes an algorithm fastRG to sample a sparse matrix \(A\) with low rank expectation \(E(A) = XSY^T\) and independent Poisson elements. This allows for quickly sampling from a broad class of stochastic blockmodel graphs (degree-corrected, mixed membership, overlapping) all of which are specific parameterizations of the generalized random product graph model defined in Section 2.2. The basic idea of fastRG is to first sample the number of edges \(m\) and then sample each edge. The key insight is that because of the the low rank expectation, it is easy to sample individual edges. The naive "element-wise" algorithm requires \(O(n^2)\) operations to generate the \(n\times n\) adjacency matrix \(A\). In sparse graphs, where \(m = O(n)\), ignoring log terms, fastRG runs in time \(O(n)\). An implementation in fastRG is available on github. A computational experiment in Section 2.4 simulates graphs up to \(n=10,000,000\) nodes with \(m = 100,000,000\) edges. For example, on a graph with \(n=500,000\) and \(m = 5,000,000\), fastRG runs in less than one second on a 3.5 GHz Intel i5. |
---|---|
ISSN: | 2331-8422 |