Solving the k-sparse Eigenvalue Problem with Reinforcement Learning
We examine the possibility of using a reinforcement learning (RL) algorithm to solve large-scale eigenvalue problems in which the desired the eigenvector can be approximated by a sparse vector with at most $k$ nonzero elements, where $k$ is relatively small compare to the dimension of the matrix to...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | We examine the possibility of using a reinforcement learning (RL) algorithm
to solve large-scale eigenvalue problems in which the desired the eigenvector
can be approximated by a sparse vector with at most $k$ nonzero elements, where
$k$ is relatively small compare to the dimension of the matrix to be partially
diagonalized. This type of problem arises in applications in which the desired
eigenvector exhibits localization properties and in large-scale eigenvalue
computations in which the amount of computational resource is limited. When the
positions of these nonzero elements can be determined, we can obtain the
$k$-sparse approximation to the original problem by computing eigenvalues of a
$k\times k$ submatrix extracted from $k$ rows and columns of the original
matrix. We review a previously developed greedy algorithm for incrementally
probing the positions of the nonzero elements in a $k$-sparse approximate
eigenvector and show that the greedy algorithm can be improved by using an RL
method to refine the selection of $k$ rows and columns of the original matrix.
We describe how to represent states, actions, rewards and policies in an RL
algorithm designed to solve the $k$-sparse eigenvalue problem and demonstrate
the effectiveness of the RL algorithm on two examples originating from quantum
many-body physics. |
---|---|
DOI: | 10.48550/arxiv.2009.04414 |