Discussion of methods for eigenvalues

If the matrix to diagonalize is large and sparse, direct methods simply become impractical, also because many of the direct methods tend to destroy sparsity. As a result large dense matrices may arise during the diagonalization procedure. The idea behind iterative methods is to project the $n-$dimensional problem in smaller spaces, so-called Krylov subspaces. Given a matrix \( \mathbf{A} \) and a vector \( \mathbf{v} \), the associated Krylov sequences of vectors (and thereby subspaces) \( \mathbf{v} \), \( \mathbf{A}\mathbf{v} \), \( \mathbf{A}^2\mathbf{v} \), \( \mathbf{A}^3\mathbf{v},\dots \), represent successively larger Krylov subspaces.

Matrix \( \mathbf{A}\mathbf{x}=\mathbf{b} \) \( \mathbf{A}\mathbf{x}=\lambda\mathbf{x} \)
\( \mathbf{A}=\mathbf{A}^* \) Conjugate gradient Lanczos
\( \mathbf{A}\ne \mathbf{A}^* \) GMRES etc Arnoldi