Lecture 13: Power Iterations
1.Reduction to triangular form
The relevance of eigendecompositions
to repeated application of the linear operator
as in
suggests that algorithms that construct powers of
might reveal eigenvalues. This is indeed the case and leads to a class
of algorithms of wide applicability in scientific computation. First,
observe that taking condition numbers gives
where ,
are the eigenvalues of maximum and minimum absolute values. While these
express an intrinsic property of the operator ,
the factor is
associated with the conditioning of a change of coordinates, and a
natural question is whether it is possible to avoid any ill-conditioning
associated with a basis set that is
close to linear dependence. The answer to this line of inquiry is given
by the following result.
Schur Theorem. For any
there exists unitary and
upper triangular such that .
Proof. Proceed by induction, starting from an
arbitrary eigenvalue and eigenvector
. Let , the
first column vector of a unitary matrix . Then
with
that by the inductive hypothesis can be written as ,
with unitary,
upper triangular. The matrix
is a product of unitary matrices, hence itself unitary. The
computation
then shows that is indeed
triangular.
The eigenvalues of an upper triangular matrix are simply its diagonal
elements, so the Schur factorization is an eigenvalue-revealing
factorization.
2.Power iteration for real symmetric
matrices
When the operator expresses some
physical phenomenon, the principle of action and reaction implies that
is symmetric,
and has real eigenvalues. Componentwise, symmetry of implies .
Consider ,
and take the adjoint to obtain ,
or
since is symmetric. Form scalar
products ,
,
and subtract to obtain
since ,
an eigenvector.
Example. Consider a linear array of identical
mass-springs. The
point mass obeys the dynamics
expressed in matrix form as ,
with symmetric.
For a real symmetric matrix the Schur theorem states that
and since a symmetric triangular matrix is diagonal, the Schur
factorization is also an eigendecomposition, and the eigenvector matrix
is a basis, .
2.1.The power iteration idea
Assume initially that the eigenvalues are distinct and ordered . Repeated application of
on an arbitrary vector is
expressed as
a linear combination of the columns of
(eigenvectors of ) with coefficients
.
For large enough ,
, ,
leads to a dominant contribution along the direcion of eigenvector
This gives a procedure for finding one eigenvector of a matrix, and
the Schur theorem proof suggests a recursive algorithm to find all
eigenvalues can be defined.
Example. Construct an
matrix with eigenvalues ,
,
and arbitrary orthonormalized eigenvectors .
∴ |
function genAQΛ(m)
X=rand(m,m); Q=qr(X).Q; Λ=diagm(2.0 .^ (0:-1:1-m))
A=Q*Λ*Q'
return A,Q,Λ
end; |
Carry out power iterations starting
from some arbitrary initial vector ,
constructing
∴ |
for k=1:n
global V
V[:,k+1] = A*V[:,k]
end |
Compute components of the vectors in the
basis, .
Row of
contains the coefficients of
along eigenvector .
Notice that for
the coefficients decrease.
∴ |
clf(); plot(log10.(abs.(C))); |
∴ |
xlabel("Iteration"); ylabel("lg|c_i|"); |
The sequence of normalized eigenvector approximants is linearly convergent at rate .
2.2.Rayleigh quotient
To estimate the eigenvalue revealed by power iteration, formulate the
least squares problem
that seeks the best approximation of one power iteration
as a linear combination of the initial vector .
Of course, if
is an eigenvector, then the solution would be ,
the associated eigenvalue. The projector onto is
that when applied to
gives the equation
The function ,
is known as the Rayleigh quotient which, evaluated for an eigenvector,
gives . To determine how
well the eigenvalue is approximated, carry out a Taylor series in the
vicinity of an eigenvector
where
is the gradient of
Compute the gradient through differentiation of the Rayleigh quotient
Noting that ,
the column of
, the gradient of
is
To compute , let ,
and since is symmetric ,
leading to
Use
also expressed as
by swapping indices to obtain
and therefore
Use symmetry of to write
and substitute above to obtain
Gathering the above results
gives the following gradient of the Rayleigh quotient
When evaluated at ,
obtain ,
implying that near an eigenvector the Rayleigh quotient approximation of
an eigenvalue is of quadratic accuracy,
2.3.Refining the power iteration idea
Power iteration furnishes the largest eigenvalue. Further eigenvalues
can be found by use of the following properties:
-
eigenpair of eigenpair of ;
-
eigenpair of eigenpair of .
If is a known initial approximation of
the eigenvalue then the inverse power iteration ,
actually implemented as successive solution of linear systems
leads to a sequence of Rayleigh quotients
that converges quadratically to an eigenvalue close to .
An important refinement of the idea is to change the shift at each
iteration which leads to cubic order of convergence
Algorithm (Rayleigh quotient iteration)
(solve linear system)
Power iteration can be applied simultaneously to multiple directions at
once
Algorithm (Simultaneous iteration)
(power iteration applied to multiple directions)
(orthogonalize new directions)