Lecture 12: The Eigenvalue Problem
1.Definitions
Linear endomorphisms ,
represented by ,
can exhibit invariant directions
for which
known as eigenvectors, with associated eigenvalue .
Eigenvectors are non-zero elements of the null space of ,
and the null-space is referred to as the eigenspace of
for eigenvalue , .
Non-zero solutions are obtained if
is rank-deficient (singular), or has linearly dependent columns in which
case
From the determinant definition as “sum of all products choosing
an element from row/column”, it results that
known as the characteristic polynomial associated with the matrix
𝑨, and of degree . The
characteristic polynomial is monic, meaning that the coefficient
of the highest power
is equal to one. The fundamental theorem of algebra states that of degree
has roots, hence
has eigenvalues (not necessarily
distinct), and associated eigenvectors.
This can be stated in matrix form as
with
the eigenvector matrix and eigenvalue matrix, respectively. By
definition, the matrix is
diagonalizable if is of full
rank, in which case the eigendecomposition of
is
1.1.Coordinate transformations
The statement ,
that eigenvector is an invariant
direction of the operator along which
the effect of operator is scaling by ,
suggests that similar behavior would be obtained under a coordinate
transformation .
Assuming is of full rank and
introducing ,
this leads to
Upon coordinate transformation, the eigenvalues (scaling factors along
the invariant directions) stay the same. Metric-preserving coordinate
transformations are of particular interest, in which case the
transformation matrix is unitary .
Definition. Matrices
are said to be similar, ,
if there exists some full rank matrix
such that .
Proposition. Similar matrices ,
,
have the same eigenvalues, and eigenvectors
of ,
of are related through .
Since the eigenvalues of
are the same, and a polynomial is completely specified by its roots and
coefficient of highest power, the characteristic polynomials of must be
the same
This can also be verified through the determinant definition
since .
1.2.Paradigmatic eigenvalue problem
solutions
Reflection matrix. The
matrix
is the two-dimensional Householder reflector across .
Vectors colinear with change
direction along the same orientation upon reflection, while vectors
orthogonal to (i.e., in the null
space ) are unchanged. It is therefore to
be expected that ,
,
and ,
.
This is readily verified
 |
|
Figure 1. Reflector in two dimensions
|
∴ |
v=[-1; 1]; q=v/norm(v); H=I-2*q*q'; |
|
(1) |
|
(2) |
Rotation matrix. The
matrix
represents the isometric rotation of two-dimensional vectors. If , with
eigenvalues ,
and eigenvector matrix .
For ,
the eigenvalues are ,
again with eigenvector matrix .
If ,
the orientation of any non-zero
changes upon rotation by . The
characteristic polynomial has complex roots
and the directions of invariant orientation have complex components
(are outside the real plane )
∴ |
R=[1 -1; 1 1]/sqrt(2); eigvals(R) |
|
(3) |
|
(4) |
Second-order differentiation
matrix. Eigenvalues of matrices arising from discretization of
continuum operators can be obtained from the operator eigenproblem.
The second-order differentiation operator
has eigenvalues
associated with eigenfunctions
Sampling of at ,
,
leads to the vector
with components . The boundary conditions at the
sampling interval end-points affect the eigenvalues. Imposing , at
and
leads to .
The derivative can be approximated at the sample points through
The derivative approximation vector
results from a linear mapping
and the matrix
has eigenvectors and eigenvalues
, .
In the limit of an infinite number of sampling points the continuum
eigenvalues are obtained, exemplifying again the correspondence
principle between discrete and continuum representations
∴ |
m=4; o=ones(m-1,1)[:,1]; D=diagm(-1 => o, 1 => o)-2*I |
|
(5) |
∴ |
function ErrorλD(m)
h=π/(m+1); ξ=1:m; o=ones(m-1); D=(diagm(-1=>o, 1=>o)-2*I)/h^2;
λD = sort(eigvals(D),rev=true)
λ = -(4/h^2)*sin.(ξ*h/2).^2
ε=norm(λD .- λ)/norm(λ)
end; |
∴ |
ε=ErrorλD.(20:20:100)'; |
∴ |
print(round.(ε; digits=3)) |
∴ |
function eigD(m)
h=π/m; o=ones(m-1,1)[:,1]; D=diagm(-1 => o, 1 => o)-2*I;
λD=sort(eigvals(D),rev=true)
λ = -4*sin.(collect(1:m).*h/2).^2
return [λD λ]
end; |
∴ |
for m=10:10:100
local Λ=eigD(m)
plot(1:m,Λ[:,1],"ob",markerfacecolor="none")
plot(1:m,Λ[:,2],"xr")
end |
∴ |
xlabel("k"); ylabel("λ[k]"); |
∴ |
pre=homedir()*"/courses/MATH661/images/"; |
∴ |
savefig(pre*"L14D2eigs.eps"); |
1.3.Matrix eigendecomposition
A solution
to the eigenvalue problem
always exists, but the eigenvectors of
do not always form a basis set, i.e.,
is not always of full rank. The factorized form of the characteristic
polynomial of
is
with
denoting the number of distinct roots of , and is the algebraic
multiplicity of eigenvalue ,
defined as the number of times the root
is repeated. Let
denote the associated eigenspace . The dimension of
denoted by is the
geometric multiplicity of eigenvalue .
The eigenvector matrix is of full rank when the vector sum of the
eigenspaces covers ,
as established by the following results.
Proposition. The geometric multiplicity is at
least 1, .
Proof. By contradiction if ,
then , but
eigenvectors cannot be null.
Proposition. If
then (the
eigenspaces of distinct eigenvalues are disjoint)
Proof. Let , hence
and , hence .
Subtraction gives
Since
it results that .
Proposition. The geometric multiplicity of an
eigenvalue is less or equal to its algebraic multiplicity,
Proof. Let
be an orthonormal basis for . By definition of a null space,
i.e., every vector of the eigenspace is an eigenvector with eigenvalue
. Then
Form the unitary matrix ,
and compute
Since is unitary, obtain
where
is the
identity matrix, and in the above
denotes a
matrix of zeros. The matrix
is similar to and has the same
eigenvalues. Since , the algebraic multiplicity of must be at least ,
i.e., .
Definition 1. An
eigenvalue for which the geometric multiplicity is less than the
algebraic multiplicity is said to be defective.
Example.
Non-defective matrices exist, for example
|
(6) |
|
(8) |
Example.
Non-defective matrices with repeated eigenvalues exist, for
example
∴ |
A=diagm([1,1,1]); eigvals(A) |
|
(10) |
Example.
Defective matrices exist, for example
has eigenvalue
with algebraic multiplicity .
Reduction to row-echelon form of
leads to
and , i.e., the
geometric multiplicity is equal to 1. The above is known as a Jordan
block.
∴ |
A=[3 1 1; 0 3 1; 0 0 3] |
|
(11) |
|
(13) |
|
(14) |
Proposition 2. A matrix
is diagonalizable if the geometric multiplicity of each eigenvalue is
equal to the algebraic multiplicity of that eigenvalue.
Proof. Recall that is
diagonalizable if the eigenvector matrix
is of full rank. Since the eigenspaces
of the distinct eigenvalues are disjoint,
the column space of is the direct
vector sum of the eigenspaces
The dimension of is therefore given by the sum of the
eigenspace dimensions
Since ,
the only possibility for to be of full
rank, ,
is for .
1.4.Matrix properties from eigenvalues
Eigenvalues as roots of the characteristic polynomial
reveal properties of a matrix .
The evaluation of leads to
hence the determinant of a matrix is given by the product of its
eigenvalues
The trace of a matrix is the sum of its diagonal elements is equal to
the sum of its eigenvalues
a relationship established by the Vieta formulas.
1.5.Matrix eigendecomposition
applications
Whereas the SVD, QR, LU decompositions can be applied to general
matrices
with not necessarily equal to , the eigendecomposition requires ,
and hence is especially relevant in the characterization of
endomorphisms. A generic time evolution problem is stated as
stating that the rate of change in the state variables
characterizing some system is a function of the current state through
the function ,
an endomorphism. An approximation of
is furnished by the MacLaurin series
Truncation at first order gives a linear ODE system ,
that can be formally integrated to give
The matrix exponential
is defined as
Evaluation of
requires
matrix multiplications or
floating point operations. However, if the eigendecomposition of
is available the matrix exponential can be evaluate in only
operations since
leads to
2.Computation of the SVD
The existence of the SVD
was establish by a constructive procedure by complete induction. However
the proof depends on determining the singular values, e.g., . The
existence of the singular values was established by an argument from
analysis, that the norm function on a compact domain must attain its
extrema. This however leaves open the problem of effectively determining
the singular values. In practive the singular values and vectors are
determined by solving the eigenvalue problem for
and
From the above the left singular vectors
are eigenvectors of ,
and the right singular vectors are eigenvectors of .
Both
and
have the same eigenvalues that are the squared singular values.