Eigen Values
Contents
4.9. Eigen Values#
In this subsection, we discuss the eigenvalues and eigenvectors of square matrices. Much of the discussion in this section will be equally applicable to real as well as complex square matrices. We will use the complex notation mostly and make specific remarks for real matrices wherever needed. Of specific interest is the eigenvalue decomposition of real symmetric matrices.
4.9.1. Eigen Values#
Definition 4.112 (Eigen value)
A scalar
A nonzero vector
An eigen value is also known as a characteristic value, proper value or a latent value.
We note that (4.9) can be written as
Thus
Definition 4.113 (Spectrum of eigen values)
The set comprising of eigen values of a matrix
Remark 4.13 (Uniqueness of eigen value for an eigen vector)
For each eigen vector
Proof. We show this by contradiction.
Assume that for
there are two eigen values and .Then
This can happen only when either
or .Since
is an eigen vector, it cannot be .Thus
.
Remark 4.14
If
Proof. We proceed as follows:
since
Remark 4.15
An eigen vector
In other words
We can put together eigen vectors of a matrix into another matrix by itself. This can be very useful tool. We start with a simple idea.
Lemma 4.50 (Matrix of eigen vectors)
Let
Then all the
Proof. Expanding the equation, we can write
Clearly we want
where
Converse: Assume that
This idea will be extended further in building the eigen value decomposition of matrices.
Observation 4.5
Let
In other words the angle between
Meanwhile
Thus
The angle
4.9.2. Eigen Space#
Definition 4.114 (Eigen space)
Let
Remark 4.16 (Eigen vectors for an eigen value)
The set comprising all the eigen vectors of
since
Definition 4.115 (Geometric multiplicity)
Let
We can see that
4.9.3. Characteristic Polynomial#
Remark 4.17 (Singularity of
A scalar
Observation 4.6 (Eigenvalues as zeros of a polynomial)
We can write the determinant as a polynomial:
where
It is easy to show that
Definition 4.116 (Characteristic polynomial)
For any square matrix
is known as its characteristic polynomial. The equation give by
is known as its characteristic equation.
The eigen values of
Observation 4.7 (Factors of characteristic polynomial)
For real square matrices, if we restrict eigen values to real values, then the characteristic polynomial can be factored as
The polynomial has
Clearly
For complex square matrices where eigen values can be complex (including real square matrices), the characteristic polynomial can be factored as
The polynomial can be completely factorized into first degree polynomials.
There are
Thus, including the duplicates there are exactly
It is quite possible that a real square matrix doesn’t have any real eigen values.
Definition 4.117 (Algebraic multiplicity)
The number of times an eigen value appears in the factorization of the characteristic polynomial
of a square matrix
Theorem 4.119 (Geometric and algebraic multiplicities)
Let
Corollary 4.16 (Distinct eigen values and multiplicity)
If an
Proof. The algebraic multiplicity of an eigen value is greater than or equal to 1.
But the sum cannot exceed
Corollary 4.17
Let an
Moreover if
then
Remark 4.18 (Spectrum from roots of characteristic polynomial)
In the factorization (4.12)
of the characteristic polynomial
Let us consider the sum of
With this there are
where
We will refer to the set (allowing repetitions)
4.9.4. The Zero Eigen Value#
Lemma 4.51 (
Proof. Let
Then there exists
such thatThus
is a non-trivial solution of the homogeneous linear system .Thus
is singular.
Converse:
Assume that
is singular.Then there exists
s.t.Thus
is an eigen value of .
Lemma 4.52 (Eigenspace of the zero eigen value)
If a square matrix
Proof. This is straight forward from the definition of eigen space (see Definition 4.114).
Remark 4.19
Clearly the geometric multiplicity of
4.9.5. Transpose#
Lemma 4.53 (Eigen values of the transpose)
Let
Proof. The eigen values of
But
Hence (using Lemma 4.33)
Thus the characteristic polynomials of
4.9.6. Powers and Inverses#
Lemma 4.54 (Power rule)
Let
Proof. We prove this by mathematical induction.
For
the statement holds trivially since is an eigen value of .Assume that the statement holds for some value of
.Thus let
be an eigen value of and let be corresponding eigen vector.Now
Thus
is an eigen value for with the same eigen vector .With the principle of mathematical induction, the proof is complete.
Lemma 4.55 (Eigenvalue of the inverse)
Let a square matrix
Proof. Since
Let
be an eigen vector of for the eigen value .Then
Thus
is also an eigen vector of for the eigen value .Now let
.Then
.Thus if
is an eigen value of then is an eigen value of .Thus if
is invertible then eigen values of and have one to one correspondence (being reciprocal).
4.9.7. Invariant Subspaces#
Definition 4.118 (Invariant subspace)
Let
In other words,
We also say that
Eigen vectors are generators of invariant subspaces.
Lemma 4.56 (Spans of eigenvectors are invariant)
Let
Then the column space of
Proof. Let us assume that
Let any vector
Then
Clearly
4.9.8. Triangular Matrices#
Lemma 4.57 (Eigenvalues of triangular matrices)
Let
Proof. We are given that
Then
is also triangular with its diagonal entries being .Using Lemma 4.38, we have
Clearly the roots of characteristic polynomial are
.
Several small results follow from this lemma.
Corollary 4.18 (Eigenvalues of triangular matrices)
Let
The characteristic polynomial of
isA scalar
is an eigen value of iff it is one of the diagonal entries of .The algebraic multiplicity of an eigen value
is equal to the number of times it appears on the main diagonal of .The spectrum of
is given by the distinct entries on the main diagonal of .
A diagonal matrix is naturally both an upper triangular matrix as well as a lower triangular matrix. Similar results hold for the eigen values of a diagonal matrix also.
Lemma 4.58 (Eigenvalues of diagonal matrices)
Let
Its eigen values are the entries on its main diagonal.
The characteristic polynomial of
isA scalar
is an eigen value of iff it is one of the diagonal entries of .The algebraic multiplicity of an eigen value
is equal to the number of times it appears on the main diagonal of .The spectrum of
is given by the distinct entries on the main diagonal of .
There is also a result for the geometric multiplicity of eigen values for a diagonal matrix.
Lemma 4.59 (Geometric multiplicity of eigenvalues for diagonal matrices)
Let
Proof. The unit vectors
They are linearly independent.
Thus if a particular eigen value appears
number of times, then there are linearly independent eigen vectors for the eigen value.Thus its geometric multiplicity is equal to the algebraic multiplicity.
4.9.9. Similar Matrices#
Some very useful results are available for similar matrices.
Lemma 4.60 (Characteristic polynomial of similar matrices)
The characteristic polynomial and spectrum of similar matrices is same.
Proof. Let
Then there exists an invertible matrix
such thatNow
Thus
is similar to .Hence due to Lemma 4.40, their determinant is equal.
In other words,
This means that the characteristic polynomials of
and are same.Since eigen values are nothing but roots of the characteristic polynomial, hence they are same too.
This means that the spectrum (the set of distinct eigen values) is same.
This result is very useful. Since if it can be shown that a matrix
Corollary 4.19 (Eigenvalues of similar matrices)
If
An eigen value has same algebraic and geometric multiplicity for both
and .The (not necessarily distinct) eigen values of
and are same.
Although the eigen values are same, but the eigen vectors are different.
Lemma 4.61 (Eigenvectors of similar matrices)
Let
for some invertible matrix
Proof. We are given that
Thus we have
Thus
Now
and is non singular.Thus
.Thus
is an eigen vector of .
4.9.10. Linear Independence of Eigenvectors#
Theorem 4.120 (Linear independence of eigenvectors with distinct eigenvalues)
Let
Proof. We first prove the simpler case with 2 eigen vectors
Let there be a linear relationship between
and given byMultiplying both sides with
we getSince
and , hence .Similarly by multiplying with
on both sides, we can show that .Thus
and are linearly independent.Now for the general case, consider a linear relationship between
given byMultiplying by
and using the fact that if , we get .Thus the only linear relationship is the trivial relationship.
This completes the proof.
For eigen values with geometric multiplicity greater than
Theorem 4.121 (Linear independence of eigenvectors from different eigen spaces)
Let
consisting of
This result puts an upper limit on the number of linearly independent eigen vectors of a square matrix.
Lemma 4.62 (Maximum number of linearly independent eigenvectors)
Let
Moreover if
then a set of
4.9.11. Diagonalization#
Diagonalization is one of the fundamental operations in linear algebra. This section discusses diagonalization of square matrices in depth.
Definition 4.119 (diagonalizable matrix)
An
Remark 4.20
We note that if we restrict to real matrices, then
In other words, if
The next theorem is the culmination of a variety of results studied so far.
Theorem 4.122 (Properties of diagonalizable matrices)
Let
which equals the number of nonzero entries on the main diagonal of . . .The characteristic polynomial of
isThe spectrum of
comprises the distinct scalars on the diagonal entries in .The (not necessarily distinct) eigenvalues of
are the diagonal elements of .The columns of
are (linearly independent) eigenvectors of .The algebraic and geometric multiplicities of an eigenvalue
of equal the number of diagonal elements of that equal .
Proof. From Definition 4.119 we note that
Due to Lemma 4.40
Due to Lemma 4.39
Now due to Lemma 4.31
Further due to Lemma 4.60 the characteristic polynomial and spectrum of
and are same.Due to Lemma 4.58 the eigen values of
are nothing but its diagonal entries.Hence they are also the eigen values of
.We recall that
Now writing
we have
Matching columns, we see that
Thus
are eigen vectors of .Since
is invertible, hence its columns are linearly independent.Hence all the eigenvectors of
are linearly independent.Since the characteristic polynomials of
and are same, hence the algebraic multiplicities of eigen values are same.From Lemma 4.61 we get that there is a one to one correspondence between the eigen vectors of
and through the change of basis given by .Thus the linear independence relationships between the eigen vectors remain the same.
Hence the geometric multiplicities of individual eigenvalues are also the same.
This completes the proof.
So far we have verified various results which are available if
a matrix
Theorem 4.123 (Diagonalizability: necessary and sufficient conditions)
An
Proof. We note that since
The necessary condition part was proven in Theorem 4.122.
We now show that if
Let the columns of
be and corresponding (not necessarily distinct) eigen values be .Then
Thus by letting
, we haveSince columns of
are linearly independent, hence is invertible.This gives us
Thus
is similar to a diagonal matrix .Hence
is diagonalizable.This validates the sufficient condition.
A corollary follows.
Corollary 4.20 (Diagonalizability and linear independence of eigen vectors)
An
Now we know that geometric multiplicities of eigen values of
Corollary 4.21 (Diagonalizability and geometric multiplicities)
Let
4.9.12. Real Symmetric Matrices#
This subsection is focused on real symmetric matrices.
Following is a fundamental property of real symmetric matrices.
Theorem 4.124 (Existence of eigenvalues)
Every real symmetric matrix has an eigen value.
The proof of this result is beyond the scope of this book.
Lemma 4.63 (Orthogonality of eigenvectors for distinct eigenvalues)
Let
Proof. By definition we have
Thus
Definition 4.120 (Orthogonally diagonalizable matrix)
A real
is a real diagonal matrix.
Lemma 4.64
Every orthogonally diagonalizable matrix
Proof. We have a diagonal matrix
Taking transpose on both sides we get
Thus
Theorem 4.125
Every symmetric matrix
We skip the proof of this theorem.
4.9.13. Hermitian Matrices#
Following is a fundamental property of Hermitian matrices.
Theorem 4.126 (Existence of eigen values for Hermitian matrices)
Every Hermitian matrix has an eigen value.
The proof of this result is beyond the scope of this book.
Lemma 4.65 (Real eigen values for Hermitian matrices)
The eigenvalues of a Hermitian matrix are real.
Proof. Let
Thus
Lemma 4.66 (Orthogonality of eigenvectors for distinct eigenvalues)
Let
Proof. By definition we have
Thus
Definition 4.121 (Unitary diagonalizable matrix)
A complex
is a complex diagonal matrix.
Lemma 4.67
Let
Proof. We have a real diagonal matrix
Taking conjugate transpose on both sides we get
Thus
Theorem 4.127
Every Hermitian matrix
We skip the proof of this theorem.
The theorem means that if
Definition 4.122 (Eigen value decomposition of a Hermitian matrix)
Let
Let
If
Remark 4.21
Let
Now if
Also
Lemma 4.68
Let
Proof.
Let
and let .Clearly
.Then
From previous remark we have
Thus we get
4.9.14. Miscellaneous Properties#
This subsection lists some miscellaneous properties of eigen values of a square matrix.
Lemma 4.69
Proof. We can see that
Thus
4.9.15. Semidefinite, Definite and Indefinite Matrices#
4.9.15.1. Positive Semidefinite Matrices#
Definition 4.123 (Positive semidefinite matrix)
A symmetric matrix
The notation
We define the set of symmetric positive semidefinite matrices as
“positive semidefinite” is often abbreviated as “p.s.d.”.
Theorem 4.128
Let
Proof. We shall just prove this for
We note that
Thus,
is symmetric. . Since it is symmetric, hence .Let
.Then,
where
is the norm induced by the dot product on .Thus,
is p.s.d..
Theorem 4.129 (Non-negativity of the diagonal elements of p.s.d. matrices)
Let
Proof. Let
Then, for every nonzero
, .In particular, this is true for standard unit vectors.
But
.Hence,
must be true.
Definition 4.124 (Square root of a positive semidefinite matrix)
Let
Let
The matrix
The definition of
We can see that
4.9.15.2. Positive Definite Matrices#
Definition 4.125 (Positive definite matrix)
A symmetric matrix
The notation
We define the set of symmetric positive definite matrices as
“positive definite” is often abbreviated as “p.d.”.
Theorem 4.130 (Positivity of the diagonal elements of p.d. matrices)
Let
Proof. Let
Then, for every nonzero
, .In particular, this is true for standard unit vectors.
But
.Hence,
must be true.
Theorem 4.131 (Principal minors criterion)
A matrix
In other words,
4.9.15.3. Negative Semidefinite Matrices#
Definition 4.126 (Negative semidefinite matrix)
A symmetric matrix
The notation
We define the set of symmetric negative semidefinite matrices as
“negative semidefinite” is sometimes abbreviated as “n.s.d.”.
4.9.15.4. Negative Definite Matrices#
Definition 4.127 (Negative definite matrix)
A symmetric matrix
The notation
We define the set of symmetric negative definite matrices as
“negative definite” is sometimes abbreviated as “n.d.”.
4.9.15.5. Indefinite Matrices#
Definition 4.128 (Indefinite matrix)
A symmetric matrix
Indefinite matrices are neither positive semidefinite nor negative semidefinite.
Theorem 4.132 (Diagonal elements and indefiniteness)
Let
Proof. Each diagonal element corresponds to
4.9.15.6. Eigenvalue Decomposition#
Theorem 4.133 (Eigenvalue characterization theorem)
Let
is positive definite if and only if all its eigenvalues are positive. is positive semidefinite if and only if all its eigenvalues are nonnegative. is negative definite if and only if all its eigenvalues are negative. is negative semidefinite if and only if all its eigenvalues are nonpositive. is indefinite if and only if at least one eigenvalue is positive and at least one eigenvalue is negative.
Proof. Let the eigenvalue decomposition of
We prove (1).
Let
.Then
where
.Since
is nonsingular, hence for every if and only if for every .Plugging
for , we see that is a necessary condition.Also, if
for every , then, the sum is positive for every nonzero since at least one . Hence, it is a sufficient condition.
Similar arguments apply for other statements.
4.9.15.7. Trace and Determinant#
Corollary 4.22 (Trace and determinant of positive definite matrices)
Let
Proof. Trace is the sum of eigenvalues. Determinant is the product of eigenvalues.
If
Corollary 4.23 (Trace and determinant of positive semidefinite matrices)
Let
4.9.16. Diagonally Dominant Matrices#
Definition 4.129 (Diagonally dominant matrix)
Let
is called diagonally dominant iffor every
. In other words, the absolute value of the diagonal entry in a row is greater than or equal to the sum of absolute values of non diagonal entries in the row. is called strictly diagonally dominant iffor every
. In other words, the absolute value of the diagonal element is bigger than the sum of absolute values of all the off diagonal elements on that row.
Example 4.28 (Strictly diagonally dominant matrix)
Let us consider
We can see that the strict diagonal dominance condition is satisfied for each row as follows:
Theorem 4.134 (Positive semidefiniteness of diagonally dominant matrices)
Let
If
is diagonally dominant whose diagonal entries are nonnegative then is positive semidefinite.If
is strictly diagonally dominant whose diagonal entries are positive then is positive definite.
Proof. Assume that
For contradiction, assume that
is not positive semidefinite.Then, there is an eigen value
and corresponding eigenvector .Consider the absolute values of entries of
given by .Let
denote the index of the largest absolute value entry of .We also have
.For the
-th row, we get the equalityThus,
.But
is nonnegative and is negative, hence this reduces to or .We have arrived at a contradiction.
Thus,
must be positive semidefinite.
Now, assume that
By first part, it is clear that
is p.s.d..We just need to show that all eigenvalues are positive. There are no zero eigenvalues.
For contradiction, assume that
is indeed an eigenvalue of .Let
be corresponding eigenvector satisfying .Let
denote the index of the largest absolute value entry of .Following the earlier argument
This is impossible. Hence,
must be positive definite.
Strictly diagonally dominant matrices have a very special property. They are always non-singular. The following result is valid for both real and complex strictly diagonally dominant matrices.
Theorem 4.135
Strictly diagonally dominant matrices are non-singular.
Proof. For contradiction, suppose that
Then there exists a vector
with such thatLet
We first show that every entry in
cannot be equal in magnitude.For contradiction, let us assume that there exists
such thatSince
hence .We can write
for every .Now for any row
in , we haveby triangle inequality.
But this contradicts our assumption that
is strictly diagonally dominant.Thus all entries in
are not equal in magnitude.Let us now assume that the largest entry in
lies at index with .Without loss of generality we can scale down
by to get another vector in which all entries are less than or equal to 1 in magnitude while -th entry has a magnitude of .In other words,
and for all other entries.From
we get for the -th rowwhich again contradicts our assumption that
is strictly diagonally dominant.Hence strictly diagonally dominant matrices are non-singular.
4.9.17. Gershgorin’s Theorem#
We are now ready to examine Gershgorin’ theorem which provides very useful bounds on the spectrum of a square matrix.
Theorem 4.136 (Gershgorin’s circle theorem)
Every eigen value
Proof. The proof is a straight forward application of non-singularity of diagonally dominant matrices.
We know that for an eigen value
,In other words, the matrix
is singular.Hence it cannot be strictly diagonally dominant due to Theorem 4.135.
Thus looking at each row
of we can say thatcannot be true for all rows simultaneously.
In other words, this condition must fail at least for one row.
This means that there exists at least one row
for whichholds true.
What this theorem means is pretty simple.
Consider a disc in the complex plane for the
-th row of whose center is given by and whose radius is given by i.e. the sum of magnitudes of all non-diagonal entries in -th row.There are
such discs corresponding to rows in .(4.13) means that every eigen value must lie within the union of these discs. It cannot lie outside.
This idea is crystallized in following definition.
Definition 4.130 (Gershgorin’s disc)
For
is called the
We note that the definition is equally valid for real as well as complex matrices. For real matrices, the centers of disks lie on the real line. For complex matrices, the centers may lie anywhere in the complex plane.
Clearly there is nothing extraordinary about the rows of
Theorem 4.137 (Gershgorin’s circle theorem for columns)
Every eigen value of a matrix
with
Proof. We know that eigen values of