Eigenvalues / Eigenvectors

Table of contents
  1. Characteristic Polynomial
    1. Roots of the Characteristic Polynomial
  2. Eigenvalues and Eigenvector
    1. Finding Eigenvalues via Characteristic Polynomial
      1. Eigenvectors and the Kernel
      2. Rank of AλI
      3. Determinant of AλI
      4. Solving the Characteristic Polynomial
    2. Eigenvalues of a Transpose
  3. Defective Matrix
    1. Linear Independence of Eigenvectors
  4. Diagonalizable Matrix (Non-Defective Matrix)
    1. Eigen-Decomposition
    2. Determinant of a Matrix Using Eigenvalues
    3. Trace of a Matrix Using Eigenvalues
    4. Spectral Theorem
  5. Eigenspace
    1. Geometric Multiplicity

Characteristic Polynomial

For λR and square matrix ARn×n,

Remeber λ is a variable here.

The characteristic polynomial of A is:

pA(λ)=det(AλI)=(1)nλn+cn1λn1++c1λ+c0=(1)nλn+(1)n1tr(A)λn1++(1)0det(A)

where ciR.

Sometimes it is defined like... pA(λ)=det(λIA)

The only difference is that this version guarantees the leading coefficient is positive, or monic:

pA(λ)=λn+cn1λn1++c1λ+c0

Unlike the (1)n above.

Roots of the Characteristic Polynomial

This is literally the roots λ you get from solving the above polynomial equation:

pA(λ)=0

Also called eigenvalues of A.


Eigenvalues and Eigenvector

Let ARn×n.

For the eigenvalue equation:

Ax=λx

The satisfying non-zero vector xRn{0} and corresponding scalar λR are called eigenvectors and eigenvalues of A, respectively.

Intuition

Intuitively, it means that for some linear transformation represented by A, there is a vector that is unchanging in any other way except for the magnitude and maybe reversed direction (if λ<0).

In other words, this transformation A can only scale the eigenvectors by its eigenvalues.

Some like to order the eigenvalues in descending order, and call them first, second, etc. eigenvalues.

Finding Eigenvalues via Characteristic Polynomial

Eigenvectors and the Kernel

The above equation can be rewritten as:

(AλI)x=0

or equivalently:

(λIA)x=0

With this homogenous system, we see that the non-trivial solution x (or eigenvectors) are in the kernel of the matrix (AλI):

xker(AλI)

Rank of AλI

We saw that the eigenvector is a non-trivial element of the kernel.

Which means that the columns of AλI are linearly dependent (they can zero themselves out by a linear combination).

Therefore, AλI cannot be full rank:

rank(AλI)<n

Determinant of AλI

From here, we know that the determinant of the matrix is non-zero if and only if the matrix is full rank.

Matrix is invertible only when columns are linearly independent, but since it is not, determinant should be zero for the inverse to be undefined.

However, we just saw that AλI cannot be full rank.

Hence:

det(AλI)=0

Solving the Characteristic Polynomial

Because we know

det(AλI)=0

we can use the characteristic polynomial to find the eigenvalues λ.

Eigenvalues of a Transpose

AT and A have the same eigenvalues.

However, the eigenvectors are not necessarily the same.


Defective Matrix

A square matrix ARn×n is defective if it does not have n linearly independent eigenvectors.

Linear Independence of Eigenvectors

Eigenvectors corresponding to distinct eigenvalues are linearly independent.

Therefore, eigenvectors with n distinct eigenvalues form a basis of Rn.


Diagonalizable Matrix (Non-Defective Matrix)

We say that a matrix ARn×n is diagonalizable (or non-defective) if it is similar to a diagonal matrix D.

It is saying that we can make A into a diagonal matrix by changing its basis.

In other words, there exists an invertible matrix P such that:

D=P1AP

This can be rewritten as:

A=PDP1

Eigen-Decomposition

Any non-defective matrix ARn×n can be decomposed into:

A=PDP1

where P is a matrix whose columns are the eigenvectors of A, and D is a diagonal matrix whose diagonal elements are the eigenvalues of A.

What?

If you rewrite the above equation as:

AP=PD

Let’s say P=[p1pn], then:

AP=[Ap1Apn]

Keeping in mind that D is a diagonal matrix, let the principal diagonal elements be c1,,cn, then:

PD=[c1p1cnpn]

Since Api=cipi,

We see that each pi is an eigenvector of A, and ci is the corresponding eigenvalue.

Determinant of a Matrix Using Eigenvalues

What follows is a quick way to calculate the determinant of a matrix usig its eigenvalues.

A property of the determinant is that it is invariant under change of basis.

So for similar matrices A and D, their determinants are the same, but it is easy to calculate the determinant of a diagonal matrix D, since it is just the product of its diagonal elements which are the eigenvalues of A.

Therefore the determinant of a matrix is the product of its eigenvalues:

det(A)=i=1nλi

λi is not necessarily unique! It can have repeated eigenvalues.

Trace of a Matrix Using Eigenvalues

Same logic applies to trace as it is also invariant under change of basis.

The trace of D is the sum of its diagonal elements (the eigenvalues of A).

Therefore the trace of a matrix is the sum of its eigenvalues:

tr(A)=i=1nλi

Again, repetition counts.

Spectral Theorem

A symmetric matrix ARn×n is always diagonalizable.

Further, the eigenvectors of A form an orthonormal basis:

PP=IP=P1

Therefore the decomposition is:

A=PDP


Eigenspace

Eigenvectors are not unique, that is, any scalar multiple of an eigenvector is also an eigenvector.

These eigenvectors that share the same eigenvalue span a subspace of Rn, called the eigenspace of A with respect to the eigenvalue λ (denoted with a subscript):

Eλ=ker(AλI)

Geometric Multiplicity

Let λ be an eigenvalue of square matrix A.

Then the geometric multiplicity of λ, is the number of linearly independent eigenvectors corresponding to λ, or

dim(Eλ)

Algebraic multiplicity is the multiplicity of λ as a root of the characteristic polynomial. For example, if a matrix had two repeated eigenvalues, then the algebraic multiplicity of that eigenvalue is 2.