Applied linear operators and spectral methods/Lecture 4

From testwiki
Jump to navigation Jump to search

More on spectral decompositions

In the course of the previous lecture we essentially proved the following theorem:

Theorem:

1) If a n×n matrix ๐€ has n linearly independent real or complex eigenvectors, the ๐€ can be diagonalized. 2) If ๐“ is a matrix whose columns are eigenvectors then ๐“๐€๐“1=Λ is the diagonal matrix of eigenvalues.

The factorization ๐€=๐“1Λ๐“ is called the spectral representation of ๐€.

Application

We can use the spectral representation to solve a system of linear homogeneous ordinary differential equations.

For example, we could wish to solve the system

d๐ฎdt=๐€๐ฎ=[2112][u1u2]

(More generally ๐€ could be a n×n matrix.)

Comment:

Higher order ordinary differential equations can be reduced to this form. For example,

d2u1dt2+adu1dt=bu1

Introduce

u2=du1dt

Then the system of equations is

du1dt=u2du2dt=bu1au2

or,

d๐ฎdt=[01ba][u1u2]=๐€๐ฎ

Returning to the original problem, let us find the eigenvalues and eigenvectors of ๐€. The characteristic equation is

det(๐€λ๐ˆ)=0

o we can calculate the eigenvalues as

(2+λ)(2+λ)1=0λ2+4λ+3=0λ1=1,λ2=3

The eigenvectors are given by

(๐€λ1๐ˆ)๐ง1=๐ŸŽ;(๐€λ2๐ˆ)๐ง2=๐ŸŽ

or,

n11+n21=0;n11n21=0;n12+n22=0;n12+n22=0

Possible choices of ๐ง1 and ๐ง2 are

๐ง1=[11];๐ง2=[11]

The matrix ๐“ is one whose columd are the eigenvectors of ๐€, i.e.,

๐“=[1111]

and

Λ=๐“1๐€๐“=[1003]

If ๐ฎ=๐“๐ฎ' the system of equations becomes

d๐ฎ'dt=๐“1๐€๐“๐ฎ'=Λ๐ฎ'

Expanded out

du1'dt=u1';du2'dt=3u2'

The solutions of these equations are

u1'=C1et;u2'=C2e3t

Therefore,

๐ฎ=๐“๐ฎ'=[C1et+C2e3tC1etC2e3t]

This is the solution of the system of ODEs that we seek.

Most "generic" matrices have linearly independent eigenvectors. Generally a matrix will have n distinct eigenvalues unless there are symmetries that lead to repeated values.

Theorem

If ๐€ has k distinct eigenvalues then it has k linearly independent eigenvectors.

Proof:

We prove this by induction.

Let ๐งj be the eigenvector corresponding to the eigenvalue λj. Suppose ๐ง1,๐ง2,,๐งk1 are linearly independent (note that this is true for k = 2). The question then becomes: Do there exist α1,α2,,αk not all zero such that the linear combination

α1๐ง1+α2๐ง2++αk๐งk=0

Let us multiply the above by (๐€λk๐ˆ). Then, since ๐€๐งi=λi๐งi, we have

α1(λ1λk)๐ง1+α2(λ2λk)๐ง2++αk1(λk1λk)๐งk1+αk(λkλk)๐งk=๐ŸŽ

Since λk is arbitrary, the above is true only when

α1=α2==αk1=0

In thast case we must have

αk๐งk=๐ŸŽαk=0

This leads to a contradiction.

Therefore ๐ง1,๐ง2,,๐งk are linearly independent.

Another important class of matrices which are diagonalizable are those which are self-adjoint.

Theorem

If ๐‘จ is self-adjoint the following statements are true


  1. ๐‘จ๐ฑ,๐ฑ is real for all ๐ฑ.
  2. All eigenvalues are real.
  3. Eigenvectors of distinct eigenvalues are orthogonal.
  4. There is an orthonormal basis formed by the eigenvectors.
  5. The matrix ๐‘จ can be diagonalized (this is a consequence of the previous statement.)


Proof

1) Because the matrix is self-adjoint we have

๐‘จ๐ฑ,๐ฑ=๐ฑ,๐‘จ๐ฑ

From the property of the inner product we have

๐ฑ,๐‘จ๐ฑ=๐‘จ๐ฑ,๐ฑ

Therefore,

๐‘จ๐ฑ,๐ฑ=๐‘จ๐ฑ,๐ฑ

which implies that ๐‘จ๐ฑ,๐ฑ is real.

2) Since ๐‘จ๐ฑ,๐ฑ is real, ๐‘ฐ๐ฑ,๐ฑ=๐ฑ,๐ฑ is real. Also, from the eiegnevalue problem, we have

๐‘จ๐ฑ,๐ฑ=λ๐ฑ,๐ฑ

Therefore, λ is real.

3) If (λ,๐ฑ) and (μ,๐ฒ) are two eigenpairs then

λ๐ฑ,๐ฒ=๐‘จ๐ฑ,๐ฒ

Since the matrix is self-adjoint, we have

λ๐ฑ,๐ฒ=๐ฑ,๐‘จ๐ฒ=μ๐ฑ,๐ฒ

Therefore, if λμ0, we must have

๐ฑ,๐ฒ=0

Hence the eigenvectors are orthogonal.

4) This part is a bit more involved. We need to define a manifold first.

Linear manifold

A linear manifold (or vector subspace) โ„ณ๐’ฎ is a subset of ๐’ฎ which is closed under scalar multiplication and vector addition.

Examples are a line through the origin of n-dimensional space, a plane through the origin, the whole space, the zero vector, etc.

Invariant manifold

An invariant manifold โ„ณ for the matrix ๐‘จ is the linear manifold for which ๐ฑโ„ณ implies ๐‘จ๐ฑโ„ณ.

Examples are the null space and range of a matrix ๐‘จ. For the case of a rotation about an axis through the origin in n-space, invaraiant manifolds are the origin, the plane perpendicular to the axis, the whole space, and the axis itself.

Therefore, if ๐ฑ1,๐ฑ2,,๐ฑm are a basis for โ„ณ and ๐ฑm+1,,๐ฑn are a basis for โ„ณ (the perpendicular component of โ„ณ) then in this basis ๐‘จ has the representation

๐‘จ=[xx|xxxx|xx00|xx00|xx]

We need a matrix of this form for it to be in an invariant manifold for ๐‘จ.

Note that if โ„ณ is an invariant manifold of ๐‘จ it does not follow that โ„ณ is also an invariant manifold.

Now, if ๐‘จ is self adjoint then the entries in the off-diagonal spots must be zero too. In that case, ๐‘จ is block diagonal in this basis.

Getting back to part (4), we know that there exists at least one eigenpair (λ1,๐ฑ1) (this is true for any matrix). We now use induction. Suppose that we have found (n1) mutually orthogonal eigenvectors ๐ฑi with ๐‘จ๐ฑi=λi๐ฑi and λi are real, i=1,,k1. Note that the ๐ฑis are invariant manifolds of ๐‘จ as is the space spanned by the ๐ฑis and so is the manifold perpendicular to these vectors).

We form the linear manifold

โ„ณk={๐ฑ|๐ฑ,๐ฑj=0j=1,2,,k1}

This is the orthogonal component of the k1 eigenvectors ๐ฑ1,๐ฑ2,,๐ฑk1 If ๐ฑโ„ณk then

๐ฑ,๐ฑj=0and๐‘จ๐ฑ,๐ฑj=๐ฑ,๐‘จ๐ฑj=λj๐ฑ,๐ฑj=0

Therefore ๐‘จ๐ฑโ„ณk which means that โ„ณk is invariant.

Hence โ„ณk contains at least one eigenvector ๐ฑk with real eigenvalue λk. We can repeat the procedure to get a diagonal matrix in the lower block of the block diagonal representation of ๐‘จ. We then get n distinct eigenvectors and so ๐‘จ can be diagonalized. This implies that the eigenvectors form an orthonormal basis.

5) This follows from the previous result because each eigenvector can be normalized so that ๐ฑi,๐ฑj=δij.

We will explore some more of these ideas in the next lecture. Template:Lecture