Sei sulla pagina 1di 11

Second Edition (March 2011)

These rough notes are for the Vector Spaces part of CIE Further Mathematics Alevel, since
these topics are not covered in many books. I hope they are helpful. I have included proofs and
explanations for most things which can be understood by students, even if they are not on the
syllabus.
Dr. Zen Harper, CIE centre at Shanghai Experimental School
Email:
zen.harper@cantab.net

V ECTOR SPACES , SPAN , LINEAR INDEPENDENCE , BASES


Definition 1 A vector space V is a collection of objects with the following properties:
 There is some special element 0 V (the zero vector) and some operation +, meaning:
for any u, v V , there is some u + v V . We have the following properties (for all
u, v, w V ):
u + v = v + u,

(u + v) + w = u + (v + w),

u + 0 = u.

 For every u V , there is some element u V such that u + (u) = 0.


 For every number R and every u V , there is some u V . We have the following
properties (for all u, v V and all , R):
1u = u,

0u = 0,

(u + v) = u + v,

( + )u = u + u,

(u) = ()u.

With these properties, we can prove that u is UNIQUE, and (1)u = u. We write u v
instead of u + (v).
T HE MOST OBVIOUS EXAMPLE : Rn , the space of all column vectors of height n.
OTHER EXAMPLES : see Example 17 below. The set of all (realvalued) functions y satd2 y
dy
isfying 2
6y = 0 is a vector space. T HIS IS ONE OF THE REASONS WHY VECTOR
dx
dx
SPACES ARE SO IMPORTANT ! This is a subspace of the vector space of all continuous functions
f : R R.
(In this case, the space has dimension 2, spanned by the functions e3x , e2x ; see later for the
definition of dimension).
Definition 2 Let V be a vector space and W V . We say that W is a vector subspace of V if
W is itself a vector space. We only need to check that a, b W = a + b W , a W
for every R.
 

 

x
x
Example 3 In R , consider the sets A =
: x = 13y , B =
: 14x + 3y = 0 ,
y
y
 

 

 

x
x
x
2
2
C=
:x+y =1 ,D =
:x +y =1 ,E =
: x, y Z .
y
y
y
 
 
u1
v
2
Then A, B are vector subspaces of R , e.g. let u =
, v = 1 A. Then u1 = 13u2 ,
 u2
 v2
u1 + v1
v1 = 13v2 , so u1 + v1 = 13(u2 + v2 ); thus u + v =
A. Also u1 = 13(u2 ), so
u2 + v2
u A.
2

C, D are NOT subspaces because 0 6 C, D. E is not a subspace because a E does not


imply 12 a E.
If we defined a set F similarly by the equation xy = 0, then a F = a F , but
a, b F does not imply a + b F , so F is not a subspace.

x
y R3 :

2x y + 3z = 0,

x + y + 4z = 0,

x 4y + 2z = 0

is a vector subspace of R3 .

Every vector subspace of R3 is either: {0} (a point), a line through 0, a plane through 0, or R3
itself.
Definition 4 Let u1 , u2 , u3 , . . . , V . A linear combination of these vectors is
some N and numbers 1 , 2 , 3 , . . . , N R. Their span (or linear span) is

PN

j=1

j uj , for

span{u1 , u2 , u3 , . . .} = { all linear combinations of u1 , u2 , u3 , . . . }


N OTE : the collection of vectors u1 , u2 , . . . could be infinite; however, a linear combination
must be a sum of only finitely many vectors.
It is obvious that the span of every (nonempty) collection of vectors is a vector subspace.



 
 
1
17
x
2
2
Example 5 Let a =
,b =
in R . Then span {a, b} = R , because
=
1
0
y
1
(y)a + 17
(x + y) b for every x, y R.
 
 
 

1
3
x
Let c =
, d =
in R2 . Then span{c, d} =
R2 : y = 2x =
2
6
y
span{c} = span{d}.
Definition 6 u1 , u2 , u3 , . . . , uN V are called linearly independent if:
N
X

j uj = 0

j = 0 for all j.

j=1

For example: every plane in R3 passing through (0, 0, 0) is the span of two linearly independent vectors.



1
0
0

0 are linearly independent in R3 .


Example 7 a = 1 , b = 1 , c =
0
2
3

Proof: Let a + b + c = 0. Then + = 0, so = 0, + = 0, 2 3 = 0.


2 3
Thus = = = 0 (check this calculation yourself!)



1
0
2

1 , c = 1 are NOT linearly independent in R3 .


Example 8 a = 2 , b =
1
2
4

Proof: 2a + 3b = c, so 2a + 3b + (1)c = 0.
Also, any u1 , u2 , u3 , . . . , uN with uj = 0 for some j is not linearly independent.
P ROOF : 0u1 + 0u2 + + 1uj + 0uj+1 + + 0uN = 0, but 0, 0, . . . , 1, 0, . . . , 0 are not
all zero.
Exercise 9 If a, b, c R3 are orthogonal and nonzero, so that a b = a c = b c = 0 and
a, b, c 6= 0, then a, b, c are linearly independent.

x11
x12
x1n
x11 x12
x21 x22
x2n are linearly
x21 x22

Fact 10 .. , .. , . . . , .. independent det ..


..
...
. .
.
.
.
in Rn
xn1
xn2
xnn
xn1 xn2

x1n
x2n

.. 6= 0.
.
xnn

Definition 11 A basis of V is a linearly independent collection of vectors u1 , u2 , u3 , . . . V


such that span{u1 , u2 , u3 , . . .} = V .
Theorem 12PLet u1 , u2 , u3 , . . . be a basis of V . Then every a V is a UNIQUE linear combination a = j j uj .
Proof: By definition, 1 , 2 , 3 , . . . EXIST, because u1 , u2 , u3 , . . . span V . If also a =
for some 1 , 2 , . . . then
X
X
X
0=aa=
j uj
j uj =
(j j )uj ,
j

j uj

so j j = 0 for each j because u1 , u2 , u3 , . . . are linearly independent. Therefore j = j


for each j. Therefore 1 , 2 , 3 , . . . are UNIQUE.
Exercise 13 Let A be any n n invertible matrix, and let v1 , v2 , . . . , vn Rn . Then:
v1 , v2 , . . . , vn is a basis of Rn

Av1 , Av2 , . . . , Avn is a basis of Rn .


Definition 14 A vector space V is called finite dimensional if there is a basis of finitely many
vectors u1 , u2 , . . . , un V .
Fact 15 If V is finite dimensional, then every basis of V has the same number n of vectors;
then n = dim(V ) is called the DIMENSION of V .



1
0
0
Example 16 dim(Rn ) = n. For n = 3, a basis is e1 = 0, e2 = 1, e3 = 0.
0
0
1
n
th
More generally, the standard basis for R is e1 , e2 , . . . , en , where (ei )i = 1 (the i row), and
(ei )j = 0 if i 6= j.
 InExample 3 above, dim(A) = dim(B) = 1. One possible basis for A is the single vector
13
. (But any nonzero vector in A will also give a basis). One possible basis for B is the
1


3/7
single vector
. In Example 5 above, {a, b} is a basis for R2 .
2

Example 17 Let Y be the set of all functions y : R R which satisfy


Then Y is a finitedimensional vector space with dim(Y ) = 2.

d2 y
dy
6y = 0.

2
dx
dx

Proof: From Differential Equations: let y1 = e3x , y2 = e2x . Then {y1 , y2 } is a basis of Y .
Fact 18 Let U, V be vector spaces with U V , and V finitedimensional. T HEN U is finite
dimensional with dim(U ) 6 dim(V ). Also
dim(U ) = dim(V ) U = V .
Theorem 19 Let u1 , u2 , . . . , um V be linearly independent. Then:
(i) u1 , u2 , . . . , um is a basis for span{u1 , u2 , . . . , um }.
(ii) If m = dim(V ) then u1 , u2 , . . . , um is a basis for V .
(iii) Every collection of vectors a, b, c, . . . V containing more than dim(V ) vectors is not
linearly independent.
Proof: (i) by definition. (ii) by Fact 18 (since the span is a subspace of V with the same
dimension). In (iii), the vectors a, b, c, . . . cannot be linearly independent (because if they
were, then their span would be a subspace of V with higher dimension, contradicting Fact 18).
Definition 20 For any matrix A, the column space is the linear span of the columns of A. Thus,
it is a vector space of column vectors. The row space
 is the linear span of the rows of A, and
thus a vector space of row vectors x1 x2 xn .

3 1 0

Example 21 The column space of 0 1 0


0 0 0


3 + 0 + 2 : , ,

1
0 is R3 , and the row space is
2



R = x y 0 z : x, y, z R .

The same set of vectors can have many different descriptions, but notice that the second description of the row space using x, y, z is simpler than the first description using , , .
Proof: (CHECK THESE CALCULATIONS !) The column space is the set of all vectors



3
1
0
1
3 +

0 + 1 + 0 + 0 =
0
0
0
2
2

x
3

over all , , , R. This is clearly a subset of R . Conversely, given any vector y R3 ,


z
we can find , , , such that x = 3 + , y = , z = 2. This is easy: let = y, = 12 y,
then let = 13 (x + ), and can be anything: the value of is unimportant.
For the row space, we have to show two things:
 first, given any
 , , R, there are some
x, y, z R such that 3 + 0 + 2 = x y 0 z . This is obvious: let x = 3,
etc. Second, given any x, y, z R, there are some , , R such that it holds. This is also
fairly easy: let = 13 x, then = y , then finally = 12 (z + ).
I MPORTANT: this row space is NOT a subspace of R4 , because Rn uses columns, not rows.
4

R ANGE SPACE , RANK , NULL SPACE ,

NULLITY, LINEAR
TRANSFORMATIONS

Definition 22 Let V, W be vector spaces, and T : V W any function. Then T is called a


linear transformation if
for all a, b V ,

T (a + b) = T (a) + T (b)

, R.

Theorem 23 Let T : Rn Rm be any linear transformation. There is a unique matrix A, with


m rows and n columns, such that T (v) = Av.
Conversely, if B is any matrix with m rows and n columns, then the function S : Rn Rm
given by S(v) = Bv is a linear transformation.
Proof: Obvious. The column vectors of A are exactly T e1 , T e2 , . . . , T en , where e1 , e2 , . . . , en
is the standard basis of Rn .
Definition 24 For any linear transformation T : V W , we have
Range space of T = {T v : v V },

Rank of T = dim( Range space of T ),

Null space of T = {v V : T v = 0},

Nullity of T = dim( Null space of T ).

Obviously the null space and range space are vector subspaces of V and W , respectively.
Theorem 25 Let T : Rn Rm be a linear transformation represented by the matrix A. Then:
Range space of T = Column space of A
Proof: Obvious, since Aej is the j th column of A, with e1 , e2 , . . . , en the standard basis of
Rn . (I already said this in the proof of Theorem 23).
Fact (RankNullity Formula): dim(V ) = dim(Null space of T ) +
 dim(Range space of T ).
Proof: (***) From undergraduate Linear Algebra, V = V /null(T ) null(T ), so


dim(V ) = dim V /null(T ) + dim null(T ) .
But T : V W induces anisomorphism Te : V /null(T ) T (V ), given by Te(v + null(T )) =
T v. Hence dim V /null(T ) = dim(T (V )).
Example 26 (Invertibility can never work for nonsquare matrices) Let A, B, C be matrices with AB = I (m m) and CA = I (n n). Then WE MUST HAVE n = m.
Proof: A, B, C must represent linear transformations A : Rn Rm and B, C : Rm Rn .
The range space of A is Rm , because u = Iu = (AB)u = A(Bu) for every u Rm .
The null space of A is {0}, because if Av = 0 then v = Iv = (CA)v = C(Av) = C0 = 0.
Thus

n = dim(Rn ) = dim(Null space of A) + dim(Range space of A)


= dim({0}) + dim(Rm ) = 0 + m = m.

Theorem 27 If A is a SQUARE matrix, then A is invertible the null space of A is {0}.

Proof: Let u be in the null space, so that Au = 0. If also A is invertible, then u = A1 0 = 0.


Conversely, suppose that the null space of A is {0}. Let A have size n n, so that A defines
a function A : Rn Rn . Now
Au = Av

Au Av = 0 A(u v) = 0
u v is in the null space of A
u v = 0

u = v.

So A, as a function on Rn , is onetoone: different vectors are sent to different vectors.


Secondly, the nullity of A is zero, so the rank of A is n 0 = n. Thus A sends Rn onto Rn .
Thus A is onetoone and onto, so defines an INVERTIBLE function on Rn . Let the inverse
function be f : Rn Rn . Hence, f (Au) = u = Af (u) for all u Rn .
It is obvious that f : Rn Rn is linear. Therefore, f can be represented by some
n n matrix B. Therefore u = f (A(u)) = f (Au) = B(Au) = (BA)u for all u Rn ,
so BA = I [because the matrix representing a linear transformation is unique]. Similarly,
u = A(f (u)) = A(Bu) = (AB)u, so AB = I. Therefore B = A1 .
Theorem 28 If A, B are SQUARE matrices and AB = I, then A, B are invertible, B = A1 .
The definition of A1 says A1 A = AA1 = I. So this DOESN T FOLLOW just from the
definition of inverse! We DON T assume that BA = I, so we have to prove it!!
Proof: Since A, B are square, THEY ARE THE SAME SIZE (otherwise AB would not be
defined). Now:
E ITHER : ( RUBBISH PROOF !!) 1 = det(I) = det(AB) = det(A)det(B), so det(A) 6= 0, so
A is invertible; therefore A1 = A1 I = A1 (AB) = (A1 A)B = IB = B.
O R : ( PROPER PROOF !) if AB = I and v is in the null space of B, then v = Iv = (AB)v =
A(Bv) = A0 = 0. Therefore, the null space of B is {0}. Therefore by Theorem 27, B is
invertible.
So A = AI = A(BB 1 ) = (AB)(B 1 ) = IB 1 = B 1 , so A = B 1 . It is always true that
1
B is invertible with (B 1 )1 = B [PROVE THIS !!!!], thus A1 = B.


1 0 0
Warning W E MUST HAVE SQUARE MATRICES !!!!! e.g. if A =
and B =
1
0
1



1 0
1
0
0 0, then AB =
, but A, B are not invertible (they couldnt possibly be, since
0 1
1 1

1 0 0
they are not square!) Notice that BA = 0 0 0 is not the identity matrix.
0 0 1

ROW OPERATIONS
Given any matrix, there are 3 kinds of row operations on the rows R1 , R2 , R3 , . . .
1. Swap Ri , Rj , for some i 6= j;
2. Replace Ri by Ri , for some 6= 0;
3. Replace Ri by Ri + Rj , where i 6= j.

1 2 1
e.g. A = 2 3 4 ,
1 0 1

2 3 4
B = 1 2 1 ,
1 0 1

A B: swap R1 , R2 .

A C: replace R2 by 3R2 .

1 2 1
C = 6 9 12 ,
1 0 1

1 2 1
D = 0 1 6 .
1 0 1

A D: replace R2 by R2 2R1 .

Effect of the row operations on determinants: let M be a square n n matrix.


1. Changes det(M ) into det(M ).
2. Changes det(M ) into det(M ).
3. Does not change det(M ).
I MPORTANT: if M is an n n matrix, then det(M ) = n det(M ), not det(M ), because
we do operation 2 to each of the n rows.
I MPORTANT: we can do row operations on matrices of any size; however, for determinants,
we must have square matrices.
To invert a matrix A by row operations: Play a game: write A, I next to each other:
(AkI). Use row operations to try to change the lefthand matrix A into I, the identity matrix.
Each time, do the same row operations to the righthand matrix. You win when you get (IkB);
in this case, B = A1 .
If you ever get a noninvertible matrix on the left or right, then the original matrix A was
also noninvertible.
If A is invertible then it is always possible to win the game (and find A1 by this method).
However, if A is invertible then there are many different possible moves leading to a win.
E XPLANATION : each row operation corresponds to multiplying on the left by an invertible
matrix M . For example:

0 1 0
a1 a2 a3 a4
b1 b 2 b3 b4
1 0 0 b1 b2 b3 b4 = a1 a2 a3 a4 ,
0 0 1

1 0
a1 + b1 a2 + b2 a3 + b3 a4 + b4
a1 a2 a3 a4
0 1 0 b1 b2 b3 b4 = b1
b2
b3
b4 , . . .
0 0 1

Thus if M1 , M2 , M3 , . . . Mn are matrices representing our moves, then our game proceeds:
(AkI)

(M1 AkM1 )

(M2 M1 AkM2 M1 )
(Mn M2 M1 AkMn M2 M1 ) = (M AkM ),

so if ever we reach (IkB) = (M AkM ), then we have I = M A and B = M , so that A1 = B.

1 1 0
3 0 .
Example 29 Find A1 , where A = 1
3 1 1

1
1 1 0
1
0
0
1
1
0



1
3 0 0 1 0 0
4 0
1

3 1 1
0 0 1
3 1 1 0
7

One possible sequence of moves is:


0 0
1 1 0
1
0
0

1 0 0
1 0
1/4 1/4 0

0 1
3 1 1
0
0 1

1 0

0 1
3 1

1 0 0
3/4
0 1 0
1/4
0 0 1 10/4

3/4 1/4 0
1 0 0

1/4 1/4 0 0 1 0


0
0 1
3 0 1

1/4 0
3/4
1/4 0 , so finally A1 = 1/4
2/4 1
5/2

0
0
1

3/4
1/4 0

1/4 1/4 0

1/4 1/4 1

3 1 0
1/4 0
1/4 0 = 14 1 1 0.
10 2 4
1/2 1

ROW ECHELON FORM


Definition 30 A matrix is in row echelon form if the number of zero entries from the left in each
row increases (strictly) as we go down the rows (unless a whole row is zero, in which case all
lower rows remain zero also).
For example, the following matrices are in row echelon form:



1 0 2 1
0 11 3 1
0 2 17

B1 = 0 3 0 0 , B2 = 0 0 0 7 , B3 =
,
0 0 0
0 0 0 4
0 0 0 0

0 1
B4 = 0 0
0 0

but the following matrices are NOT in row echelon form:

0 0 1
4 0 0
1 1 1 0
0 2 0 , 0 0 1 , 0 0 1 1 .
0 0 0
0 0 2
0 1 0 0
Fact 31 If A is any matrix, we can use row operations to convert it to another matrix B in row
echelon form. B is called a row echelon form reduction of A. We always have
Rank of A = The number of rows of B which are not all zero.
Explanation (***) It is not hard to show that row operations do not change the rank (because
the matrix representing any row operation is invertible). Although, note that the range space
DOES change; it is only its dimension which is fixed. (Exercise: if V is a vector subspace of W
and T : W W is an invertible linear transformation, then dim(V ) = dim(T (V )).)
Thus Rank(A) = Rank(B). Now, the formula for Rank(B) becomes obvious if you think
long enough about how B acts on vectors, and use the Row Echelon Form property: its similar
to the proof that upper triangular matrices with nonzero diagonal entries are invertible by
back substitution starting from the lowest variables.
(Sorry this is just a messy bit of computation; I cant think of a really nice proof right now.
I think its possible to do something with induction and removing unnecessary columns, but it
still seems fairly messy).
Thus, the row echelon form is useful for calculating the rank of a matrix (and hence also the
nullity). In the example just above, B1 , B2 , B3 , B4 have ranks 3, 2, 1, 1, and nullities 1, 2, 2, 1
respectively (remember the RankNullity Formula above).

F INDING BASES FOR THE RANGE SPACE AND NULL SPACE


Let A be any matrix, B any row echelon form reduction of A.
Fact 32 If B has exactly k rows not identically zero, choose any linearly independent collection
of k columns from B. The corresponding k columns of A give a basis of the range space of A.
8

Proof: (***) Let bn1 , bn2 , . . . , bnk be the column vectors of columns n1 , n2 , . . . , nk of B,
and an1 , an2 , . . . , ank the corresponding column vectors of A. We already know that k is the
dimension of the range space (by Fact 31); so we only need to prove linear independence.
anj = Aenj and bnj = Benj for each j, where e1 , e2 , . . . are the standard basis vectors. Now
B = M A for some invertible M , so bnj = Benj = M Aenj = M (Aenj ) = M anj .
C LAIM : an1 , an2 , . . . , ank are linearly independent.
P
 P
P
P
P ROOF : let j j anj = 0. Then 0 = M

a
= j j M anj = j j bnj . Therej j nj
fore 1 = 2 = = k = 0 because bn1 , bn2 , . . . , bnk are linearly independent (by our
choice). Thus an1 , an2 , . . . , ank are linearly independent.
Example 33 let us reduce a matrix to row echelon form:

1 1 2 3 4
1 1 2 3 4
1 1 2 3 4
1 2 0 3 2

0 1 2 0 6 0 1 2 0 6 = B,
A=
0
0 0 0 0 0
0 0 0 1 2
0 0 0 0
0
0 0 1 2
0 0 0 1 2
0 0 0 0 0
by the row operations New R2 = (Old R2 ) + R1 , followed by Swap(R3 , R4 ). Now B has 3
nonzero rows.
B are linearly independent; therefore, a basis for the range
Columns
1,
3,4 of
1
2
3
1 0 3

space of A is
0 , 0, 0 . Other choices are possible: columns 1, 2, 4, or columns
0
0
1
1, 2, 5, or columns 1, 3, 5 would also work (since they are obviously linearly independent.
N OTE : columns 2, 4, 5 or columns 3, 4, 5 are also linearly independent, so would also work;
but this is not obvious.
R EMARK : Finding linearly independent columns of a reduced row echelon matrix is very
easy; just look at the lowest nonzero entries in each column. Its similar to showing that upper
triangular matrices with nonzero diagonal entries are always invertible by direct solution of linear equations. Try lots of examples and it becomes obvious! Its really a kind of algorithmic
understanding.
Theorem 34 Every basis for the null space of B is also a basis for the null space of A.
Proof: B = M A, so Au = 0 = Bu = M Au = M (0) = 0. Therefore (null space of A)
(null space of B). But A = M 1 B, so similarly (null space of B) (null space of A). Thus
A, B have the same null space. So any basis for one null space will also be a basis for the other.
This is useful because it is always easy to find a basis for the null space of a matrix in row
echelon form.
Example:
of A from Example 33 above. We know the nullity is 5 3 = 2.
find
the null
space

1
0
0
1


, v = (since it is obvious that these are linearly independent); now choose

Let u =

the other numbers so that u, v lie in the null space of B (the reduced form of A):

1
1
+
2x
+
3y
+
4z
0

2x + 6z

,
0 = Bu = B
x =

y + 2z
y
0
z
so that 2x + 3y + 4z + 1 = 0, 2x + 6z = 0, y + 2z = 0. This is easy to solve: we get y = 2z,
x = 3z, (6 6 + 4)z = 1, so that z = 1/8, x = 3/8, y = 1/4.


1
0
0
0 1
1


1
1
1

, 1/4
3/8
Similarly, if 0 = B p, then p = 4 , q = 2 , r = 4 . Hence, the pair

1/4 1/2
q
1/8
1/4
r

8
0
0 4


is a basis for the null space of A. We can clear the fractions to get
3, 1 .
2 2
1
1

E IGENVECTORS , EIGENVALUES ,

DIAGONALISATION

Definition 35 Let A be an n n matrix, so that A gives a linear transformation A : Rn Rn .


If Au = u for some 0 6= u Rn and R, then u is an eigenvector of A with eigenvalue .
I MPORTANT:

WE MUST HAVE

u 6= 0.

Exercise 36 If 6= 0, then u is an eigenvector of A u is an eigenvector of A.


Theorem 37 is an eigenvalue of A I A is singular det(I A) = 0.
Proof: If is an eigenvalue of A with u 6= 0 an eigenvector, then (I A)u = u u = 0.
Therefore, I A is not invertible (because if it were, we would have u = (I A)1 (0) = 0,
a contradiction because u 6= 0.) Thus also det(I A) = 0.
Conversely, suppose that det(I A) = 0, so that I A is singular. Then, there must
be some u 6= 0 such that (I A)u = 0, by Theorem 27. Then u is an eigenvector with
eigenvalue .
So, YOU CAN FIND EIGENVALUES BY SOLVING det(I A) = 0. This is a polynomial
equation in , of degree n.
[*** Notice that if we use the complex numbers C, then eigenvalues always exist, because
polynomials in C always have solutions in C. But this is NOT ON THE SYLLABUS. ***]
Fact 38 Let 1 , 2 , . . . be DISTINCT eigenvalues of A. Then the corresponding eigenvectors
u1 , u2 , . . . are linearly independent.
P
Proof: (***) Assume that j j uj = 0 for some j . Then p(A)uj = p(j )uj for every
polynomial p, so
!
X
X
X
0 = p(A)0 = p(A)
j uj =
j p(A)uj =
j p(j )uj .
j

10

Now, for each fixed k, there is some polynomial pk such that


(
1
(j = k),
pk (j ) =
0
(j 6= k).
Q
Q
For example, take the Lagrange polynomial pk (z) = j (z j ) j6=k (k j )1 . Now with
p = pk above, we get 0 = k uk , so that k = 0 (because uk 6= 0 by definition of eigenvectors).
Therefore k = 0 for all k, so u1 , u2 , . . . are linearly independent as required.
Fact 39 Let A be an nn matrix with DISTINCT eigenvalues 1 , 2 , . . . , n . Let u1 , u2 , . . . , un
be eigenvectors (with j the eigenvalue for uj ). Then A = QDQ1 , where D is a diagonal
matrix with entries 1 , 2 , . . . , n , and Q has column vectors u1 , u2 , . . . , un .
*** [If some of the eigenvalues of A are repeated, so we do not have n distinct eigenvalues,
then it doesnt work; but this is NOT ON THE SYLLABUS !! Sometimes diagonalisation is not
possible; instead we have the Jordan Canonical Form.] ***
Proof: (***) By definition Q has column vectors {uj }, so uj = Qej . Thus j uj = Auj =
AQej . But also Dej is the j th column of D, so Dej = j ej . Thus
QDej = Q(j ej ) = j Qej = j uj = Auj = AQej .
Since this is true for every j (and {e1 , . . . , en } is a basis of Rn ), QD = AQ.
Now the columns of Q are u1 , u2 , . . . , un , which are linearly independent (by Fact 38), and
hence a basis of Rn . Therefore Q is invertible, by two methods:
E ITHER: Fact 10 (RUBBISH !!), OR: the Range Space of Q is the column space of Q (Theorem 25), which has dimension n; so the rank of Q is n. Now look at the proof of Theorem 27.
But QD = AQ. Therefore, A = QDQ1 .
In the exam, you will only have to calculate eigenvectors and eigenvalues for 2 2 and 3 3
matrices, and the eigenvalues will always be distinct and in R (not C).
Main use of diagonalisation: if A = QDQ1 then An = QDn Q1 , and Dn is easy to find
because D is DIAGONAL.




2 1
1 0
1
Example 40 Let A = QDQ , where Q =
and D =
. Find An .
3 0
0 2




0 1
0 1
1
=3
. Therefore,
We have Q =
3 2
3 2






 

0 1
1 0
2 1
0 2n
2 1
2n
0
n
1
1
A =3
=3
=
3 2
0 2n
3 0
3 2n+1
3 0
2 2n+1 1

 

1
0
1 0
0
U SEFUL CHECK : when n = 0, we get A =
=
. Hence A0 = I.
22 1
0 1


1/2 0
1
Note also that the formula is true for n < 0, e.g. A =
.
1 1
1

1
20(3)1

11

Reference= Linear Algebra By Gilbert Strang

Potrebbero piacerti anche