Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Lecture Notes
Lecture 1 Introduction
Textbook
Students are strongly advised to acquire a copy of the Textbook:
D. C. Lay. Linear Algebra and its Applications. Pearson, 2006. ISBN 0-521-28713-4.
Other editions can be used as well; the book is easily available on Amazon (this includes
some very cheap used copies) and in Blackwell’s bookshop on campus.
These lecture notes should be treated only as indication of the content of the
course; the textbook contains detailed explanations, worked out examples,
etc.
About Homework
1
The Undergraduate Student Handbook says:
As a general rule for each hour in class you should spend two hours on independent study.
This will include reading lecture notes and textbooks, working on exercises and preparing
for coursework and examinations.
In respect of this course, MATH10212 Linear Algebra B, this means that students are
expected to spend 8 (eight!) hours a week in private study of Linear Algebra.
Normally, homework assignments will consist of some odd numbered exercises from the
sections of the Textbook covered in the lectures up to Wednesday in particular week.
The Textbook contains answers to most odd numbered exercises (but the numbering of
exercises might change from one edition to another).
Homework N (already posted on the webpage and BB9 space of the course) should be
returned to Supervision Classes teachers early in Week N, for marking and discussion at
supervision classes on Tuesday and Wednesday.
Communication
The Course Webpage is
http://www.maths.manchester.ac.uk/~avb/math10212-Linear-Algebra-B.html
The Course Webpage page is updated almost daily, sometimes several times a day. Re-
fresh it (and files it is linked to) in your browser, otherwise you may miss the changes.
Email: Feel free to write to me with questions, etc., at the address
1
http://www.maths.manchester.ac.uk/study/undergraduate/information-for-current-students/
undergraduatestudenthandbook/teachingandlearning/timetabledclasses/.
MATH10212 • Linear Algebra B • Lecture 1 • Linear systems • Last change: 14 March 2019 3
alexandre.borovik@manchester.ac.uk
but only from your university e-mail account.
Emails from Gmail, Hotmail, etc. automatically go to spam.
• Over the course, we shall develop increasingly compact notation for operations of
Linear Algebra. In particular, we shall discover that
p1 g1 + p2 g2 + p3 g3
and
T p1
p1 p 2 p3 = p2 .
p3
MATH10212 • Linear Algebra B • Lecture 1 • Linear systems • Last change: 14 March 2019 4
write
p1 g 1 + p2 g 2 + p3 g 3 = pi g i ,
omitting the summation sign entirely. This particular trick was invented by Albert
Einstein, of all people. I do not use “physics” tricks in my lectures, but am
prepared to give a few additional lectures to physics students.
• Unlike, say, Calculus, Linear Algebra focuses more on the development of a special
mathematics language rather than on procedures.
x1 + x 2 = 3
x1 − x2 = 1
Replacement Replace one equation by the sum of itself and a multiple of another
equation.
We shall prove later in the course that a system of linear equations has either
• no solution, or
under the assumption that the coefficients and solutions of the systems are real or complex
numbers.
A system of linear equations is said to be consistent it if has solutions (either one or
infinitely many), and a system in inconsistent if it has no solution.
to replace one system with an equivalent system (that is, with the same
solution set) which is easier to solve.
MATH10212 • Linear Algebra B • Lecture 2 • Linear systems • Last change: 14 March 2019 6
Checking solutions
Given a solution of the system of linear equations, how to check that it is correct?
MATH10212 • Linear Algebra B • Lecture 2 • Linear systems • Last change: 14 March 2019 7
Matrix notation
x1 − 2x2 + 3x3 = 1
x1 + x2 = 2
x2 + x3 = 3
Replacement Replace one row by the sum of itself and a multiple of another row.
The two matrices are row equivalent if there is a sequence of elementary row operations
that transforms one matrix into the other.
Note:
If the augmented matrices of two linear systems are row equivalent, then the two systems
have the same solution set.
A nonzero row or column of a matrix is a row or column which contains at least one
nonzero entry.
MATH10212 • Linear Algebra B • Lecture 2 • Linear systems • Last change: 14 March 2019 8
Two linear systems are equivalent if and only if the augmented matrix of one of them
can be obtained from the augmented matrix of another system by frow operations and
insertion / deletion of zero rows.
MATH10212 • Linear Algebra B • Lecture 3 • Solution of Linear Systems • Last change: 14 March 2019 9
Definition. A matrix is in echelon form (or row echelon form) if it has the following
three properties:
2. Each leading entry of a row is in column to the right of the leading entry of the
row above it.
Each matrix is row equivalent to one and only one reduced echelon form.
MATH10212 • Linear Algebra B • Lecture 3 • Solution of Linear Systems • Last change: 14 March 2019 10
A pivot is a nonzero number in a pivot position which is used to create zeroes in the
column below it.
A rule for row reduction:
1. Pick the leftmost non-zero column and in it the topmost nonzero entry; it is a
pivot.
2. Using scaling, make the pivot equal 1.
3. Using replacement row operations, kill all non-zero entries in the column below
the pivot.
4. Mark the row and column containing the pivot as pivoted.
5. Repeat the same with the matrix made of not pivoted yet rows and columns.
6. When this is over, interchange the rows making sure that the resulting matrix is
in echelon form.
7. Using replacement row operations, kill all non-zero entries in the column above
the pivot entries.
The variables x1 and x2 correspond to pivot columns in the matrix and a re called basic
variables (also leading or pivot variables). The other variable, x3 is a free variable.
Free variables can be assigned arbitrary values and then leading variables expressed in
terms of free variables:
x1 = 1 + 5x3
x2 = 4 − x3
x3 is free
A linear system is consistent if and only if the rightmost column of the augmented matrix
is not a pivot column—that is, if and only if an echelon form of the augmented matrix
has no row of the form
0 · · · 0 b with b nonzero
If a linear system is consistent, then the solution set contains either
(ii) infinitely many solutions, when there is at least one free variable.
5. Express each basic variable in terms of any free variables appearing in the equation.
MATH10212 • Linear Algebra B • Lecture 4 • Vector equations • Last change: 14 March 2019 12
The scalar multiple cv of a vector v and a real number (“scalar”) c is the vector
obtained by multiplying each entry in v by c. For example in R3 ,
1 1.5
1.5 0 =
0 .
−3 −2
The vector whose entries are all zeroes is called the zero vector and denoted 0:
0
0
0 = .. .
.
0
Algebraic properties of Rn
1. u + v = v + u
2. (u + v) + w = u + (v + w)
3. u + 0 = 0 + u = u
4. u + (−u) = −u + u = 0
MATH10212 • Linear Algebra B • Lecture 4 • Vector equations • Last change: 14 March 2019 13
5. c(u + v) = cu + cv
6. (c + d)u = cu + du
7. c(du) = (cd)u
8. 1u = u
Linear combinations
y = c1 v1 + · · · cp vp
x2 + x3 = 2
x1 + x2 + x3 = 3
x1 + x2 − x3 = 2
Denote
0 1 1
a1 = 1 , a2 = 1 , a3 =
1
1 1 −1
MATH10212 • Linear Algebra B • Lecture 4 • Vector equations • Last change: 14 March 2019 14
and
2
b = 3 ,
2
then the vector equation can be written as
x1 a1 + x2 a2 + x3 a3 = b.
Therefore
solving a linear system is the same as finding an expression of the vector of the right part
of the system as a linear combination of columns in its matrix of coefficients.
A vector equation
x1 a1 + x2 a2 + · · · + xn an = b.
has the same solution set as the linear system whose augmented matrix is
a1 a2 · · · an b
That is, Span{v1 , . . . , vp } is the collection of all vectors which can be written in the form
c1 v1 + c2 v2 + · · · + cp vp
with c1 , . . . , cp scalars.
We say that vectors v1 , . . . , vp span Rn if
Span{v1 , . . . , vp } = Rn
= x1 a1 + x2 a2 + · · · + xn an
x2 + x3 = 2
x1 + x2 + x3 = 3
x1 + x2 − x3 = 2
was written as
x1 a1 + x2 a2 + x3 a3 = b.
where
0 1 1
a1 = 1 , a2 = 1 , a3 = 1
1 1 −1
and
2
b = 3 .
2
In the matrix product notation it becomes
0 1 1 x1 2
1 1 1 x2 = 3
1 1 −1 x3 2
or
Ax = b
where
x1
A = a1 a2 a3 , x = x2 .
x3
x1 a1 + x2 a2 + · · · + xn an = b
MATH10212 • Linear Algebra B • Lecture 5 • The matrix equation Ax = b • Last change: 14 March 2019 17
which has the same solution set as the system of linear equations whose augmented
matrix is
a1 a2 · · · an b .
Row-vector rule for computing Ax. If the product Ax is defined then the ith entry
in Ax is the sum of products of corresponding entries from the row i of A and from the
vector x.
Theorem 1.4.5: Properties of the matrix-vector product Ax.
Ax = 0.
x1 + 2x2 − x3 = 0
x1 + 3x3 + x3 = 0
MATH10212 • Linear Algebra B • Lecture 6 • Solution sets of linear equations • Last change: 14 March 2019 18
x1 + 2x2 − x3 = 0
x1 + 3x3 + x3 = 5
Ax = b
is consistent for some given b, and p be a solution. Then the solution set of Ax = b is
the set of all vectors of the form
w = p + vh ,
where vh is any solution of the homogeneous equation
Ax = 0.
MATH10212 • Linear Algebra B • Lecture 7 • Linear independence • Last change: 14 March 2019 19
{ v1 , . . . , vp }
x1 v1 + · · · + xp vp = 0
The set
{ v1 , . . . , vp }
in Rn is linearly dependent if there exist weights c1 , . . . , cp , not all zero, such that
c1 v1 + · · · + cp vp = 0
Ax = 0
Therefore the columns of matrix A are linearly independent if and only if the equation
Ax = 0
S = { v1 , . . . , vp }
of two or more vectors is linearly dependent if and only if at least one of the vectors in
S is a linear combination of the others.
Theorem 1.7.8: dependence of “big” sets. If a set contains more vectors than
entries in each vector, then the set is linearly dependent. Thus, any set
{ v1 , . . . , vp }
T : Rn −→ Rm
x 7→ Ax.
In short:
T (x) = Ax.
The range of a matrix transformation. The range of T is the set of all linear
combinations of the columns of A.
Indeed, this can be immediately seen from the fact that each image T (x) has the form
T (x) = Ax = x1 a1 + · · · + xn an
T : Rn −→ Rm
is linear if:
T (0) = 0
and
T (cu + dv) = cT (u) + dT (v).
The identity matrix. An n × n matrix with 1’s on the diagonal and 0’s elsewhere is
called the identity matrix In . For example,
1 0
I1 = 1 , I2 = ,
0 1
1 0 0 0
1 0 0 0 1 0 0
I3 = 0 1 0 , I4 =
0
.
0 1 0
0 0 1
0 0 0 1
MATH10212 • Linear Algebra B • Lecture 8 • Linear transformations • Last change: 14 March 2019 21
e1 , e2 , . . . , en .
For example, in R3
1 0 0
e1 = 0 ,
e2 = 1 ,
e3 = 0 .
0 0 1
Obsetrve that if x is an arbitrary vector in Rn ,
x1
..
x = . ,
xn
then
x1
x2
x = x3
..
.
xn
1 0 0
0 1 0
= x1 0 + x2 0 + · · · + xn 0
.. .. ..
. . .
0 0 1
= x1 e1 + x2 e2 + · · · + xn en .
In x = x for all x ∈ Rn .
Therefore the linear transformation associated with the identity matrix is the identity
transformation of Rn :
Rn −→ Rn
x 7→ x
T : Rn −→ Rm
be a linear transformation.
Then there exists a unique matrix A such that
In fact, A is the m × n matrix whose jth column is the vector T (ej ) where ej is the jth
column of the identity matrix in Rn :
A = T (e1 ) · · · T (en )
x = x 1 e 1 + x2 e 2 + · · · + xn e n
T (x) = T (x1 e1 + x2 e2 + · · · + xn en )
= T (x1 e1 ) + · · · + T (xn en )
= x1 T (e1 ) + · · · + xn T (en )
and then switch to matrix
notation:
x
.1
= T (e1 ) · · · T (en ) ..
xn
= Ax.
The matrix A is called the standard matrix for the linear transformation T .
Definition. A transformation
T : Rn −→ Rm
is onto Rm if each b ∈ Rm is the image of at least one x ∈ Rn .
A transformation T is one-to-one if each b ∈ Rm is the image of at most one x ∈ Rn .
Theorem: One-to-one transformations: a criterion. A linear transformation
T : Rn −→ Rm
T (x) = 0
T : Rn −→ Rm
a11 · · · ···
a1j a1n
.. .. ..
. . .
= ai1 · · · aij ··· ain
. .. ..
.. . .
am1 · · · amj · · · amn
Notice that in aij the first subscript i denotes the row number, the second subscript j
the column number of the entry aij . In particular, the column aj is
a1j
..
.
aj = aij .
.
..
amj
are diagonal!
Sums. If
A = a1 a2 · · · an
and
B = b1 b2 · · · bn
are m × n matrices then we define the sum A + B as
A + B = a1 + b1 a2 + b2 · · · an + bn
a11 + b11 · · · a1j + b1j · · · a1n + b1n
.. .. ..
. . .
= ai1 + bi1 · · · aij + bij · · · ain + bin
.. .. ..
. . .
am1 + bm1 · · · amj + bmj · · · amn + bmn
ca11 · · · ···
ca1j ca1n
.. .. ..
. . .
= cai1 · · · caij ··· cain
. .. ..
.. . .
cam1 · · · camj · · · camn
1. A + B = B + A
2. (A + B) + C = A + (B + C)
3. A + 0 = A
4. r(A + B) = rA + rB
5. (r + s)A = rA + sA
6. r(sA) = (rs)A.
MATH10212 • Linear Algebra B • Lecture 10 • Matrix multiplication • Last change: 14 March 2019 25
T : Rn −→ Rm , x 7→ Bx
and
S : Rm −→ Rp , y 7→ Ay.
Their composition
(S ◦ T )(x) = S(T (x))
is a linear transformation
S ◦ T : Rn −→ Rp .
From the previous lecture, we know that linear transformations are given by matrices.
What is the matrix of S ◦ T ?
Multiplication of matrices. To answer the above question, we need to compute A(Bx)
in matrix form.
Write x as
x1
x = ...
xn
and observe
Bx = x1 b1 + · · · + xn bn .
Hence
A(Bx) = A(x1 b1 + · · · + xn bn )
= A(x1 b1 ) + · · · + A(xn bn )
= x1 A(b1 ) + · · · + xn A(bn )
= Ab1 Ab2 · · · Abn x
transforms x into A(Bx). Hence C is the matrix of the linear transformation S ◦ T and
it will be natural to call C the product of A and B and denote
C =A·B
Ab1 , . . . , Abn :
MATH10212 • Linear Algebra B • Lecture 10 • Matrix multiplication • Last change: 14 March 2019 26
AB = A b1 · · · bn
= Ab1 · · · Abn .
Mnemonic rules
1. A(BC) = (AB)C
2. A(B + C) = AB + AC
3. (B + C)A = BA + CA
4. r(AB) = (rA)B = A(rB) for any scalar r
5. Im A = A = AIn
Theorem 2.1.3: Properties of transpose. Let A and B denote matrices whose sizes
are appropriate for the following sums and products. Then we have:
MATH10212 • Linear Algebra B • Lecture 10 • Matrix multiplication • Last change: 14 March 2019 27
1. (AT )T = A
2. (A + B)T = AT + B T
4. (AB)T = B T AT
MATH10212 • Linear Algebra B • Lecture 11 • The inverse of a matrix • Last change: 14 March 2019 28
CA = I and AC = I
det A = ad − bc
x = A−1 b.
(A−1 )−1 = A
(AB)−1 = B −1 A−1
• Form the augmented matrix A I and row reduce it.
The Invertible Matrix Theorem 2.3.8: For an n × n matrix A, the following are
equivalent:
(a) A is invertible.
(l) AT is invertible.
MATH10212 • Linear Algebra B • Lecture 12 • Characterizations of invertible matrices • Last change: 14 March 201930
B = A−1 and A = B −1 .
T : Rn −→ Rn
S(x) = A−1 x
Addition of partitioned matrices. If matrices A and B are of the same size and
partitioned the same way, they can be added block-by-block.
Similarly, partitioned matrices can be multiplied by a scalar blockwise.
Multiplication of partitioned matrices. If the column partition of A matches the
row partition of B then AB can be computed by the usual row-column rule, with blocks
treated as matrix entries.
1 2 a α β
A = 3 4 b , B = γ δ
p q z Γ ∆
Example
1 2 a
A11 A12
A= 3 4 b =
,
A21 A22
p q z
α β
B1
B = γ δ =
B2
Γ ∆
A11 A12
B1
AB =
A21 A22
B2
A11 B1 + A12 B2
=
A21 B1 + A22 B2
1 2 α β a
3 4 γ + Γ ∆
δ b
=
α β
p q + z Γ ∆
γ δ
MATH10212 • Linear Algebra B • Lecture 13 • Partitioned matrices • Last change: 14 March 2019 32
AY1 + CY2 = B.
Please notice that matrix A is invertible – at this stage it should be an easy mental
calculation for you. Multiplying the both parts of the last equation by A−1 , we get
and
Y1 = A−1 B − A−1 CY2
MATH10212 • Linear Algebra B • Lecture 13 • Partitioned matrices • Last change: 14 March 2019 33
The entries of
x
Y2 = 3
x4
can be taken for free parameters: x3 = t, x4 = s, and x1 and x2 expressed as
x1 −1 −1 t
=A B−A C
x2 s
Observe that
−1 3 −2
A =
−1 1
and
−1 3 −2 3 4
A C = ·
−1 1 3 4
3 4
=
0 0
0 0 1
Of course, when we solving a singles system of equations, the formulae of the kind
do not save time and effort, but in many applications you have to solve millions of systems
of linear equations with the same matrix of coefficients but different right-hand parts,
and in this situation shortcuts based on compound matrices become very useful.
MATH10212 • Linear Algebra B • Lecture 14 • Subspaces • Last change: 14 March 2019 34
(a) 0 ∈ H.
Rn is a subspace of itself.
{ 0 } is a subspace, called the zero subspace.
Ax = 0
Theorem 2.8.12: The null space. The null space of an m×n matrix A is a subspace
of Rn .
Equivalently, the set of all solutions to a system
Ax = 0
For each x ∈ H, the coordinates of x relative to the basis B are the weights
c1 , . . . , c p
such that
x = c1 b1 + · · · + cp bp .
The vector
c1
..
x B=.
cp
The determinant of a 1 × 1 matrix A = a11 is defined simply as being equal its
only entry a11 :
det a11 = a11 .
equals
1 0 0 1
2 · det + 7 · det
0 4 1 0
which further simplifies as
2 · 4 + 7 · (−1) = 8 − 7 = 1.
Submatrices. By definition, the submatrix Aij is obtained from the matrix A by crossing
out row i and column j.
For example, if
1 2 3
A = 4 5 6 ,
7 8 9
MATH10212 • Linear Algebra B • Lecture 15 • Introduction to determinants • Last change: 14 March 2019 37
then
1 3 2 3
A22 = , A31 =
7 9 5 6
det A = a11 det A11 − a12 det A12 + · · · + (−1)1+n a1n det A1n
n
X
= (−1)1+j a1j det A1j
j=1
Then
det A = a11 C11 + a12 C12 + · · · + a1n C1n
is cofactor expansion across the first row of A.
Theorem 3.1.1: Cofactor expansion. For any row i, the result of the cofactor
expansion across the row i is the same:
(−1)i+j
which appear in the formula for cofactors, form the easy-to-recognise and easy-to-remember
“chess board” pattern:
+ − + ···
− + −
+ − +
.. ...
.
Proof: This proof contains more details than the one given in the textbook. It is based
on induction on n, which, to avoid the use of “general” notation is illustrated by a simple
example. Let
a11 0 0 0 0
a21 a22 0 0 0
A= a 31 a 32 a 33 0 0 .
a41 a42 a43 a44 0
a51 a52 a53 a54 a55
All entries in the first row of A–with possible exception of a11 —are zeroes,
a12 = · · · = a15 = 0,
But the smaller matrix A11 is also lower triangle, and therefore we can conclude by
induction that
a22 0 0 0
a32 a33 0 0
a42 a43 a44 0 = a22 · · · a55
det
Corollary. The determinant of a diagonal matrix equals is the product of its diagonal
elements:
d1 0 · · · 0
0 d2 0
det .. .. = d1 d2 · · · dn .
..
. . .
0 · · · dn
det In = 1.
det 0 = 0.
MATH10212 • Linear Algebra B • Lectures 15–16 • Properties of determinants • Last change: 14 March 2019 40
(a) If a multiple of one row of A is added to another row to produce a matrix B, then
det B = det A.
det B = − det A.
det B = k det A.
Example. Let
1 2 0
A = 0 1 0 ,
0 3 4
then
1 0 0
AT = 2 1 3
0 0 4
and
1 0
det A = 1 · det
3 4
0 0
−2 · det
0 4
0 1
+0 · det
0 3
= 1·4−2·0+0·0
= 4.
Similarly
T 1 3
det A = 1 · det +0+0
0 4
= 1 · 4 = 4;
MATH10212 • Linear Algebra B • Lectures 15–16 • Properties of determinants • Last change: 14 March 2019 41
by expansion across the first row. Now we do it by expansion across the third column:
1+3 0 1
det A = 7 · (−1) · det
1 0
2+3 2 0
+0 · (−1) · det
1 0
3+3 2 0
+4 · (−1) · det
0 1
= −7 − 0 + 8
= 1.
det B = − det A.
det B = k det A.
Proof: Column operations on A are row operation of AT which has the same determinant
as A.
Computing determinants. In computations by hand, the quickest method of comput-
ing determinants is to work with both columns and rows:
• If a row or a column has a convenient scalar factor, take it out of the determinant.
MATH10212 • Linear Algebra B • Lectures 15–16 • Properties of determinants • Last change: 14 March 2019 42
• If convenient, swap two rows or two columns— but do not forget to change the sign
of the determinant.
• Adding to a row (column) scalar multiples of other rows (columns), simplify the
matrix to create a row (column) with just one nonzero entry.
Example.
1 0 3 1 0 0
C3 −3C1
det 1 3 1 = det 1 3 −2
1 1 −1 1 1 −4
1 0 0
2 out of C3
= 2 · det 1 3 −1
1 1 −2
1 0 0
−1 out of C3
= −2 · det 1 3 1
1 1 2
expand 1+1 3 1
= −2 · 1 · (−1) · det
1 2
R1 −3R2 0 −5
= −2 · det
1 2
R1 ↔R2 1 2
= −2 · (−1) · det
0 −5
= 2 · (−5)
= −10.
As an exercise, I leave you with this question: wouldn’t you agree that the determinant
of the similarly constructed matrix of size n × n should be ±(n − 1)? But how can one
determine the correct sign?
MATH10212 • Linear Algebra B • Lectures 15–16 • Properties of determinants • Last change: 14 March 2019 44
And one more exercise: you should now be able to instantly compute
1 2 3 4 0
1 2 3 0 5
1
det 2 0 4 5 .
1 0 3 4 5
0 2 3 4 5
det A 6= 0.
Recall that this theorem has been already known to us in the case of 2 × 2 matrices A.
Example. The matrix
2 0 7
A = 0 1 0
1 0 4
has the determinant
1 0 0 1
det A = 2 · det + 7 · det
0 4 1 0
= 2 · 4 + 7 · (−1)
= 8−7=1
Example. Take
1 0 1 5
A= and B = ,
3 2 0 1
then
det A = 2 and det B = 1.
But
1 51 0
AB =
0 13 2
1 5
=
3 17
MATH10212 • Linear Algebra B • Lectures 15–16 • Properties of determinants • Last change: 14 March 2019 45
and
det(AB) = 1 · 17 − 5 · 3 = 2,
which is exactly the value of
(det A) · (det B).
Corollary. If A is invertible,
For any n × n matrix A and any b ∈ Rn , denote by Ai (b) the matrix obtained from A
by replacing column i by the vector b:
Ai (b) = a1 · · · ai−1 b ai+1 · · · an .
Theorem 3.3.7: Cramer’s Rule. Let A be an invertible n×n matrix. For any b ∈ Rn ,
the unique solution x of the linear system
Ax = b
is given by
det Ai (b)
xi = , i = 1, 2, . . . , n.
det A
x1 + x2 = 3
x1 − x2 = 1
MATH10212 • Linear Algebra B • Lectures 15–16 • Properties of determinants • Last change: 14 March 2019 46
and Aji is the matrix obtained from A by deleting row j and column i.
Notice that for a 2 × 2 matrix
a b
A=
c d
the cofactors are
C11 = (−1)1+1 · d = d
C21 = (−1)2+1 · b = −b
C12 = (−1)1+2 · c = −c
C22 = (−1)2+2 · a = a,
Quiz
Is Γ positive, negative, or zero?
0 0 71
Γ = det 0 93 0
87 0 0
Ax = λx
(A − λI)x = 0
(A − λI)x = 0
{ v1 , . . . , vp }
is linearly independent.
MATH10212 • Linear Algebra B • Lecture 17 • The characteristic equation • Last change: 14 March 2019 48
det(A − λI) = 0
is the characteristic equation for A.
Characterisation of eigenvalues
Theorem. A scalar λ is an eigenvalue of an n × n matrix A if and only if λ satisfies the
characteristic equation
det(A − λI) = 0
A = P BP −1 .
Example:
Occasionally the similarity relation is denoted by symbol ∼.
Similarity is an equivalence relation: it is
• reflexive: A ∼ A
• symmetric: A ∼ B implies B ∼ A
• transitive: A ∼ B and B ∼ C implies A ∼ C.
Conjugation. Operation
AP = P −1 AP
is called conjugation of A by P .
Properties of conjugation
• IP = I
• (AB)P = AP B P ; as a corollary:
• (A + B)P = AP + B P
• (c · A)P = c · AP
• (Ak )P = (AP )k
• (A−1 )P = (AP )−1
• (AP )Q = AP Q
A = P DP −1
MATH10212 • Linear Algebra B • Lectures 18–21 • Diagonalisation 50
2. u + v = v + u.
3. (u + v) + w = u + (v + w).
7. c(u + v) = cu + cv.
8. (c + d)u = cu + du.
9. c(du) = (cd)u.
10. 1u = u.
Pn = {a0 + a1 x + · · · + an xn }
Definitions
Most concepts introduced in the previous lectures for the vector spaces Rn can be trans-
ferred to arbitrary vector spaces. Let V be a vector space.
Given vectors v1 , v2 , . . . , vp in V and scalars c1 , c2 , . . . , cp , the vector
y = c1 v1 + · · · cp vp
That is, Span{v1 , . . . , vp } is the collection of all vectors which can be written in the form
c1 v1 + c2 v2 + · · · + cp vp
with c1 , . . . , cp scalars.
Span{v1 , . . . , vp } is a vector subspace of V , that is, it is closed with respect to addition
of vectors and multiplying vectors by scalars.
V is a subspace of itself.
{ 0 } is a subspace, called the zero subspace.
x1 v1 + · · · + xp vp = 0
The set
{ v1 , . . . , vp }
in V is linearly dependent if there exist weights c1 , . . . , cp , not all zero, such that
c1 v1 + · · · + cp vp = 0
S = { v1 , . . . , vp }
of two or more vectors is linearly dependent if and only if at least one of the vectors in
S is a linear combination of the others.
A basis of V is a linearly independent set which spans V . A vector space V is called
finite dimensional if it has a finite basis. Any two bases in a finite dimensional vector
space have the same number of vectors; this number is called the dimension of V and
is denoted dim V .
If B = (b1 , . . . , bn ) is a basis of V , every vector x ∈ V can be written, and in a unique
way, as a linear combination
x = x1 b1 + · · · xn bn ;
the scalars x1 , . . . , xn are called coordinates of x with respect to the basis B.
MATH10212 • Linear Algebra B • Lectures 23 and 24 • Linear transformations • Last change: 14 March 2019 53
T (0) = 0
and
T (cu + dv) = cT (u) + dT (v).
V −→ V
x 7→ x.
V −→ V
x 7→ 0.
If
T : V −→ W
is a linear transformation, its kernel
Ker T = {x ∈ V : T (x) = 0}
is a subspace of W .
MATH10212 • Linear Algebra B • Lecture 25 • Symmetric matrices and inner product • Last change: 14 March 201954
An example:
1 2 3
A = 2 3 4 .
3 4 5
u v = uT v
v1
v2
= u1 u2 · · · un ..
.
vn
= u1 v1 + u2 v2 + · · · + un vn
(a) u v = v u
(b) (u + v) w = u w + v w
In particular,
kvk2 = v v.
MATH10212 • Linear Algebra B • Lecture 25 • Symmetric matrices and inner product • Last change: 14 March 201955
uv = 0
That is, vectors u and v are orthogonal if and only if the Pythagoras Theorem holds for
the triangle formed by u and v as two sides (so that its third side is u + v).
ku + vk2 = (u + v) (u + v)
= uu + uv + vu + vv
= uu + uv + uv + vv
= u u + 2u v + v v
= kuk2 + 2u v + kvk2 .
Hence
ku + vk2 = kuk2 + kvk2 + 2u v,
and
ku + vk2 = kuk2 + kvk2 iff u v = 0.
{ u1 , . . . , up }
in Rn is orthogonal if
ui uj = 0 whenever i 6= j.
Theorem 6.2.4. If
S = { u1 , . . . , up }
is an orthogonal set of non-zero vectors in Rn then S is linearly independent.
Proof. Assume the contrary that S is linearly dependent. This means that there exist
scalars c1 , . . . , cp , not all of them zeroes, such that
c1 u1 + · · · + cp up = 0.
Forming the dot product of the left hand side and the right hand side of this equation
with ui , we get
(c1 u1 + · · · + cp up ) ui = 0 ui = 0
MATH10212 • Linear Algebra B • Lecture 25 • Symmetric matrices and inner product • Last change: 14 March 201956
In view of orthogonality of the set S, all dot products on the left hand side, with the
exception of ui u, equal 0. Hence what remains from the equality is
ci ui ui = 0
Since ui is non-zero,
ui ui 6= 0
and therefore ci = 0. This argument works for every index i = 1, . . . , p. We get a con-
tradition with our assumption that one of ci is non-zero.
v1 , . . . , vp
λ1 , . . . , λ p .
Then
{ v1 , . . . , vp }
is an orthogonal set.
Proof. We need to prove the following:
u v = 0.
λu v = (λu) v
= (Au) v
= (Au)T v
= (uT AT )v
= (uT A)v
= uT (Av)
= uT (µv)
= µuT v
= µu v.
MATH10212 • Linear Algebra B • Lecture 25 • Symmetric matrices and inner product • Last change: 14 March 201957
Hence
λu v = µu v
and
(λ − µ)u v = 0.
Since λ − µ 6= 0, we have u v = 0.
Proof. Let
v = x1 u1 + · · · + xn un ,
then, for i = 1, 2, . . . , n,
v ui = x1 u1 ui + · · · + xn un ui
= xi ui ui
= xi .
AAT = I.
• AT is also orthogonal.
• det A = ±1.
(a) A is orthogonal .
u v,
1. u v = v u
2. (u + v) w = u w + v w
3. (cu) v = c(u v)
Inner product space. A vector space with an inner product is called an inner product
space.
Example: School Geometry. The ordinary Euclidean plane of school geometry is an
inner product space:
• Dot product:
u v = kukkvk cos θ,
where θ is the angle between vectors u and v.
Example: C[0, 1]. The vector space C[0, 1] of real-valued continuous functions on the
segment [0, 1] becomes an inner product space if we define inner product by the formula
Z 1
f g = f (x)g(x)dx.
0
In this example, the tricky bit is to show that the dot product on C[0, 1] satisfies the
axiom:
then f (x) = 0 for all x ∈ [0, 1]. This requires the use of properties of continuous func-
tions; since we study linear algebra, not analysis, I am leaving it to the readers as an
MATH10212 • Linear Algebra B • Lectures 25 and 26 • Inner product spaces • Last change: 14 March 2019 60
exercise.
Notice that
kuk2 = u u,
and
kuk = 0 ⇔ u = 0.
u v = 0.
ku + vk2 = (u + v) (u + v)
= uu + uv + vu + vv
= uu + uv + uv + vv
= u u + 2u v + v v
= kuk2 + 2u v + kvk2 .
Hence
ku + vk2 = kuk2 + kvk2 + 2u v,
and
ku + vk2 = kuk2 + kvk2
if and only if
u v = 0.
u v = kukkvk cos θ
hence
|u v| 6 kukkvk.
The following Theorem shows that this is is a general property of inner product spaces.
MATH10212 • Linear Algebra B • Lectures 25 and 26 • Inner product spaces • Last change: 14 March 2019 61
|u v| 6 kukkvk.
The Cauchy-Schwarz Inequality: an example. In the vector space Rn with the dot
product of
u1 v1
.. ..
u = . and v = .
un vn
defined as
u v = uT v = u1 v1 + · · · + un vn ,
the Cauchy-Schwarz Inequality becomes
q q
|u1 v1 + · · · + un vn | 6 u1 + · · · + un · v12 + · · · + vn2 .
2 2
The Cauchy-Schwarz Inequality: one more example. In the inner product space
C[0, 1] the Cauchy-Schwarz Inequality takes the form
Z 1
sZ 1
s
Z 1
f (x)g(x)dx 6 f (x)2 dx · g(x)2 dx,
0 0 0
or, equivalently,
Z 1 2 Z 1 Z 1
2 2
f (x)g(x)dx 6 f (x) dx · g(x) dx .
0 0 0
Proof of the Cauchy-Schwarz Inequality given here is different from the one in the
textbook by Lay. Our proof is based on the following simple fact from school algebra:
Let
q(t) = at2 + bt + c
be a quadratic function in variable t with the property that
q(t) > 0
q(t) > 0.
Hence
(2u v)2 − 4kuk2 kvk2 6 0
which can be simplified as
|u v| 6 kukkvk.
Distance. We define distance between vectors u and v as
d(u, v) = ku − vk.
2. d(u, v) > 0.
3. d(u, v) = 0 iff u = v.
Axioms 1–3 immediately follow from the axioms for dot product, but the Triangle In-
equality requires some attention.
Proof of the Triangle Inequality. It suffices to prove
ku + vk 6 kuk + kvk,
or
ku + vk2 6 (kuk + kvk)2 ,
or
(u + v) (u + v) 6 kuk2 + 2kukkvk + kvk2 ,
or
u u + 2u v + v v 6 u u + 2kukkvk + v v,
or
2u v 6 2kukkvk
But this is the Cauchy-Schwarz Inequality. Now observe that all rearrangements are
reversible, hence the Triangle Inequality holds.
MATH10212 • Linear Algebra B • Lecture 27 • Orthogonalisation and the Gram-Schmidt Process • Last change: 14 March 201963
u = c1 v1 + · · · + cn vn
iff
u vi
ci =
vi vi
u = c1 v1 + · · · + cn vn
iff
c i = u vi
v1 = x1
x2 v1
v2 = x2 − v1
v1 v1
x3 v1 x3 v2
v3 = x3 − v1 − v2
v1 v1 v2 v2
..
.
xp v1 xp v2 xp vp−1
vp = xp − v1 − v2 − · · · − vp−1
v1 v1 v2 v2 vp−1 vp−1
Main Theorem about inner product spaces. Every finite dimensional inner product
space has an orthonormal basis and is isomorphic to on of Rn with the standard dot
product
v
.1
u v = u v = u1 · · · un .. .
T
vn
MATH10212 • Linear Algebra B • Lectures 28–30 • Revision • Last change: 14 March 2019 64
Ax = b,
where b ∈ Rm and
x1
x = ...
xn
is the column vector of unknowns.
Especially important is the homogeneous system of simultaneous linear equations
Ax = 0.
The set
null A = {x ∈ Rn : Ax = 0}
is called the null space of A; it is a vector subspace of Rn and coincides with the
solution space of the homogeneous system Ax = 0.
If a1 , . . . , an are columns of A, so that
A = a1 · · · an ,
Col A = {c1 a1 + · · · + cn an : ci ∈ R}
= Span {a1 , . . . , an }
and equals to the span of the system of vectors a1 , . . . , an , that is, the set of all possible
linear combinations of vectors a1 , . . . , an .
The column space of A has another important characterisation:
TA : Rn −→ Rn
v 7→ Av
MATH10212 • Linear Algebra B • Lectures 28–30 • Revision • Last change: 14 March 2019 65
then the column space of A gets yet another interpretation: it is the image of the linear
transformation TA :
Col A = Im TA
= {v ∈ Rn : ∃v ∈ Rn s.t. TA (w) = v},
null A = ker TA
= {v ∈ R : TA (v) = 0}.
MATH10212 • Linear Algebra B • Lectures 28–30 • Revision • Last change: 14 March 2019 66
Linear dependence
If a1 , . . . , ap are vectors in a vector space V , an expression
c1 a1 + · · · + cp ap , ci ∈ R
c1 a1 + · · · + cp ap = 0
and linear dependencies between vectors a1 , . . . , ap are nothing else but solutions of the
homogeneous system of linear equation
Ax = 0.
Ax = 0
Ri ↔ Rj , i 6= j
Ri ← Ri + λRj , i 6= j
Ri ← λRi , λ 6= 0
Ac = 0 ⇔ (EA)c = 0.
Rank of a matrix
We took it without proof that if U is a vector subspace of Rn , then the maximal systems
of linearly independent subsets in U contain the same number of vectors. This number
is called the dimension of U and is denoted dim U .
A maximal system of linearly independent vectors in U is called a basis of U .
Each basis of U contains dim U vectors.
An important theorem:
dim null A + dim Col A = n
Therefore
dim null A + rank A = n
.
Theorem. If rank A = p then among columns of A there are p linearly independent
columns
Theorem. Since elementary row transformations of A do not change dependency of
columns, e=elementary row operation do not change rank A.