Matrices and Linear Algebra
4/5
()
About this ebook
Linear algebra is one of the central disciplines in mathematics. A student of pure mathematics must know linear algebra if he is to continue with modern algebra or functional analysis. Much of the mathematics now taught to engineers and physicists requires it.
This well-known and highly regarded text makes the subject accessible to undergraduates with little mathematical experience. Written mainly for students in physics, engineering, economics, and other fields outside mathematics, the book gives the theory of matrices and applications to systems of linear equations, as well as many related topics such as determinants, eigenvalues, and differential equations.
Table of Contents:
l. The Algebra of Matrices
2. Linear Equations
3. Vector Spaces
4. Determinants
5. Linear Transformations
6. Eigenvalues and Eigenvectors
7. Inner Product Spaces
8. Applications to Differential Equations
For the second edition, the authors added several exercises in each chapter and a brand new section in Chapter 7. The exercises, which are both true-false and multiple-choice, will enable the student to test his grasp of the definitions and theorems in the chapter. The new section in Chapter 7 illustrates the geometric content of Sylvester's Theorem by means of conic sections and quadric surfaces. 6 line drawings. lndex. Two prefaces. Answer section.
Related to Matrices and Linear Algebra
Titles in the series (100)
Calculus: An Intuitive and Physical Approach (Second Edition) Rating: 4 out of 5 stars4/5A History of Mathematical Notations Rating: 4 out of 5 stars4/5Dynamic Probabilistic Systems, Volume II: Semi-Markov and Decision Processes Rating: 0 out of 5 stars0 ratingsThe Calculus Primer Rating: 0 out of 5 stars0 ratingsInfinite Series Rating: 4 out of 5 stars4/5Calculus Refresher Rating: 3 out of 5 stars3/5History of the Theory of Numbers, Volume II: Diophantine Analysis Rating: 0 out of 5 stars0 ratingsMathematics for the Nonmathematician Rating: 4 out of 5 stars4/5Analytic Inequalities Rating: 5 out of 5 stars5/5Advanced Calculus: Second Edition Rating: 5 out of 5 stars5/5The History of the Calculus and Its Conceptual Development Rating: 4 out of 5 stars4/5Fourier Series and Orthogonal Polynomials Rating: 0 out of 5 stars0 ratingsAn Introduction to Lebesgue Integration and Fourier Series Rating: 0 out of 5 stars0 ratingsElementary Matrix Algebra Rating: 3 out of 5 stars3/5First-Order Partial Differential Equations, Vol. 2 Rating: 0 out of 5 stars0 ratingsA Catalog of Special Plane Curves Rating: 2 out of 5 stars2/5Laplace Transforms and Their Applications to Differential Equations Rating: 5 out of 5 stars5/5Optimization Theory for Large Systems Rating: 5 out of 5 stars5/5Theory of Approximation Rating: 0 out of 5 stars0 ratingsGauge Theory and Variational Principles Rating: 2 out of 5 stars2/5Methods of Applied Mathematics Rating: 3 out of 5 stars3/5First-Order Partial Differential Equations, Vol. 1 Rating: 5 out of 5 stars5/5Differential Forms with Applications to the Physical Sciences Rating: 5 out of 5 stars5/5Numerical Methods Rating: 5 out of 5 stars5/5Euclid and His Modern Rivals Rating: 4 out of 5 stars4/5Chebyshev and Fourier Spectral Methods: Second Revised Edition Rating: 4 out of 5 stars4/5A History of Geometrical Methods Rating: 0 out of 5 stars0 ratingsTopology for Analysis Rating: 4 out of 5 stars4/5An Adventurer's Guide to Number Theory Rating: 4 out of 5 stars4/5Theory of Games and Statistical Decisions Rating: 4 out of 5 stars4/5
Related ebooks
Introduction to Linear Algebra and Differential Equations Rating: 3 out of 5 stars3/5Introduction to Real Analysis Rating: 3 out of 5 stars3/5An Introduction to Ordinary Differential Equations Rating: 4 out of 5 stars4/5A Concept of Limits Rating: 4 out of 5 stars4/5Set Theory: The Structure of Arithmetic Rating: 5 out of 5 stars5/5Boolean Algebra and Its Applications Rating: 4 out of 5 stars4/5Determinants and Matrices Rating: 3 out of 5 stars3/5Linear Algebra Rating: 1 out of 5 stars1/5Introduction to Vector and Tensor Analysis Rating: 4 out of 5 stars4/5Vector Spaces and Matrices Rating: 0 out of 5 stars0 ratingsDifferential Geometry Rating: 5 out of 5 stars5/5An Introduction to Algebraic Structures Rating: 2 out of 5 stars2/5Matrix Theory Rating: 0 out of 5 stars0 ratingsComplex Analysis Rating: 3 out of 5 stars3/5Sets, Sequences and Mappings: The Basic Concepts of Analysis Rating: 0 out of 5 stars0 ratingsDifferential Equations Rating: 4 out of 5 stars4/5Vector Geometry Rating: 0 out of 5 stars0 ratingsIntroduction to Algebraic Geometry Rating: 4 out of 5 stars4/5Matrices and Linear Transformations: Second Edition Rating: 3 out of 5 stars3/5Introduction to the Theory of Abstract Algebras Rating: 0 out of 5 stars0 ratingsIntroduction to Matrices and Vectors Rating: 0 out of 5 stars0 ratingsVector Analysis Rating: 4 out of 5 stars4/5Advanced Calculus: An Introduction to Classical Analysis Rating: 5 out of 5 stars5/5Applied Complex Variables Rating: 5 out of 5 stars5/5Journey into Mathematics: An Introduction to Proofs Rating: 4 out of 5 stars4/5Foundations of Mathematical Analysis Rating: 3 out of 5 stars3/5Cohomology and Differential Forms Rating: 5 out of 5 stars5/5Multivariable Calculus, Linear Algebra, and Differential Equations Rating: 4 out of 5 stars4/5The Linear Algebra Survival Guide: Illustrated with Mathematica Rating: 5 out of 5 stars5/5Vector Calculus Rating: 4 out of 5 stars4/5
Mathematics For You
Basic Math & Pre-Algebra For Dummies Rating: 4 out of 5 stars4/5Must Know High School Algebra, Second Edition Rating: 0 out of 5 stars0 ratingsGeometry For Dummies Rating: 5 out of 5 stars5/5Algebra I Workbook For Dummies Rating: 3 out of 5 stars3/5Algebra - The Very Basics Rating: 5 out of 5 stars5/5Quantum Physics for Beginners Rating: 4 out of 5 stars4/5Mental Math Secrets - How To Be a Human Calculator Rating: 5 out of 5 stars5/5Relativity: The special and the general theory Rating: 5 out of 5 stars5/5Real Estate by the Numbers: A Complete Reference Guide to Deal Analysis Rating: 0 out of 5 stars0 ratingsGame Theory: A Simple Introduction Rating: 4 out of 5 stars4/5The Golden Ratio: The Divine Beauty of Mathematics Rating: 5 out of 5 stars5/5Basic Math & Pre-Algebra Workbook For Dummies with Online Practice Rating: 4 out of 5 stars4/5The Little Book of Mathematical Principles, Theories & Things Rating: 3 out of 5 stars3/5Precalculus: A Self-Teaching Guide Rating: 5 out of 5 stars5/5Is God a Mathematician? Rating: 4 out of 5 stars4/5The Everything Guide to Algebra: A Step-by-Step Guide to the Basics of Algebra - in Plain English! Rating: 4 out of 5 stars4/5Introducing Game Theory: A Graphic Guide Rating: 4 out of 5 stars4/5Calculus Made Easy Rating: 4 out of 5 stars4/5The Thirteen Books of the Elements, Vol. 1 Rating: 0 out of 5 stars0 ratingsSee Ya Later Calculator: Simple Math Tricks You Can Do in Your Head Rating: 4 out of 5 stars4/5A Mind for Numbers | Summary Rating: 4 out of 5 stars4/5Logicomix: An epic search for truth Rating: 4 out of 5 stars4/5Summary of The Black Swan: by Nassim Nicholas Taleb | Includes Analysis Rating: 5 out of 5 stars5/5Limitless Mind: Learn, Lead, and Live Without Barriers Rating: 4 out of 5 stars4/5The Everything Everyday Math Book: From Tipping to Taxes, All the Real-World, Everyday Math Skills You Need Rating: 5 out of 5 stars5/5
Reviews for Matrices and Linear Algebra
5 ratings0 reviews
Book preview
Matrices and Linear Algebra - Hans Schneider
1
The Algebra
of Matrices
1. MATRICES: DEFINITIONS
This book is entitled Matrices and Linear Algebra, and linear
will be the most common mathematical term used here. This word has many related meanings, and now we shall explain what a linear equation is. An example of a linear equation is 3x1 + 2x2 = 5, where x1 and x2 are unknowns. In general an equation is called linear if it is of the form
where x1, …, xn are unknowns, and a1, …, an and b are numbers. Observe that in a linear equation, products such as x1x2 or x3⁴ and more general functions such as sin x1 do not occur.
In elementary books a pair of equations such as
is called a pair of simultaneous equations. We shall call such a pair a system of linear equations. Of course we may have more than three unknowns and more than two equations. Thus the most general system of m equations in n unknowns is
The aij are numbers, and the subscript (i, j) denotes that aij is the coefficient of xj in the ith equation.
, or π. The coefficients could just as well be complex numbers. This case would arise if we considered the equations
a number system.
In Chapter 2 we shall study systems of linear equations in greater detail. In this chapter we shall use linear equations only to motivate the concept of a matrix. Matrices will turn out to be extremely useful not only in the study of linear equations but also in much else.
If we consider the system of equations (1.1.2), we see that the arrays of coefficients
convey all the necessary information. Conversely, given any arrays like
we can immediately write down a corresponding system of equations
stand for the real or complex numbers. With this as motivation we adopt the following
( 1.1.4)DEFINITIONA matrix that occur in the matrix A are called the entries of A.
Examples of matrices are
is
where each aij , that is, either a real or complex number.
The horizontal array
is called the ith row of A and we shall denote it by ai*. Similarly, the vertical array
is called the jth column of A, and we shall often denote it by a*j. Observe that aij is the entry of A occurring in the ith row and the jth column.
If the matrix A has m rows and n columns it is called an m × n matrix. In particular, if m = n the matrix is called a square matrix. At times an n × n square matrix is referred to as a matrix of order n. Two other special cases are the m × 1 matrix, referred to as a column vector, and the 1 × n matrix, which is called a row vector. Examples of each special case are
Usually we denote matrices by capital letters (A, B, and C), but sometimes the symbols [aij], [bkl], and [cpq] are used. The entries of the matrix A will be denoted by aij, those of B by bkl, and so forth.
It is important to realize that
are all distinct matrices. To emphasize this point, we make the following
( 1.1.6)DEFINITIONLet A be an m × n matrix and B a p × q matrix. Then A = B if and only if m = p, n = q, and
Thus for each pair of integers m, n we may consider two sets of matrices. One is the set of all m × n , and the other is the set of all m × n m, n m, nm, n m, n m, nm, n m, n m, n.
We shall now point out how the use of matrices simplifies the notation for systems of linear equations. Let
All the information in (1.1.2) is contained in A, x, and b. (It is convenient here to call x a matrix or column vector even though its entries are unknowns.) Thus we could, purely symbolically at the moment, write
Ax = b.
Thus we reduce two linear equations in three unknowns to one matrix equation. If we have m linear equations in n unknowns, as in (1.1.3), we can still use the matrix form
where
For the time being, (1.1.7) is merely a symbolic way of expressing the equations (1.1.3). As another example, let
Then Ax = b is shorthand for
In Section 3 we shall see that the left side of (1.1.7) may be read as the product of two matrices. Note that b and x are column vectors; b is a column vector with m elements and x is a column vector with n elements. This method of writing the linear equations concentrates attention upon the essential item, the coefficient array.
EXERCISES
1. Find the matrices (A, b, x), corresponding to the following
systems of equations.
2. What systems of equations correspond to the following pairs of matrices?
2. ADDITION AND SCALAR
MULTIPLICATION OF MATRICES
We wish to see the effect on their corresponding matrices of adding two systems of equations. Consider the following two systems of equations with their corresponding matrices:
Adding the corresponding equations, we obtain a third system of equations,
and its matrices
Here we see how C may be obtained directly from A and B without reference to the original system of equations as such. We simply add the entries in like position. Thus
c22 = + 4 = a22 + b22
and we shall write C = A + B.
We shall define this sum A + B only when A and B have the same number of rows and columns. Two matrices with this property will be called additively conformable. For such matrices we make the following
( 1.2.1)DEFINITIONIf A and B are two m × n matrices, then the sum C = A + B is given by
With m and n m, n, the set of all m × n , together with the operation + defined above for these additively conformable matrices. We have the following
( 1.2.2)THEOREM(1) Closurem, n is closed under addition in the sense that the sum of any two m × n matrices is defined and is again an m × n matrix.
(2) Associativity: (A + B) + C = A + (B + C).
(3) Existence of a zero matrix 0: There exists an m × n matrix each of whose entries is zero,
with the property that
A + 0 = A
for any matrix A m, n.
(4) Existence of negatives: Given any matrix A m, n, there is a unique X m, n such that
A + X = 0
(As with numbers, we denote X by − A. −A is called the additive inverse of A, and for A = [aij] we have −A = [−aij].)
(5) A + B = B + A.
Note that we are using 0 in two senses: (1) the real number zero, and (2) the matrix all of whose entries are the real number zero.
As an example of (4), let m = 3, n = 2, and
PROOF
We will give a proof of (2); the other assertions are equally easy.
Let aij, bij, and cij be the general entries in A, B, and C, respectively. Then the general term of (A + B) + C is given by (aij + bij) + cij, while the general term of A + (B + C) is given by aij + (bij + cij). But the real and the complex numbers satisfy the associative law for addition, so that
Consequently, (A + B) + C = A + (B + C).
Remark Those readers who know some abstract algebra will recognize that Theorem (m, n is an abelian group under addition.
We now examine the result of multiplying the system of equations
and its matrix form
by some number α (which is called a scalar in this connection). Let α = 2. Upon multiplying we obtain
and the matrix form
Inspection of (1.2.4′),
leads immediately to the
( 1.2.5)DEFINITION OF SCALAR MULTIPLICATIONLet A = [aij] be an m × n , and let α . Then B = αA = Aα is defined by
Directly from the definition we have the following
(1.2.6)THEOREM
(1) α(A + B) = αA + αB.
(2) (α + β)A = αA + βA.
(3) (αβ)A = α(βA).
(4) 1A = A.
To illustrate (1) let α = 2,
Then
So in this particular case we have α(A + B) = αA + αB. The proofs are straightforward and are left to the reader.
Although αA = Aα for α a scalar and A a matrix, if A is a row vector, we shall always write the product as αA. However, if B is a column vector, the product of B and α will be written as Bα.
EXERCISES
1. None of the following matrices are equal. Check! Which of them are additively conformable? Calculate the sum of all pairs of additively conformable matrices of the same order.
2. Add the following pairs of systems of equations by finding their matrices A, b, and x. After performing the matrix addition translate the result back into equation form. Check your result by adding the equations directly.
3. Check parts (3) and (4) of Theorem (1.2.2).
4. Prove part (5) of Theorem (1.2.2). [Hint: In this and exercise 3 use the definition of matrix addition as illustrated in the proof of part (2).]
5. Let C be a n × n matrix. The trace of C, tr(C. Deduce the following results:
(a) tr(A + B) = tr(A) + tr(B).
(b) tr(kA) = k tr(A).
6.
Find the scalar product of the given systems of equations by 3. First perform the multiplications. Find the associated matrices and calculate the scalar product using the definition. Translate the result back into equations. Compare the respective answers.
7. Check the assertions following the definition of scalar multiplication.
8. For the system of equations
find the product of the corresponding matrix equation with the scalar (1 − i).
3. MATRIX MULTIPLICATION
Suppose we now consider the system of equations
and its matrix form
We wish to define a multiplication between matrices so that (1.3.2) is not merely symbolic but will give us back (1.3.1) when the multiplication is carried out.
Suppose we also have
and the corresponding matrices
so that
We can substitute (1.3.1) into (1.3.3) to obtain a new set of equations in terms of z and x. We want to be able to make the corresponding substitution of (1.3.2) into (1.3.4) and have the following hold for the matrices A and B:
If we compute the equations resulting from the substitution of (1.3.1) into (1.3.3) we obtain
so that if the product BA is to be defined we should have
Observe that (1.3.3) was a system of three equations in the two unknowns y1 and y2, and that (1.3.1) had one equation for each of y1 and y2. Thus in our substitution the number of unknowns in (1.3.3) equals the number of equations in (1.3.1).
In terms of matrices this means that the number of columns of B equals the number of rows of A. Further, after the substitution has been carried out in (1.3.6), it is clear that the number of equations equals the number of equations in (1.3.3), while the number of unknowns is the same as in (1.3.1). Thus our new matrix BA will have the same number of rows as B and the same number of columns as A.
With this in mind we shall call an m × n matrix B and an n′ × p matrix A multiplicatively conformable if and only if n = n′, that is, if the number of columns of B equals the number of rows of A. We shall define multiplication only for multiplicatively conformable matrices. Further, BA will be an m × p matrix; that is, BA will have as many rows as B and as many columns as A.
Keeping (1.3.7) and subsequent remarks in mind we make the following
( 1.3.8)DEFINITIONLet B be an m × n matrix and A be an n × p matrix, so that B has as many columns as A has rows. Let
Then the product
is defined by
that is,
We emphasize that the product of an m × n matrix with an n × p matrix yields an m × p matrix. Observe further that in (1.3.7) the product BA is defined, while the product AB is not, unless m = p.
To illustrate further the definition of matrix multiplication, let us look at our previous example with letters in place of the numerical coefficients. The system (1.3.1) becomes
while (1.3.3) and (1.3.6) are rewritten, respectively, as
The corresponding coefficient matrices are
Note that if A is the matrix of (1.1.5) and b and x are the column vectors of (1.1.8), then Ax is precisely the left side of equations (1.1.3). This justifies our use of Ax = b in Section 1.
We can immediately verify the following special cases:
(1) row vector × matrix = row vector.
(2) matrix × column vector = column vector.
(3) row vector × column vector = 1 × 1 matrix (a scalar).
(4) column vector × row vector = matrix.
( 1.3.12)EXAMPLELet
and C = BA. Then
We remark that C can be obtained from B and A by a row into column multiplication. Thus c11 is the sum of the entrywise products going across the first row of B and down the first column of A, so in (1.3.12)
c11 = 1.0 + 1.1 + 2.2 = 5 = b1.a.1,
where b1* is the first row of B and a*1 is the first column of A. Similarly, c32 can be obtained by going across the third row of B and down the second column of A. Again, in (1.3.12) we have
c32 = 1.1 + 4(–1) + 9.0 = –3 = b3.a.2.
In general
The formula (1.3.13) holds not just in this one example but whenever the product C = BA is defined.
Also, in (1.3.12)
Thus
These formulas again illustrate a general situation. Thus, if C = BA, where B is m × n and A is n × p, then
To prove (1.3.14) we need merely compare the components of ci* with those of bi1a1* + … + binan*.
The right side of equation (1.3.14) is obtained by multiplying each row of A by a scalar and then adding. A sum of the form β1a1* + … + βnan* is called a linear combination of the rows a1* to an* (Linear combinations will be of importance in Chapter 3 and will be studied there in great detail.)
In view of the importance of (1.3.14) we shall sum it up by a theorem.
( 1.3.15)THEOREMIf C = BA, then the ith row of C is a linear combination of the rows of A with coefficients from the ith row of B.
Similarly, we observe that in Example (1.3.12),
Again this is a special case of a general formula. Thus if C = BA, where B is m × n and A is n × p, then
Extending the notion of linear combinations to columns in the obvious way we have the analogue to Theorem (1.3.15) in
( 1.3.17)THEOREMIf C = BA, then the jth column of C is a linear combination of the columns of B with coefficients from the jth column of A.
The reader should also check that in Example (1.3.12),
and in general
Note that for each k, b*kak* is an m × p matrix.
Let us now return to some properties of matrix multiplication. We can now prove
( 1.3.18)THEOREMFor all matrices A, B, and C, and any scalar α:
(1) A(BC) = (AB)C.
(2) A(B + C) = AB + AC.
(3) (A + B)C = AC + BC.
(4) α(AB) = (αA)B.
whenever all the products are defined.
PROOF OF
(1) Let AB = D, BC = G, (AB)C = F, and A(BC) = H. We must show that F = H.
The proofs of (2), (3), and (4) are similar and are left to the reader as an exercise.
As an example of (1) in the theorem, let
Then
We see in this example that F = H.
( 1.3.19)REMARKIf A, B, and C are n × n matrices, then all sums and products in (n,n under matrix addition, matrix multiplication, and scalar multiplication.
The identity contained in part (1) of Theorem (1.3.18) is called the associative law for multiplication. The identities contained in parts (2) and (3) together are called the distributive law. The commutative law for multiplication would mean that AB = BA whenever the products are defined, but in the beginning of Section 4 we shall show by an example that there are matrices A and B such that AB and BA are both defined but AB ≠ BA. (In this connection, see exercise 4 of this section.)
EXERCISES
1. Check the following products.
2.Which of the matrices in exercise 1 of Section 2are multiplicatively conformable? Calculate three such products.
3. Check parts (2) and (3) of Theorem (1.3.18). (Hint: Use the definition.)
4. Compute the following products:
5. Give an example of a pair of matrices A and B such that AB = BA and an example of a pair C and D such that CD ≠ DC.
6. A square n × n matrix whose only nonzero entries occur in positions aii, i = 1, 2, …, n is called a diagonal matrix (that is, akl = 0 if k ≠ l). In general a11, a22, a33, …, an n is called the main diagonal. Each of the matrices in 4(a) and (b) is diagonal. What about their products? Prove, in general, that the product of two n × n diagonal matrices D1, D2 is diagonal and that D1 and D2 commute. (Hint: Let D1 = [dij] and D2 = [d′kl]. Use the definition of matrix multiplication to show that there is at most one nonzero term in ci1 = ∑jdijd′ji.) The matrix
is an example of a diagonal matrix.
7. Square matrices whose only nonzero entries occur on or above the main diagonal are called upper triangular matrices (those for which aij = 0 if i > j). The matrices in 4(c) to (e) are examples of upper triangular matrices. Prove that the product of two upper triangular matrices T1, T2 is an upper tria].
8. A strictly upper triangular matrix is a triangular matrix whose diagonal entries are zero (that is, aij = 0 if i ≥ j). Prove that the product of a strictly upper triangular matrix S and an upper triangular matrix T is strictly upper triangular. Since matrix multiplication is not commutative one must check two products: ST and TS. An example of a strictly upper triangular matrix is
9. Let
Calculate J², J³, J⁴, JA, BJ, J²A, and BJ².
10. Let S be any n × n strictly upper triangular matrix. Prove that Sn = 0.
11. Let Eμν = [eij], where
(b) Show that if B = [bμν] is any n × n .
(c) Let A be an arbitrary n × n matrix with entry aij ≠ 0. Show that
(Note: Here i and j are fixed.)
12. Use mathematical induction to prove
. Show that A² = 0.
14. Let
Compute the products AB and BC. Note that A ≠ B.
15. Find two square matrices, C and D, such that (C + D)² ≠ C² + 2CD + D². Why does equality not hold in general?
16. Find all 2 × 2 matrices X such that
17. Find 2 × 2 matrices A ≠ 0 and B ≠ 0 such that A²B² = 0.
18. Let
and let B be any 3 × 3 matrix. Show that the columns of BA are equal. What about the rows of AB?
19. Let B and C be n × n matrices such that C² = 0 and BC = CB. If A = B + C, show that for all integers k > 0, Ak + ¹ = Bk(B + (k + 1)C).
4. SQUARE MATRICES, INVERSES,
AND ZERO DIVISORS
m, n is the set of all matrices having m rows and n columns. If m = nn, n is said to be a square matrix. If A and B n, n we observe that AB and BA n, n. But a simple example shows that in general
AB ≠ BA;
that is, multiplication is not commutative. For let
This example also shows that even if A ≠ 0 and B ≠ 0, we may have AB = 0. Of course, 0B = 0 also. Hence if C = 0 and A and B are the matrices of the above example, then AB = CB. We see that if AB = CB it is not necessarily true that A = C.
n, n we have a special matrix, called the identity matrix,
defined by
For example,
In has the property that for any A n n,
InA = A = A In.
We note in passing that if B m, n,
ImB = BandBIn = B,
and that for any scalar α,
(αIm)B = αBandB(αIn) = αB.
Recall that for any number x ≠ 0, there is a number y = x−1, called the multiplicative inverse (or reciprocal) of x, such that
xx−1 = 1.
n, n, however. For let
Then finding an X 2,2 such that
is not possible. Regardless of the X chosen,
and so the second row of AX must be [0 0]. Thus we can never have equality in (1.4.1). However, in some cases we can have
AX = I = XA.
For example, let
Keeping this last example in mind, we formulate the following
( 1.4.2)DEFINITIONLet A be a n × n matrix. The n × n matrix X is said to be an inverse of A if and only if
XA = IandAX = I.
( 1.4.3)DEFINITIONThe square matrix A is termed nonsingular if and only if A possesses an inverse. If A has no inverse, A is called singular.
The actual computation of inverses will be taken up later.
( 1.4.4)THEOREMLet A be an n × n matrix. If
XA = IandAY = I,
then X = Y.
In particular, if A is nonsingular, the inverse is unique.
PROOF
X = XI = X(AY) = (XA)Y = IY = Y. In particular, if A is nonsingular and if X and Y are two inverses of A, then XA = I and AY = I, so that X = Y follows.
We shall denote the unique inverse of a nonsingular matrix A by A−1. Thus A−1A = I = AA−1. Later we shall show that if XA = I, then X = A−1.
In the example before Definition (1.4.2) we had
The computation there shows that X = A−1. As another example, let
Then
because
( 1.4.5)LEMMAIf the matrices A and B are nonsingular, then the product AB is nonsingular and
(AB)-1 = B-1A-1.
PROOF
(B−1 A−1)(AB) = B−1(A−1A)B = B−1IB = B−1B = I.
Similarly,
(AB)(B-1A-1) = A(BB)-1A-1 = AIA-1 = AA-1 = I.
By Theorem (1.4.4), (AB)−1, if it exists, is the unique matrix X such that X(AB) = I = (AB)X. Hence (AB)−1 exists and (AB)−1 = B−1A−1.
More generally we can prove by induction that if A1, A2, …, As are nonsingular, then the product
AsAs-1...A1
is nonsingular and
(AsAs-1...A1)-1 = A1-1A2-1...As-1.
If we recall