Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
S Chaturvedi
October 3, 2016
Contents
1 Finite dimensional Vector Spaces
1.1 Vector space . . . . . . . . . . . . . . . . . . . . . .
1.2 Examples . . . . . . . . . . . . . . . . . . . . . . .
1.3 Linear combinations, Linear Span . . . . . . . . . .
1.4 Linear independence . . . . . . . . . . . . . . . . .
1.5 Dimension . . . . . . . . . . . . . . . . . . . . . . .
1.6 Basis . . . . . . . . . . . . . . . . . . . . . . . . . .
1.7 Representation of a vector in a given basis . . . . .
1.8 Relation between bases . . . . . . . . . . . . . . . .
1.9 Subspace . . . . . . . . . . . . . . . . . . . . . . . .
1.10 Basis for a vector space adapted to its subspace . .
1.11 Direct Sum and Sum . . . . . . . . . . . . . . . . .
1.12 Linear Operators . . . . . . . . . . . . . . . . . . .
1.13 Null space, Range and Rank of a linear operator . .
1.14 Invertibility . . . . . . . . . . . . . . . . . . . . . .
1.15 Invariant subspace of a linear operator . . . . . . .
1.16 Eigenvalues and Eigenvectors of a linear operator .
1.17 Representation of a linear operator in a given basis
1.18 Change of basis . . . . . . . . . . . . . . . . . . . .
1.19 Diagonalizability . . . . . . . . . . . . . . . . . . .
1.20 From linear operators to Matrices . . . . . . . . . .
1.21 Rank of a matrix . . . . . . . . . . . . . . . . . . .
1.22 Eigenvalues and Eigenvectors of a matrix . . . . . .
1.23 Diagonalizability . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
3
. 3
. 4
. 4
. 4
. 4
. 5
. 5
. 5
. 6
. 6
. 6
. 7
. 7
. 7
. 7
. 7
. 8
. 8
. 8
. 9
. 9
. 9
. 10
1.24
1.25
1.26
1.27
1.28
1.29
1.30
1.31
1.32
1.33
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
11
12
13
14
14
14
14
15
15
.
.
.
.
16
19
19
19
.
.
.
.
.
.
.
22
22
22
23
23
25
27
28
1.1
Vector space
1.2
Examples
1.3
1.4
Linear independence
1.5
Dimension
1.6
Basis
In a finite dimensional vector space V of dimension n any n linearly independent vectors x1 , x2 , , xn V of linearly independent vectors are said to
provide a basis for V . In general there are infinitely many bases for a given
vector space.
1.7
x1
x 7 x =
xn
In particular
1
0
0
0
e1 7 e1 =
, , en 7 en =
0
1
1.8
Let e1 , e2 , , en and e01 , e02 , , e0n be two bases for a vector space V . Since
e1 , e2 , , en is a basis each e0i can be written as a linear combination of
e1 , e2 , , en :
n
X
0
ei =
Sji ej
j=1
The matrix S must necessarily be invertible as since e01 , e02 , , e0n is a basis
and each ei can be written as a linear combination of e01 , e02 , , e0n :
ei =
n
X
j=1
1 0
Sji
ej
Two bases are thus related to each other through an invertible matrix S
there are as many bases in a vector space of dimension n as there are n n
invertible matrices. The set of all n n invertible real (complex) matrices
form a group denoted by GL(n, R)(GL(n, C)) Under a change of basis the
components x of a vector x in e1 , e2 , , en are related to the components
x0 of x in the basis e01 , e02 , , e0n as follows
x0i
n
X
1
Sji
xj
j=1
1.9
Subspace
1.10
1.11
1.12
Linear Operators
1.13
1.14
Invertibility
1.15
1.16
1.17
n
X
Aji ej
j=1
The matrix A of the coefficients Aij is called the representation of the linear
operator in the basis e1 , e2 , , en
It can further be seen that the if the linear operators A and B are respectively represented by A and B respectvely in a given basis in V then the
operator AB is represented in the same basis by AB.
1.18
Change of basis
Clearly the representation of a linear operator depends on the chosen basis. If we change the basis the matrix representing the operator will also
change. Let A and A0 be the representation of the linear operator in the
bases e1 , e2 , , en and e01 , e02 , , e0n related to each other as
e0i
n
X
Sji ej
j=1
1.19
Diagonalizability
1.20
From the discussion above it is evident that for any vector space V of dimension n, whatever be its nature, after fixing a basis, we can make the following
identifications:
x V x Cn
Linear operator A on V A Mn (C)
Rank of A Rank of A
Invertibility of A Invertibility of A
Diagonalizability of A Diagonalizability of A
Eigenvalues and eigenvectors of A Eigenvalues and eigenvectors A
In mathematical terms, every finite dimensional vector space of dimension n
is isomorphic to Cn .
1.21
Rank of a matrix
1.22
says that the complex field is algebraically complete and is the main reason
behind considering vector spaces over the complex field.
The roots 1 , , n of the characteristic equution, the eigenvalues or the
spectrum of A may all be distinct or some of them may occur several times.
An eigenvalue that occurs more than once is said to be degenerate and the
number of times it occurs is called its (algebraic) multiplicity or degeneracy.
Having found the eigenvalues one proceeds to construct the corresponding
eigenvectors. Two situations may arise
An eigenvalue k is non degenerate. In this case, there is essentially (
or upto multiplication by a scalar) there is only one eigenvector corresponding to that eigenvalue.
An eigenvalue k occurs fold degenerate. In this case one may or may
not find linearly independent eigenvectors. Further, there is much
greater freedom in choosing the eigenvectors any linear combination
of the eigenvectors corresponding to a degenerate eigenvalue is also an
eigevector corresponding to that eigenvalue.
Given the fact that the eigenvectors corresponding to distinct eigenvalues are
always linearly independent, we can make the following statements:
If the eigenvalues of an n n matrix A are all distinct then the corresponding eigenvectors, n in number are linearly independent and hence
form a basis in Cn
If this is not so, the n eigenvectors may or may not be linearly independent. (Special kinds of matrices for which the existence of n linearly
independent eigenvectors is guranteed regardless of the degeneracies or
otherwise in its spectrum, will be considered later.
1.23
Diagonalizability
In view of what has been said above, a matrix whose eigenvalues are all
distinct can certainly be diagonalized. When this is not so i.e. when one or
more eigenvalues are degenerate we may or may not be able to diagonalize
depending on whether or not it has n linearly independent eigen vectors. If
the matrix can not be diagonalized what is the best we can do? This leads us
to the Jordan canonical form ( of which the diagonal form is a special case).
1.24
Ji =
Needless to say that the diagonal form is a special case of the Jordan form
in which each box is of 1 dimension.
Further details concerning the sizes of the blocks, explicit construction of
S which effects the Jordan form have to be worked out case by case and will
be omitted.
1.25
Cayley Hamilton theorem states that every matrix satisfies its characteristic
equation. Thus if the characteristic equation of a 3 3 matrix is 3 + c2 2 +
c1 + c0 then A satisfies A3 + c2 A2 + c1 A + c0 I = 0. This thus expresses A3 ,
and hence any higher power of A, as a linear combination of the A2 , A and
I. This result very useful in explicit computation of functions f (A) of any
n n an matrix. We illustrate below the procedure for a 3 3 matrix.
Recall that if A has eigenvalues 1 , , n with corresponding eigenvectors x1 , , xn then f (A) has eigenvalues f (1 ), , f (n ) with x1 , , xn
as eigenvectors.
Now consider a function f (A) of, say, a 3 3 matrix. Cayley Hamilton
theorem tells us that computing any function of A ( which can meaningfully
be expanded in a power series in A) reduces, in this instance, to computing
powers of A upto two.
f (A) = a2 A2 + a1 A + a0 I
Only thing that remains is to determine the three coefficients a2 , a1 , a0 and
to do that we need three equations. If the three eigenvalues are distinct, by
virtue of what was said above one obtains
f (1 ) = a2 21 + a1 1 + a0
f (2 ) = a2 22 + a1 2 + a0
f (3 ) = a2 23 + a1 3 + a0
which when solved for a2 , a1 , a0 yield the desired result.
What if one of the eigenvalues 1 is two fold degenerate i.e what if the
eigenvalues turn out to be 1 , 1 , 2 ? We then get only two equations for the
three unknowns. It can be shown that in such a situation the third equation
needed to supplement the two equations
f (1 ) = a2 21 + a1 1 + a0
12
f (2 ) = a2 22 + a1 2 + a0
is
f ()
|=1 = 2a2 1 + a1
What if all the three eigenvalues are the same i.e. if the eigenvallues turn
out to be 1 , 1 , 1 . The three desired equations then would be
f (1 ) = a2 21 + a1 1 + a0
f ()
|=1 = 2a2 1 + a1
2 f ()
|=1 = 2a2
2
One can easily recognise how the pattern outlined above extends to more
general situations.
1.26
A scalar product is a rule which assigns a scalar, denoted by (x, y), to any
pair of vectors in x, y V such that
(x, y) = (y, x) (hermitian symmetry)
(x, y + z) = (x, y) + (x, z) (linearity)
(x, x) 0. Equality holds if and only x is 0 (Positivity)
Examples:
Cn : (x, y) = x y
Mn (C) : (x, y) = Tr[x y]
Rb
Pn (t) : (x, y) = a dt w(t)x (t)y(t), for any fixed w(t) such that w(t)
0 for t (a, b)
p
A vector x is said to be normalized if its norm ||x|| (x, x) = 1. If
a vector is not normalized, it can be normalized by dividing it by its norm.
Two vectors x, y H are said to be orthogonal if (x, y) = 0. A vector space
V equipped with a scalar product is called a Hilbert space H. On a given
vector space one can define a scalar product in infinitely many ways. Hilbert
spaces corresponding to the same vector space with distinct scalar products
are regarded as distinct Hilbert spaces.
13
1.27
Orthogonal complement
1.28
Orthonormal Bases
n
X
Aji ej
j=1
1.29
Two orthonormal bases e1 , e2 , , en and e01 , e02 , , e0n are related to each
other by a unitary matrix :
e0i =
n
X
Uji ej , U U = I
j=1
1.30
1.31
Gram Matrix
1.32
i.e. the matrix for A is simply the complex conjugate transpose of the matrix
A for A. Remember that this is so only if the basis chosen is an orthonormal
basis and is not so otherwise.
1.33
Positive operator : An opearator A such that (x, Ax) 0 for all pairs
x is called a positive ( or non negative) operator. For such an operator
it can be shown to have the following properties:
its eigenvalues are 0
It is necessarily self adjoint and hence inherits all the properties
of a self adjoint operator.
Projection operator: A self adoint opearator P such that P 2 = P is
called a projection operator Such an operator can be shown to have the
following properties:
its eigenvalues are either 1 or 0
Being self adjoint, it has all the properties of a self adjoint operator.
the number of 1s in its spectrum give its rank.
A projection operator P of rank m fixes an m dimensional subspace of H. The operator Id P is also a projection operator of
rank n m and fixes the orthogonal complement of the subspace
corresponding to P.
If P1 , P2 , , Pn denote the projection operators corresponding
to the one dimensional subspaces determined by an orthonormal
basis e1 , e2 , , en H then
Tr[Pi Pj ] = ij Pi ,
P1 + P2 + + Pn = Id
n
X
Aij xi xj , Aij R, x Rn
i,j
17
18
1.34
Two self adjoint operators A, B, A = A, B = B can be diagonalized simultaneously by a unitary transformation if and only if the commute [A, B] = 0.
The task of diagonalizing two commuting self adjoint operators essential
consists in consisting a common eigenbasis which, as we know , can always
be chosen as an orthonormal basis. The unitary operator which effects the
simultaneous diagonalization is then obtained by putting the elements of
the common basis side by side as usual. If one of the two has degenerate
eigenvalues then its eigenbasis is also the eigenbasis of the other. More work
is needed If neither of the two has a non degenerate spectrum- suitable linear
cobinations of the eigenvectors corresponding to a degenerate eigenvalues
have to be constructed so that they are also the eigenvectors of the other.
The significance of this result in the context of quantum mechanics arises
in the process of labelling the elements of a basis in the Hilbert space by the
eigenvalues of a commuting set of operators.
1.35
d2
x = Kx,
dt2
1.36
n X
m
X
ij ei fj ; i = 1, n, j = 1, , m
i=1 j=1
(It is assumed that the formal symbol satisfies certain common sense
properties such as (u + v) z = u z + v z; (u) z = (u z) =
u (z) etc. )
Here a few comments are in order:
Elements x of V1 V2 can be divided into two categories, product or
separable vectors i.e those which can be written as uv; u V1 , v V2
and non separable or entangled vectors i.e. those which can not be
written in this form.
Operators A and B on V1 and V2 may respectively be extended to
operators on V1 V2 as A I and I B.
Operators on V1 V2 can be divided into two categories : local operators
i.e. those which can be written as A B and non local operators i.e.
those which can not be written in this way.
If the operators A and B on V1 and V2 are represented by the matrices
A and B in the bases e1 , , en and f1 , , fm then the operator A
B is represented in the lexicographically ordered basis ei fj ; i =
20
A11 B A1n B
AB =
An1 B Ann B
This construction can easily be extended to tensor products of three or
more vector spaces.
Tensor products of vector spaces arise naturally in the description of
composite systems in quantum mchanics. The notion of entanglement
pays a crucial role in quantum information theory.
21
2
2.1
P
n
An infinite series of the form
n=0 an (x x0 ) is called a power series
around x0 . It converges for values of x lying in the interval |x x0 | < R
where R, the radius of convergence of the power series is given by
lim
R =n |
an
|
an+1
Stated in words, the infinite series converges for values of x lying in the
interval x0 R to x0 + R where R is to be computed as above.
2.2
pn (x x0 ) , q(x) =
qn (x x0 )n
n=0
n=0
If this is not so i.e if either p(x) or q(x) or both are not analytic at x = x0
then x0 is said to be a singular point.
If x = x0 is a singular point such that (x x0 )p(x) and (x x0 )2 q(x) are
analytic at x = x0 then x0 is called a regular singular point of the differential
equation.
22
2.3
an x n
n=0
2.4
X
y(x, ) =
an xn+
n=0
23
()xn+
n=0
The two independent solutions y1 (x) and y2 (x) are then given by
y1 (x) = y(x, 1 );
y2 (x) = y(x, 2 )
y2 (x) =
y(x, )
|=1
0.aN = 0
2. B
0.aN 6= 0
24
In the case A one can simply put aN = 0 ( If one doesnt, one simply
generates an expression proportional to the first solution)
In the case B the second solution is obtained by putting a0 = ( 2 )b0 ,
solving for an s in terms of b0 to obtain
y(x, ) = b0
()xn+
n=0
y(x, )
|=2
Note that in the cases II and III B, the second solution, in general, has
the structure
X
()xn+2 +
y2 (x) = log(x)
n=0
2.5
dy
d2 y
+ x + (x2 2 )y = 0
2
dx
dx
Here the roots of the indicial equation turn out to be and and one has
Case I : 2 not an integer
Case II : = 0
Case III A 2 an odd integer
Case III B 2 an even integer i.e an integer
25
x 2n+
x 2n
X
X
(1)n
(1)n
J (x) =
; J (x) =
;
n!(n + + 1) 2
n!(n + 1) 2
n=0
n=0
and the general solution of the Bessel equation can be written as
y(x) = c1 J (x) + c2 J (x)
In cases II, IIIA, IIIB, while the first solution (corresponding to the larger
root of the indicial equation) is still J (x), to obtain the second solution
one has to follow the procedure outlined above. However, in the context of
Bessels equation, one finds that the two soultions continue to remain valid
in the Case III A as well. So the only problematic cases that remain are Case
II and Case III B i.e. when is zero or an integer. Here ( and only in this
context) to find the second solution the following trick works. One defines a
suitable linear combination of J (x) and J (x) as follows
J (x) cos J (x)
sin
and for cases I and IIIA the general solution of the Bessel equation can
equally well be written as
Y (x) =
where
I (x) =
X
n=0
x 2n+
1
n!(n + + 1) 2
and
I (x) I (x)
2
sin
As before when is zero or an integer, K (x) becomes an indeterminate
form ( by virtue of the fact that IN (x) = IN (x) ) and has to be computed
as a limit. The functions K (x) are called modified Bessel functions of the
second kind
There are several equations of mathematical physics which are related to
the Bessel equation by suitable changes of variables. It can be shown that
the family of equations
K (x) =
x2
dy
d2 y
+ (1 2s)x + [(s2 r2 p2 ) + a2 r2 x2r ]y = 0
2
dx
dx
d2 u
du
+ t + (t2 p2 )u = 0
2
dt
dt
2.6
(1)
are said to have the Sturm Liouville form. They have the stucture of the
eigenvalue problem for the second order differenial operator
1 d
d
L
s(x)w(x)
(2)
w(x) dx
dx
If w(x) and s(x) are such that
27
2.7
If in addition to the two conditions on s(x) and w(x) above one further
stipulates that
s(x) is a polynomial in x of degree 2 with real roots
C1 (x)
d
1
[s(x)w(x)]; K1 a constant , is a polynomial of
K1 w(x) dx
degree 1.
then it can be shown that
Cn (x)
1
dn n
[s (x)w(x)];
Kn w(x) dxn
n = 0, 1, 2, [Rodriguez Formula]
28
4. satisfy recursion relations of the form Cn+1 (x) = (An x + Bn )Cn (x) +
Dn Cn1 (x)
A systematic analysis of the four conditions on s(x) and w(x) leads one
to eight distinct systems of orthogonal polynomials - Hermite, Legendre,
Laguerre, Jacobi, Gegenbauer and Tchebychef. [For details see, for instance
Mathematics for Physicists Dennery and Krzywicki]
s(x)
1
x
(1 x2 )
x
(1 x2 )
(1 x2 )
(1 x2 )
(1 x2 )
w(x)
2
ex
ex
1
x
x e
(1 x) (1 + x)
(1 x2 )1/2
(1 x2 )1/2
(1 x2 )1/2
Interval
(, )
[0, )
[1, 1]
[0, )
[1, 1]
[1, 1]
[1, 1]
[1, 1]
The table below gives the values of Kn appearing in the Rodriguez formula
and those of hn appearing in the orthogonality relation
Z b
dxw(x)Cn (x)Cm (x) = hn nm
a
29
hn
2 n!
1
2
2n + 1
n
It is often not possible to remember detailed expressions for these polynomial. However their suitably defined generating functions
F (x, t) =
an Cn (x)tn
have simple analytical expressions which can easily be remembered and these
can be used directly to deduce various properties of the corresponding polynomials. The table below gives the generating functions for the first three
polynomial systems
Cn (x)
Hn (x)
an
1
n!
Ln (x)
Pn (x)
F (x, t)
2
e2xtt
xt
1
e 1t
1t
1
1 2xt + t2
30