Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2.2 Range Space, Null Space and Matrix Rank . . . . . . . . . . . . . . . . . . . . . . . 2.3 Eigenvalue decompostion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.3.1 General matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
1 1 3
3
2.3.2
Singular Value Decomposition and Projection Operator Positive (Semi)Denite matrices . . . . . . . . . . . . . . Matrices with special structure . . . . . . . . . . . . . . . Matrix inversion Lemmas . . . . . . . . . . . . . . . . . . Systems of linear equations . . . . . . . . . . . . . . . . .
Inconsistent systems
Hermitian matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
. . . . .
6 10 11 12 12
12 13
2.8.1 2.8.2
Consistent systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.1 Introduction
In many digital signal processing, problems are modelised under linear algebra equations. As a consequence, it is classical to manipulate expressions such as:
Y = AX + N
where
the matrix
AX = B.
Complementary,
is the matrix or noise vector, the multivariate normal distribution has to be introduced.
This chapter was redacted using mainly the references [1] and [2].
or
or
the
transpose
v.
Let
be a
and
conjugate
mn
(.)
= (.)
, a vector is A Cmn ,
respectively
|.|
or
det(.).
Denition 1.
A,
also called the column space, is the subspace spanned by (i.e. all linear
A: R(A) = Cm1 | = A
for
Cn1
(2.2.1)
A.
CHAPTER 2.
Denition 2.
A,
N ( A ) = Cn 1 | A = 0
(2.2.2)
Denition 3.
follow
r
(i)
rank (A) A.
The latter number is by
R (A)
r = dim R (A)
(ii)
(2.2.3)
A,
(2.2.4)
r = dim R AT = dim R AH
(iii)
is the dimension of the nonzero determinant of maximum size that can be built from the element of
A.
Denition 4.
is said to be:
Rank decient whenever Full column rank if Full row rank if Nonsingular if
r = n m.
r = m n.
r = m = n.
Proposition 5. Premultiplication or postmultiplication of A by a nonsingular matrix does not change the rank of A. This directly follows from the denition of rank(A) because the aforementionned multiplications does not change the number of linearly independent columns (or rows) of A.
Then:
Proposition 6. Let
(2.2.5)
Proof.
independent columns of
B,
hence
rank (AB) rB .
AT ,
B by A can not increase the number of linearly A by B can not wich means that rank (AB) rA .
A=
k=1
H xk yk
Proof.
Since
can be written as
A= x1 xN
y1
. . .
yN
2.3.
EIGENVALUE DECOMPOSTION
then
rank (AB) = rank (B)
(2.2.7) submatrix, the postmultiplication of
Proof.
which by
A contains a nonsingular n n rank (B) (from R1). Hence, rank (AB) rank (B)
However, by 6,
symetric.
A Cm m
is said
Hermitian
if
AH = A.
is said
Denition 10.
A matrix
U Cm m
UH U = UUH = I
If
U Cmn ,
with
m > n,
is such that
UH U = I
is semi-unitary.
2.3.1
General matrices
A scalar
Denition 11.
of a matrix
AC
mm
if
In paticular, an eigenvalue
A:
(2.3.2)
|A I| = 0
and
is a vector in
N (A I).
p
The pair
(, x)
{(i , xi )}i=1 are p eigenpairs of A (with p m) Axi = xi (i = 1, , p) in the following compact form:
Observe that if
AX = X
where
(2.3.3)
X = [x1 xp ]
and
1
. . .
0 p
A and B
Denition 14.
A Cmm tr (A) =
is dened as
Aii
i=1
(2.3.4)
CHAPTER 2.
tr (A) =
i=1
(2.3.5)
Interestingly, while the matrix product is not commutative, the trace is invariant to commuting the factors in a matrix product.
Proof.
A straightforward calculation based on the trace denition and diagonal elements of a matrices product.
Proof.
I A 0 I I B I A B I 0 I
I A B I I A B I
I 0 B I I A 0 I
I AB 0 I 0
0 I
(2.3.8)
0 I BA
(2.3.9)
As matrices of left-hand side of (2.3.8) and (2.3.9) have the same determinant (see lemma 17), equal to , then the right-hand side matrices also have the same determinant.
Proposition 19. Let A Cmn and B Cnm , the non-zero eigenvalues of AB and BA are identical.
Proof.
Let
=0
be an eigenvalue of
AB,
then
2.3.2
Hermitian matrices
An important property of the class of Hermitian matrices, which does not necessarily hold for general matrices, is the following.
(ii) The m eigenvectors of A = A Cmm form an orthogonal set. In other words, the matrices whose columns are the eigenvectors of A is unitary.
It follows from (i) and (ii) and from equation (2.3.3) that for an Hermitian matrix we can write:
AU = U
where
UU = U U = I
are
real numbers.
Equivalently:
A = UU
which is so-called Eigen-Value Decomposition (EVD) of
(2.3.10)
A = A .
special case of the Singular Value Decomposition (SVD) of a general matrix discussed in next section. Another usefull result, associated with Hermitian matrices is the following:
2.3.
EIGENVALUE DECOMPOSTION
1 2 m
v Av 1 v v
(2.3.11)
This ratio (2.3.11) is called the Rayleigh quotient. As this ratio (equation 2.3.11) is invariant to the multiplication of v by any non-null complex number, we can rewrite equation (2.3.11) in the form:
m v Av 1 for any v Cm1 with v v = 1
(2.3.12)
The equalities in equation (2.3.12) are evidently achieved when v is equal to the eigenvectors of A associated with m and 1 respectively. Proof.
Let the EVD of
w = U v =
then we need to prove that
w1
. . .
wm
m
m w w =
k=1
for any
k |wk | 1
w Cm1
satisfying
w w =
k=1
However, this is readily verifyed as follows:
|wk | = 1
1
k=1
and
k |wk | =
k=1 m
(1 k ) |wk | 0
k |wk | m =
k=1
that concludes the proof. The following result extands the previous proposition.
(k m ) |wk | 0
k=1
V Cmn , with m > n, be a semi-unitary matrix (i.e. V V = I), and let A = A have its eigenvalues ordered (1 2 m ) then:
m n
k tr VH AV
k=mn+1 k=1
(2.3.13)
where the equalities are achieved, for instance, when the columns of V are the eigenvectors of A corresponding to (mn+1 , . . . , m ) and, respectively, to (1 , . . . , n ). The rato
tr (V AV) tr (V AV) = tr (V V) n
A = UUH S = UH V sH 1
. . .
and let
(m n)
sH m
CHAPTER 2.
(hence
sH k
is the
k th
row of
S).
tr V AV = tr V UU V = tr S S = tr SS
where
=
k=1
k ck
(2.3.14)
ck
obvioulsy
sH k sk , k = 1, . . . , m ck 0 k = 1, . . . , m
(2.3.15)
(2.3.16)
and
(2.3.17)
ck 1 k = 1, . . . , m
To see this, let row of
(2.3.18)
G Cm(mn)
[SG]
H gk
denote the
k th
G.
Then, by construction,
sH k
which is (2.3.18).
H gk
sk gk
H H = ck + gk gk = 1 = ck = 1 gk gk 1
Finally, by combining (2.3.14) with (2.3.16)-(2.3.18) we can readily verify that where the equality was achieved for
tr VH AV
satises (2.3.13),
c1 = = cmn = 0; cmn+1 = = cm = 1
and respectively
c1 = = cn = 1; cn+1 = = cm = 0
These conditions on
{ck }
is equal to
and
, respectively.
With this observation, the proof is concluded. Proposition (21) is clearly a special case of Proposition (22).
mn
A Cm n
U Cmm
and
V Cnn
A = UVH
By appropriate permutation, the diagonal element of
(2.4.1)
1 2 min(m,n)
The factorisation (2.4.1) is called the (2.4.1) satisfy:
a signicant result for both a theoretical and practical standpoint. We reiterate that the matrices in equation
UH U = UUH = I (m m) VH V = VVH = I (n n) ij = i 0 0
for for
i=j i=j
U.
AAH .
2.4.
Aare
the columns of
V.
AH A. A
are the diagonal elements
{i }
of
Note that
{i }
min(m, n)
eigenvalues of
AAH
or
AH A.
where
is the triple
(k , uk , vk )
uk
is the
k th
columns of
and
vk
is the
k th
V.
If
k > 0, k = 0,
Hence, for a matrix of rank
k = 1, . . . , r k = r, . . . , min (m, n)
A=
where
U1
r
U2
mr
1 0
0 0
H V1 H V2
}r H = U1 1 V1 }n r
(2.4.2)
1 Rrr
Proposition 23. Consider the SVD of A Cmn in (2.4.11) where r min(m, n). Then:
(i) U1 is an orthogonal basis of R(A) (ii) U2 is an orthogonal basis of N (AH ) (iii) V1 is an orthogonal basis of R(AH ) (iv) V2 is an orthogonal basis of N (AH ).
Proof.
We can see that (iii) and (iv) follow from the properties (i) and (ii) applied to
AH .
R(A) = R (U1 )
and respectively,
(2.4.3)
N (AH ) = N (U2 )
To show (2.4.3), note that
(2.4.4)
R (A) there
Then
exists
such that
= A
H = U1 1 V1 = U1 and so R (U1 ). Then we show that R (A) R (U1 ) . 1 Now, for R (U1 ), it exists a such that = U1 . From (2.4.11), it follows that U1 = AV1 and 1 = A V1 = A and then R (A), that is to say R (U1 ) R (A). But we had also R (A) R (U1 ), then R (A) = R (U1 ).
Same philosophy for the demonstration of (2.4.4):
1 H H H N A H A H = 0 V 1 1 U H 1 = 0 1 V1 V1 1 U1 = 0 U1 = 0
Now, any vector
can be written as
=
since
U1
U2
H H 0 = UH 1 = U1 U1 + U1 U2 = , so = 0, and thus = U2 . We can then conclude that N A N (U2 ). H H H Now, N (U2 ) there exists such that = U2 , then A = V1 1 U1 U2 = 0 N A
U1
U2
is non singular.
However,
The previous results, readily derived by using the SVD, has a number of intersting corollaries which complement the discussion on the range and nullspace in section 2.2.
CHAPTER 2.
A Cmn the subspaces R(A) and N (AH ) are orthogonal to each other and they together span C . Consequently, we say that N (AH ) is the orthogonal complement of R(A) in Cm , and vice
m
dim N (AH ) = m r
(2.4.5)
dim R(A) = r
(2.4.6)
The SVD of a matrix also provides a conveniant representation for the projectors onto the range and null space of
and
AH .
Let
Denition 25.
matrix (i)
y Cm1
that:
R(A)
is the
and
y R(A)
2
is minimum :
y y
Hereafter,
= min
over
R(A)
= xH x
denotes the
Proposition 26. Let A Cmn . The orthogonal projector onto R(A) is given by:
= U1 UH 1
(2.4.7)
Proof.
Let
y Cm 1
be an arbitrary vector. As
R(A) = R(U1 ),
R(A)
min y U1
Because
(2.4.9)
y U1
= =
H y H U1 UH 1 y
2
H UH I U 1 UH 1 y +y 1 y
+ UH 2 y
If readily follows that the solution to the minimization problem (2.4.9) is given by
= UH 1 y. Hence the y onto R (A) and the minimum distance from y to R (A) is UH 2 y . H immediately from (2.4.7) and the fact that N A = R (U2 ). y
onto
R (A)
H y U1 UH 1 y = U2 U2 y
which is in
R ( U2 )
R(A)
in 25 and in 26. As an aside, we remark that the orthogonal projector in (2.4.7) and (2.4.8) are next denition.
idempotent matrices ;
see
Denition 27.
The matrix
A Cmm
is
idempotent if
A2 = A
(2.4.10)
Furthermore, observe by making use of 19 that the idempotent matrix in (2.4.7), for example, has values equal to
eigen-
and
(m 1)
their eigenvalues are either zero or one. Finally, we present a result that even alone would be enough to make the SVD an essential matrix analysis tool.
2.4.
A Cmn , with elements Aij . Let the SVD of A(with eigenvalues arranged in a U1
p
U2
mp
1 0
0 2
H V1 H V2
}p H = U 1 1 V 1 }n p
(2.4.11)
= tr AH A =
i=1 j =1
|Aij | =
k=1
2 k
(2.4.12)
denote the square of the so-called Frobenius norm. Then the best rank-p approximant of A in the Frobenius norm metric, that is the solution of
min A B
B 2
(2.4.13)
is given by
B0 = U1 1 V1
(2.4.14)
Futhermore, B0 above is the unique solution to the approximation problem (2.4.13) if and only if p > p+1 . Proof.
It follows from 8 and (2.8.5) that we can parametrise
B = CDH
where
C Cmp
and
D Cnp
is of
course non-unique but, as we will see, this fact does not introduce any problem. By making use of (2.4.15) we can rewrite the problem (2.4.13) in the following form:
min A CDH
C,D
(2.4.16)
The reparametrised problem is essentially constraint free. Indeed, the full column rank condition that must be satised by
and
D,
for a given
C.
A CDH
= tr
D A H C CH C
CH C
DH CH C
CH A + AH I C CH C
H 1
CH A
(2.4.17)
By result (iii) in denition 29 in the next section, the matrix is positive semidenite for any
DA C C C
C C
D.
D CH C respect to D for:
CH A
D0 = AH C CH C
(2.4.18)
tr AH I C CH C
Next we minimise (2.4.19) with respect to
CH A
denote an orthogonal basis of
(2.4.19)
C.
Let
S Cm p
R (C);
that is,
S S=I
and
S = C
for some non singuliar
pp
matrix
I C CH C
CH = I SSH
(2.4.20)
By combining (2.4.19) and (2.4.20) we can restate the problem of minimising (2.4.19) with respect to
as:
S; SH S=I
max tr SH AAH S S
is given by
(2.4.21)
S0 = U1
which yields
10
CHAPTER 2.
C0 = U1 1
It follows that:
(2.4.22)
B0
= = =
H H C0 DH CH 0 = C0 C0 C0 0 A = S0 S0 A H H U1 UH U V + U V 1 1 1 2 2 2 1 H U 1 1 V 1
Futhermore, we observe that the minimum value of the Frobenius distance in (2.4.13) is given by
min(m,n)
A B0
H 2 U2 2 V2
=
k=p+1
2 k
If
p > p+1
B0 derived above is unique. Otherwise it is not unique. p = p+1 we can obtain B0 by using either the singular vectors associated with p or those p+1 , which will generally lead to dierent solutions.
A = AH Cmm
{k }k=1
Denition 29.
(i) (ii)
We say that
k 0 (k > 0 for pd) for k = 1, . . . , m. H A 0 (H A 0 for pd) for any (iii) There exists a matrix C such that
non-zero vector
Cm 1 .
A = CCH
with
(2.5.1)
rank (C) = m for pd) |A (i1 , . . . , ik )| 0 (> 0 for pd) for all k = 1, . . . , m and all indices i1 , . . . , ik [1, m], where A (i1 , . . . , ik ) is the submatrix formed from A by eliminating the i1 , . . . , ik rows and columns of A. (A (i1 , . . . , ik ) is called the principal submatrix of A). The condition for A to be positive denite can be simplied to requiring that |A (k + 1, . . . , m)| > 0 (for k = 1, . . . , m 1) and |A| > 0. (A (k + 1, . . . , m) is called a leading submatrix of A).
(iv) Of the previous dening conditions, (iv) is apparently more involved, and named the
Sylvester Criterion.
Denition 30.
Let
A = AH
A = CCH
is called the square root of If
A.
Sometimes such a
is denoted
C = A1/2 . B.
There are then an innite number of
is a square root of
A,
then so is
CB
square roots. Two often-used particular choices for square roots are: (i) Hermitian square root:
C = CH .
A = C2 ,
and : (2.5.3)
C = U1/2 UH
because as
A = U1/2 1/2 UH
= U1/2 UH
U1/2 UH
CC
Cholesky factor.
A.
If
If
is called the
Cholesky factor of
is positive denite, the Cholesky factor is unique. This decomposition is also called
2.6.
11
A Cmn
A=
where
1 z1
. . .
1 zn
. . .
(2.6.1)
m 1 z1
m1 zn
zk C
Denition 32.
A Cmn
is called:
Examples:
a1 a0 a1
.. .
a2 a1
.. .. .. . . .
.. .. .. . . .
..
a1n
. . . . . .
a1 a0 a1
a2 a1 a0
am1
is a Toeplitz matrix.
a1 a2
The discrete convolution operation can be constructed as a matrix multiplication, where one of the inputs is converted into a Toeplitz matrix. For example, the convolution of
and
m1
y (n) =
k=0
with
Hmn
h2 . . . hm2 = hm1 0 . . . . . . 0
h0 h1
0 h0
.. .
0 0
. . .
xn1 xn2 . . . . . . x 0
hm2 hm1 0
. . .
h1 hm2 hm1 0 0
0 h0
0 a A= b c
b c c d d e
d e f
hm1
is a Hankel matrix.
Proposition 33. The eigenvectors of a symmetric Toeplitz matrix A Rmm are either symmetric or skew. . .
0 1 0 0 1 0 1
0 1
0 0
12
CHAPTER 2.
Proposition 34. Let A Cmm , B Cnn , C Cmn , D Cnm . Then, providing that the matrix inverses
= I 0 0 I A 1 I 0 + A 1 C I I B1 D B DA1 C
1
DA1
B1
A CB1 D
I CB1
Let A Cmm , B Cnn , C Cmn , D Cnm and providing that the matrix inverses appearing below exist,
A CB1 D
1
Lemma 35.
= A1 + A1 C B DA1 C
DA1
A Cmn , B Cmp ,
and
X Cn p .
AX = B
where matrices We say that:
and
m = n,
m > n, m < n,
In the following, we rst examin the case where (2.8.1) has an exact solution and then the case where (2.8.1) can not be exactly satised.
2.8.1
Consistent systems
Proposition 36. The linear system (2.8.1) is consistent, that is it admits an exact solution X, if and only if
rank A B = rank (A)
(2.8.2)
Proposition 37. Let X0 be a particular solution to (2.8.1). Then the set of all solutions to (2.8.1) is given by:
X = X0 +
(2.8.3)
where C
np
Proposition 38. The system of linear equation (2.8.1) has a unique solution if and only if (2.8.2) holds and
A has a full column rank: rank (A) = n m
(2.8.4)
Proposition 39. Consider a linear system that satises the consistency condition in (2.8.2) . Let A have a rank r min (m, n), and let
A= U1
r
U2
mr
1 0
0 0
H V1 H V2
}r H = U 1 1 V 1 }n r
(2.8.5)
(2.8.7)
2.8.
13
Denition 40.
The matrix
1 A = U 1 1 V1
(2.8.8)
Denition 41.
AA A = A A AA = A A A and AA A
.
are Hermitian
A 1 = A
The computation of a solution to (2.8.1) is an important problem. This solution must be numerically stable. That is to say, the solution must be as unsensible as possible to perturbations (for example limited resolution of computer). In eect, what we can hope is to compute an exact solution to a slightly perturbated system of linear equations:
(A + A ) (X + X ) = (B + B )
One can show that the perturbations factor given by:
(2.8.9)
A B
multiplied by a proportionality
cond (A) =
where
1 n A
and where
(2.8.10)
and
cond(.)
means
condition number . If the corresponding ratio the problem is said well-conditioned . If this
2.8.2 Inconsistent systems
introduced in (2.9.2) is small (not to much gretter than one), ratio is large, the problem is said ill-conditioned .
The systems of linear equation that appear in applications are quite often perturbated and usually they do not admit any exact solution. Such systems are said incosistent and frequently they are overdetermined and have a matrix
rank (A) = n m
linear equations:
(2.8.11)
In what follows, we present two approaches to obtain an approxiamte solution to an inconsistent system of
AX
under the condition (2.8.11).
(2.8.12)
Denition 42.
The least square (LS) approximate solution to (2.8.12) is given by the minimiser
XLS
of the
following criterion
AX B
AH B
(2.8.13)
The inverse matrix in the above equation exists in view of (2.8.11). Remark.
Note that
XLS
always the
The reason is 1 and its eect on the perturbated systems. That is why the QR n
Q Cmm
R 0
Q1
n
Q2
mn
R 0
(2.8.14)
is called the
QR decomposition .
XLS = R1 QH 1 B
14
CHAPTER 2.
XLS
RXLS = QH 1 B
(2.8.15)
X = (X1 X2 Xn )
denote an
n1
random
is
m mi
So the mean vector
T
(2.9.1)
is
(2.9.2)
is said to be
multivariate normal
p (x) = (2 )
The characteristic function of
T
n/2
(det (R))
1/2
1 T exp (x m) R1 (x m) 2
(2.9.3)
( )
(2.9.4)
( )
= =
exp j T m j T R
1 exp 2 mT R1 m exp
1 + j R 1 m 2
R + j R1 m .
(2.9.5)
The characteristic function itself is a multivariate normal function of the frequency variable
2.9.1
Let
Linear transformations
be a linear transformation of a multivariate normal random variable:
Y = AT X AT : n m (m n)
thus,
(2.9.6)
Y : N AT m, AT RA
a
(2.9.7)
m n, rank m. If m > n,
If
AT RA
be nonsingular and
have
AT RA
mn
components of
Remark.
Linear transformations allow to synthetise random vector for normaly distributed multivariate vector.
2.9.
15
2.9.2
Diagonalising transformations
The correlation matrix is symetric and nonnegative denite. Therefor there exists an orthogonal matrix that:
U such
(2.9.8)
2 UT RU = diag 2 1 n
It follows that the vector
Y=U X
2 Y = UT X : N UT m, UT RU = diag 2 1 n
We say that the random variables
Y1 , Y2 , . . . , Yn
T
E
Writting the density for
Y UT m Y,
Y UT m Yk
2 = UT RU = diag 2 1 n
(2.9.10)
UT m
and variances
2 k:
n
p (y) =
k=1
22 k
1/2
exp
1 y k UT m 22 k
2 k
(2.9.11)
This approach provides a possibility to generate independents random variables. The transformation
Y=
UT X
is called a
Karhunen-Loeve
transform.
2.9.3
Quadratic forms
Linear functions of multivariate normal vectors remain multivariate normal. But what about quadratic functions? In some very important cases the quadratic function have a Let
2 distribution
(read Chi-Squared).
denote a
N (m, R)
Q = (X m) R1 (X m)
is a chi-squared distribution with
2 n.
is
p(q ) =
2.9.4
Let
(2.9.13)
be random vectors in
Rn
such that
vec [x y]
is a
2n-dimensional
The
z = x + jy
has a
= E [ z] R = E (z ) (z )
H T
C = E (z ) (z )
where
the location parameter (mean) vector the covariance matrix the relation matrix
can be an arbitrary
n-
must be Hermitian,
should be symetric.
P = R CH R1 C
is also non-negative denite. The density function for a complex normal distribution can be computed as:
f (z) =
exp
1 H T [z ] [z ] 2
R CH
C R
[z ] [z ]
16
CHAPTER 2.
A circular symetric complex normal distribution coresponds to the case of a null relation matrix If
C = 0.
z = x + jy
vec [x y]
structure:
x y
N
where .
() ()
1 2 2 ,
Remark.
normal, centered (
i.e.
z = x + jy ,
z C. z
and
N 0, 2 /2
f (z ) =
zz 1 exp 2 2 2 2
1 |z | exp 2 2 2
Bibliography
[1] P. Stoica and R. L. Moses,
http://matrixcookbook.com, 2008.
17