Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
book
Satish Bysany
Aalto University School of Electrical Engineering
March 1, 2011
[section]
Introduction
This is as set of notes describing solutions to Steven Kays book Fundamentals of Statistical Signal Processing: Estimation Theory. A brief review of
notation is in order.
1.1
Notation
I is identity matrix.
0 represents a matrix or vector of all zeros.
e is a column vector of all ones.
J is exchange matrix, with 1s on the anti-diagonal and 0s elsewhere.
ej is a column vector whose j th element is 1, rest all 0.
.
ab = aH b is the dot product of a and b
t f (t)
f
(t)
t1
f
(t)
f (t) = t2
t
..
tM f (t)
t h(t)
t f1 (t)
f2 (t)
f (t) = t
t
..
t fM (t)
Chapter 2
2.1
Problem 2.1
The data x = {x[0], x[1], . . . , x[N 1]} are observed where the x[n]s are
i.i.d. as N (0, 2 ). We wish to estimate the variance 2 as
2 =
N 1
1 X 2
x [n]
N
(1)
n=0
Solution
Now take the E() operator on both sides of Eq(1) and using the fact
that, for any two random variables X and Y ,
E(X + Y ) = E(X) + E(Y )
N 1
N 1
1 X
1 X 2 N 2
2
=
E x [n] =
=
= 2
N
N
N
n=0
(2)
n=0
Hence the estimator 1 is unbiased. Note that, this result holds even if the
x[n]s are not independent!
Next, applying the variance operator var() on both sides of Eq(1) and
using the fact that, for independent random variables X and Y ,
var(aX + bY ) = a2 var(X) + b2 var(Y )
2
N 1
1 X
var
2 = 2
var x2 [n]
N
(3)
n=0
Let X N (0, 1) be normal distribution with zero-mean and unit variance. Then, by definition, Y = X 2 21 is chi-square
distributed
with
1 degree of freedom. We know that mean 2n = n, var 2n = 2n, so,
var(Y ) = var(X 2 ) = 2 1 = 2.
Introducing Z = X, implies that var(Z) = 2 var(X) = 2 . Since
E(Z) = E(X) = 0, we conclude Z N (0, 2 ).
Now consider var(Z 2 ) = var( 2 X 2 ) = 4 var(X 2 ) = 2 4 . Since each of
x[n] N (0, 2 ), we have,
var(x2 [0]) = var(x2 [1]) = = var(x2 [N 1]) = 2 4
Hence, Eq(3) simplifies to
var
N 1
1 X
2 4 N
2 4
= 2
(2 4 ) =
=
N
N2
N
(4)
n=0
As N , var
2 0.
2.2
Problem 2.5
Two samples {x[0], x[1]} are independently observed from N 0, 2 distribution. The estimator
1 2
2 =
x [0] + x2 [1]
(5)
2
is unbiased. Find the PDF of
2 to determine if it is symmetric about 2
Solution Consider two standard normal random variables X0 and X1 ,
that is, Xi N (0, 1) , i = 0, 1. Then, by definition, X = X02 + X12 is
2 (n)-distributed with n = 2 degrees of freedom. Its PDF is
1
fX (x) = ex/2
2
x>0
2
X
2
3
from Eq(5)
= 0, = 2 , the PDF of
2 is
y
1
1
1
2 1 y
2
2a
= 2 ey/ = ey/
f 2 (y; ) = fX
= 2
e
a
a
Taking a =
y>0
Its obvious that f 2 (y; ) 6= f 2 (y; ), so the PDF is not symmetric about
= 2 . Note carefully that the PDF is symmetric about , not 2 .
3
3.1
Chapter 3: CRLB
Formulas
(6)
L = L(t) =
ln fX (x; t) =
fX (x; t)
t
t
fX (x; t) t
2
2
= L(t) = ln fX (x; t)
L
t2
t2
(7)
(8)
(9)
E(L)
(10)
1
L = g(t)(h(x) t) t = h(x)
E(L)
(11)
where g() and h() are some functions. Note that the above applies only for
unbiased estimates, so E(t) = t = E[h(x)]. The minimum variance is also
given by,
1
1
var(t) =
=
= g(t) = E(L)
(12)
g(t)
E(L)
Note: t is an estimate of t. Hence, t cannot depend on t itself (if it does,
such an estimate is useless!). So the result t = h(x) intuitively makes sense,
because t depends only on the observed, given data x and not at all on t.
But the mean and variance of t generally do depend on t and that is ok !
For the MVUE case, mean E(t) = t and variance var(t) = g(t) both are
purely functions of t alone.
Replacing the scalar random variable X by a vector of random variables
x, the results still hold.
Facts
Identity, if the regularity condition is satisfied, then
E L 2 = E L
Fisher information I(t) for data x is defined by
I(t) = E(L)
So, the minimum variance is the reciprocal of Fisher information. The
more the information, the lower is the CRLB.
For a deterministic signal s[n; t] with an unknown parameter t in zeromean AWGN w[n] N 0, 2 ,
x[n] = s[n; t] + w[n]
n = 1, 2, . . . , N
2
=
2
k t
sk2
t s[n; t]
For an estimate t of t, if the CRLB is known, then for any transformation = g(t) for some function g() has the new CRLB
CRLB = CRLBt
2
g(t)
t
E
ln p(x; ) = 0
is satisfied (where the expectation is taken w.r.t p(x; )) then the covariance
matrix of any unbiased estimator satisfies
C I1 () 0 var(i ) [I1 ()]i,i
Note: [I1 ()]i,i means first you calculate the whole matrix inverse and
then take the (i, i)th element. The covariance matrix of any vector y is given
by
y = E(y)
Cy = E (y y )(y y )T
Furthermore, an estimator that attains the lower bound,
C = I1 ()
ln p(x; ) = I()(g(x) )
for some M -dimensional function g and some M M matrix I. That estimator, which is the MVUE, is = g(x), and its covariance matrix is
I1 ().
3.2
Problem 3.1
E
ln p(x; ) 6= 0 > 0
E
ln p(x; ) =
ln p(x; ) p(x; ) dx =
p(x; ) dx
(13)
p(xi ; ) dxi = 1
0
p(x; ) dx =
Z
0
Z
p(x1 ; ) dx1
p(xN ; ) dxN
0
7
because the xi s are independent. Note that the limits of the integral depend
on , so we cannot interchange the order of differentiation and integration,
Z
Z
p(xi ; ) dxi
p(xi ; ) dxi 6=
0
0
Hence, the regularity condition fails to hold. In fact, LHS= 1/, but RHS=0!
3.3
Problem 3.3
The data x[n] = Arn + w[n] for n = 0, 1, , N 1 are observed, where w[n]
is WGN with variance 2 and r > 0 is known. Find the CRLB of A. Show
that an efficient estimator exists and find its variance.
Solution
is
N
1
Y
1
1
n 2
p(x; A) =
exp 2 (x[n] Ar )
2
(2 2 )1/2
i=0
!
N 1
1
1 X
n 2
exp 2
(x[n] Ar )
=
2
(2 2 )N/2
= ln p(x; A) = ln 2 2
1
ln p(x; A) = 2
A
N
1
X
N
1
X
N/2
1
2 2
n=0
N
1
X
(x[n] Arn )2
n=0
rn (x[n] Arn )
n=0
(
r
2n
r2N 1
r2 1
n=0
r=
6 1
r=1
is deterministic and known (because both r and N are known), the above
equation simplifies to
!
N 1
1 X n
ln p(x; A) = 2
r x[n] AS
(14)
A
n=0
!
N
1 n
X
S
r
L = 2
x[n] A
(15)
S
n=0
= g(A)(h(x) A)
8
(16)
N
1
X
n=0
rn
x[n]
S
is depends on x but not on A. Hence, from Theorem 3.1, the MVUE estimate
A is
N 1
1 X n
r x[n]
A = h(x) =
S
n=0
2
S
and
CRLB =
1
2
=
g(A)
S
3.4
Problem 3.5
If x[n] = A+w[n] for n = 1, 2, . . . , N are observed, where w = [w[1], w[2], . . . , w[N ]]T
N (0, C), find the CRLB for A. Does an efficient estimator exist and if so,
what is its variance ?
Solution
1
=
ln p(x; A) =
(x Ae)T C1 (x Ae)
A
2 A
T
m
Qm
m =
(x Ae)T = (0
AeT ) = eT
A
A
A
So
a
[1, 1, 1] d
g
(17)
1
= [a + d + g, b + e + h, c + f + i] 1
1
(18)
=a+d+g+b+e+h+c+f +i
(19)
So, denoting = eT C1 e,
eT C1 x
A
ln p(x; A) = g(A)(h(x) A)
A
Hence, there exists a MVUE (the efficient estimator) given by
eT C1 x
eT C1 x
MVUE = A = h(x) =
= T 1
e C e
and its variance is
=
var(A)
1
1
= PN PN
1
i=1
j=1 (C )i,j
10
3.5
Problem 3.9
the CRLB is
var A =
1
2 (2 1)
=
2( 1)
eT C1 e
3.6
Problem 3.13
p1
X
Ak nk + w[n]
k=0
11
Solution:
#2
"
p1
X
1
1
Ak n k
p(x; A) =
exp 2 x[n]
2
2 2
n=0
k=0
"
#2
p1
N
1
X
X
1
1
k
=
x[n]
A
n
exp
2 2
(2 2 )N/2
N
1
Y
n=0
k=0
"
#2
p1
N 1
X
1
1 X
= ln p(x; A) = ln
x[n]
Ak nk
(2 2 )N/2 2 2 n=0
k=0
" (
)
#
p1
N
1
X
1 X
k
i
ln p(x; A) = 0 2
=
2 x[n]
Ak n
0n
Ai
2
n=0
k=0
Because
p1
X
Ak n k =
A1 n1 + A2 n2 + . . . + Ai ni + . . . + AN nN
Ai
Ai
k=0
i
= 0 + 0 + ... +
Ai n + 0
Ai
= ni
Hence, the simplification:
N 1
1 X i
ln p(x; A) = 2
n
Ai
(
x[n]
n=0
2
Aj Ai
ln p(x; A) =
1
2
=
Hence, by definition,
I(A) is given by
(i, j)th
N
1
X
ni 0 nj
n=0
N
1
X
1
2
p1
X
)
Ak n
k=0
ni+j
n=0
N 1
2
1 X i+j
= E
ln p(x; A) = 2
n
Ai Aj
[I(A)]i,j
n=0
for i, j = 0, 1, . . . , p 1. Note that the Fisher information matrix is symmetric, so the order of evaluation of partial derivatives can be interchanged. See
12
pg. 42, Eq (3.22) in the textbook for a special case of the above for p = 2.
Note that for the (0, 0)th entry of the matrix, the above expression gives
N
1
X
i+j
n=0
N
1
X
n=0
Chapter 5
4.1
Problem 5.2
p(x; ) =
N
Y
p(xn ; 2 )
n=1
N
Y
xn
exp(x2n /2 2 )u(xn )
2
n=1
!
!!
N
N
Y
1
1 X 2
=
xn u(xn )
exp 2
xn
2
2
n=1
n=1
= h(x)g(T (x), )
13
T (x) =
x2n
n=1
4.2
Problem 5.5
N
Y
p(xn ; )
n=1
(
1/(2)N
=
0
< xn < , n N
otherwise
1
bool( < xn < , N)
(2)N
But,
xn < = > x1 and > x2 and > xN
= ( > x1 ) ( > x2 ) ( > xn )
= > max{x1 , x2 , . . . , xN }
Similarly,
< xn = > xn
= > max{x1 , x2 , . . . , xN }
14
p(x; ) =
Chapter 7: MLE
ln p(x; t) = 0
t
for t. This equation may have multiple solutions and you should choose the
one appropriately.
Theorem. If an efficient estimator (the estimator which attains CRLB)
exists, then MLE procedure will find it.
The MLE is
asymptotically unbiased i.e., E(t) t as N .
15
16