Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
j=0
h
i
( j)s(k j), i = 1, . . . , M, (1)
where {h
i
(k), k = 0, 1, . . . , L
i
1} and {y
i
(k), k = 0, 1, . . . , L
i
+ N} are the impulse response and output of the i-th channel,
respectively, and {s(k), k = 0, 1, . . . , N} is the common input to the M channels.
Let d = N + L, where L = max{L
1
, L
2
, . . . , L
M
}, the output can be uniquely represented by the (d +1) 1 vector:
y
i
=
_
y
i
(0) y
i
(1) . . . y
i
(L
i
+ N) 0 . . . 0
_
T
, i = 1, . . . , M. (2)
The channel coecients can be similarly represented by (L +1) 1 vectors and the input signal by a (N +1) 1 vector:
h
i
=
_
h
i
(0) h
i
(1) . . . h
i
(L
i
) 0 . . . 0
_
T
, i = 1, . . . , M, (3)
s =
_
s(0) s(1) . . . s(N)
_
T
. (4)
By dening an operator Toepliz() which forms a (W + 1) (D + W + 1) Toeplitz matrix from the (D + 1) 1 vector
a = [a(0) a(1) . . . a(D)]
T
, i.e.,
Toepliz(a, D, W) =
(W +1) rows
we can rewrite (1) into a matrix form:
Y
i
= H
i
S, i = 1, . . . , M, (5)
where
Y
i
= Toepliz( y
i
, d, d),
H
i
= Toepliz(h
i
, L, d),
S = Toepliz(s, N, d + L). (6)
The purpose of blind system identication is to nd {h
i
} from {Y
i
}.
As in [26], it is assumed that the maximal channel order L is known and the channel polynomials {H
i
(z)} do not share
any common roots, where
H
i
(z) =h
i
(0) +h
i
(1)z + +h
i
(L
i
)z
L
i
, i = 1, . . . , M.
Note that, unlike in [24], no assumption about the statistics of the input signal is imposed. The input signal is virtually
arbitrary as long as it is not an all-zero sequence. In addition, the restrictions on the length of the input as in [5,6] are not
required since the data model (5) caters for input signals of arbitrary lengths.
738 W. Qiu et al. / Digital Signal Processing 20 (2010) 736742
3. The proposed method (L
2
method)
Dene the operator Stack() which stacks up M matrices of the same size (A
1
, A
2
, . . . , A
M
) into a new matrix, i.e.,
Stack(A
1
, . . . , A
M
) =
A
1
.
.
.
A
M
.
We now apply subspace principles to nd the following vector of the system impulse responses:
h = Stack(h
1
, h
2
, . . . , h
M
).
From (5), we have
Y = HS,
where Y and H are two Sylvester matrices given by
H = Stack(H
1
, H
2
, . . . , H
M
),
Y = Stack(Y
1
, Y
2
, . . . , Y
M
). (7)
We then form a covariance matrix,
R = Y Y
H
= HS S
H
H
H
(8)
and compute its eigendecomposition:
R = (U
r
U
o
)
_
o
_
(U
r
U
o
)
H
(9)
where
r
contains r non-zero eigenvalues, U
r
and U
o
contain the principal and null eigenvectors, respectively. Note that
when noise is absent, the diagonal matrix
o
= (
r+1
, . . . ,
M(d+1)
) in (9) is an all-zero matrix. Otherwise, its diagonal
entries will be nonzero positive numbers. Although it is impossible here to accurately estimate the noise variance, these
diagonal entries do indicate the noise level and can be represented by
=
1
M(d +1) r
_
M(d+1)
i=r+1
i
. (10)
It has been shown in [2] that H has full column rank since the polynomials {H
i
(z), i = 1, . . . , M} are co-prime and it has
more rows (M(d +1)) than columns (d + L +1). In addition, one can easily verify that, due to the Sylvester structure of S,
the matrix S S
H
is of full rank as long as s is not an all-zero vector. Therefore, r = rank(H) = d + L + 1, and (8)(9) imply
that the columns of H and the columns of U
r
= [u
r+1
u
r+2
. . . u
M(d+1)
] span the same subspace. Due to the orthogonality
between U
r
and U
o
, we have the following relationship:
U
H
o
H =0. (11)
It is known [3] that such an equation involving the full rank generalized Sylvester matrix H and null-space U
o
has a unique
(up to a constant scalar) solution for H (or equivalently h). We now solve it by transforming it to
M(d+1)
i=r+1
_
_
u
H
i
H
_
_
2
2
=0,
which is equivalent to
M(d+1)
i=r+1
_
_
G
H
i
h
_
_
2
2
=0,
where
G
i
= Stack(G
i1
, G
i2
, . . . , G
i M
), i = r +1, . . . , M(d +1), (12)
G
ik
= Toepliz(g
ik
, d, L), k = 1, 2, . . . , M, (13)
g
ik
=
_
u
i
_
(k 1)(d +1)
_
u
i
_
(k 1)(d +1) +1
_
. . . u
i
_
k(d +1) 1
__
T
. (14)
This leads to
h
H
Q h =0,
W. Qiu et al. / Digital Signal Processing 20 (2010) 736742 739
where
Q =
M(d+1)
i=r+1
G
i
G
H
i
. (15)
This means that, under the constraint h
2
= 1, h is the eigenvector of Q corresponding to the zero eigenvalue. When noise
exists, h can be identied (up to a constant scalar) by the least eigenvector of Q . Our proposed method (Proposed_L
2
) for
general (i.e., not necessarily sparse) systems is now summarized as follows:
Step 1. Form the covariance matrix R = Y Y
H
using (6)(7).
Step 2. Compute U
o
the null space of R, form matrix Q using (12)(15), and estimate h (up to a constant scalar) by the
eigenvector of Q corresponding to the smallest eigenvalue.
4. The variant for sparse systems (L
1
method)
In the previous section, the approach to solve (11) is the L
2
-norm optimization which minimizes the quadratic function
x
H
Q x and, therefore, has an ecient closed-form solution. When h is sparse (i.e., most of its coecients are zeros), the
L
2
-norm optimization will almost never nd a sparse solution; producing instead a non-sparse solution with many nonzero
elements [14]. In this situation, the L
1
-norm optimization which has been proven [10,14] to be a powerful tool in producing
sparse solutions can be applied.
Unfortunately, (11) cannot be directly used in L
1
-norm optimization since this will lead to trivial (i.e., all-zero) solutions.
Constraining the solution to h
2
= 1 does not help since this will create a non-convex L
1
-norm optimization problem
and, therefore, convergence is not guaranteed. The challenge here is to build a linear relationship which avoids trivial L
1
solutions between the unknown (h) and the available data { y
i
}. In fact, it can be shown that h and the data vector
y = Stack( y
1
, y
2
, . . . , y
M
)
are related by a matrix dened by the input s dened in (4). This is because that (5) can be re-written as
y
i
= Sh
i
, i = 1, 2, . . . , M,
where
S =
_
Toepliz(s, N, L)
_
T
. (16)
Therefore, we have
y = Sh (17)
where S is a block diagonal matrix dened by S as
S =
S 0 0 . . . 0
0 S 0 . . . 0
.
.
.
.
.
.
0 0 0 S
. (18)
Now we need to have an estimate of s so that S can be constructed. Going back to the original data model (1), it is obvious
that the polynomial dened by the input signal s:
S(z) = s(0) + s(1)z + + s(N)z
N
is actually the greatest common division (GCD) of the polynomials dened by the system outputs { y
i
}:
Y
i
(z) = y
i
(0) + y
i
(1)z + + y
i
(L
i
+ N)z
L
i
+N
, i = 1, . . . , M.
Therefore, S(z) (or equivalently s) can be found from {Y
i
(z)} using a GCD estimation algorithm. We will use the algorithm
in [15] for this purpose.
Since a GCD scaled by any nonzero constant is also a GCD, s can only be estimated in this way up to a constant scalar.
With the estimated s, the reformulated data model (17) allows h to be obtained (also up to a constant scalar) by L
1
-
minimization. Note that this is strict convex optimization and convergence is guaranteed. The minimization generally takes
the following form [16]:
h = argmin
x
_
x
1
+ y Sx
2
2
_
(19)
where balances the sparsity of the solution with the delity to the data, and should be inversely proportional to the noise
level. That is, the more accurate the data is, the bigger weight it should have in the optimization objective function. The
value of needs to be tuned for each particular application and some research towards determining analytically [17] is
currently going on. In the proposed method we set = / where is a positive constant and is given by (10).
We are now ready to summarize the Proposed_L
1
method for identifying systems with sparse impulse responses:
740 W. Qiu et al. / Digital Signal Processing 20 (2010) 736742
Fig. 1. NRMSE vs. SNR, non-sparse systems.
Step 1. Use the method in [15] to estimate s.
Step 2. Form matrix S using estimated s according to (16) and (18), and obtain the estimate of h (up to a constant scalar)
through L
1
-norm minimization (19).
5. Simulation
We evaluate the performances of the proposed methods against that of the LS method in [5]. Although the LS method
was primarily developed for continuous input sequences, it can also be applied for impulsive inputs certifying certain length
constraints. Note that here we are not intending to have a comprehensive performance comparison between the proposed
methods with the LS method since they are targeting different applications. What we would like to conrm is that our
proposed methods provide meaningful results and they are capable of dealing with arbitrarily short input signals.
In the simulation, the input consists of QPSK symbols and the three-ray multi-path channel system with M = 4 and
L = 5 in [5, Table III], is used and modied as follows to generate channels with desired sparsity. The sparsity parameter K
is introduced to specify the number of nonzero elements in each channel vector h
i
. In order to keep the order of channels
to be L, h
i
(0) and h
i
(L) are kept unchanged. The other K 2 nonzero indexes are uniformly chosen from {1, . . . , L 1} and
the remaining channel coecients are set to zero. This process is done independently for each channel.
For each test, one realization of the input s and channels with desired sparsity is generated and 100 runs are then
conducted where, in each run, an independent noise vector w
i
is added to the output y
i
(i = 1, . . . , M), with w
i
containing
i.i.d. Gaussian variables with zero mean and variance
2
i
.
The SNR for each channel is dened as
SNR
i
= 10log
10
y
i
2
2
(d +1)
2
i
, i = 1, . . . , M,
and chosen to be the same for all the channels. The normalized root-mean-square-error for the j-th test is measured by
NRMSE
j
=
1
h
2
_
1
100
100
i=1
_
_
h
o
i
h
_
_
2
2
where h
o
i
denotes the normalized estimate in the i-th run. The normalization is to remove the effect of the arbitrary
constant scalar on the error evaluation and is done by xing one element of the estimate
h
i
to the true value. That is,
h
o
i
=
h
i
(h(l)/
h
i
(l)), where h(l) = 0, l {0, 1, . . . , M(L +1)}. The nal NRMSE is the averaged NRMSE
j
over 50 independent
tests. In the simulation, = 0.1 is xed for all the scenarios and the convex programming package from [18] is used in
solving (19).
Figs. 1 and 2 show the cases when the system is not sparse (K = 6). In Fig. 1 the length of input signal is xed to N = 15
while SNR varies from 25 dB to 45 dB. The results show how the three methods improve while SNR increases. Also shown
W. Qiu et al. / Digital Signal Processing 20 (2010) 736742 741
Fig. 2. NRMSE vs. N, non-sparse systems.
Fig. 3. NRMSE vs. sparsity.
here is that, in this non-sparse situation, the performances of the proposed methods (Proposed_L
2
and Proposed_L
1
) are
barely distinguishable.
Fig. 2 shows the case when SNR = 30 dB while N varies from 1 to 35. The LS method is not applicable when N < L +1. It
can be seen that the proposed methods deal with short inputs well and their performances remain relatively constant when
N increases. The LS method is shown to improve signicantly while N increases and outperform the proposed methods
when N 20 (i.e., N 4L). Once again, Proposed_L
2
and Proposed_L
1
are shown to perform very closely for non-sparse
systems.
Fig. 3 shows the case when SNR = 30 dB and N = 15 while K varies from 2 to 6. It can be seen that when h is
relatively sparse (i.e., smaller K), Proposed_L
1
(proposed specically for sparse systems) performs better than Proposed_L
2
.
When the system impulse responses become dense (i.e., K is close to L + 1), the two methods give virtually identical
performances.
742 W. Qiu et al. / Digital Signal Processing 20 (2010) 736742
6. Conclusion
We have proposed a deterministic method for blind identication of multichannel systems. A variant of the proposed
method is also presented which utilizes the L
1
-norm convex optimization to improve the estimation accuracy when the
system to be identied has sparse impulse responses. The capability of the proposed methods of dealing with impulsive
input signals of arbitrarily short lengths has been demonstrated. The proposed methods, therefore, have the potential to
improve the performances of communication, exploration seismology and mechanical signature analysis systems.
References
[1] K.A. Meraim, W. Qiu, Y. Hua, Blind system identication, Proc. IEEE 85 (8) (1997) 13101322.
[2] L. Tong, G. Xu, T. Kailath, Blind identication and equalization based on second-order statistics: A time domain approach, IEEE Trans. Inform. The-
ory 40 (2) (1994) 340349.
[3] E. Moulines, P. Duhamel, J.-F. Cardoso, S. Mayrargue, Subspace methods for the blind identication of multichannel FIR lters, IEEE Trans. Signal
Process. 43 (2) (1995) 516525.
[4] J. Fang, A.R. Leyman, Y.H. Chew, Blind equalization of SIMO FIR channels driven by colored signals with unknown statistics, IEEE Trans. Signal Pro-
cess. 54 (6) (2006) 19982008.
[5] G. Xu, H. Liu, L. Tong, T. Kailath, A least-squares approach to blind channel identication, IEEE Trans. Signal Process. 43 (12) (1995) 29822993.
[6] Y. Hua, Fast maximum likelihood for blind identication of multiple FIR channels, IEEE Trans. Signal Process. 44 (3) (1996) 661672.
[7] S. Braun (Ed.), Mechanical Signature Analysis: Theory and Applications, Academic Press, New York, 1986.
[8] L.C. Wood, S. Treitel, Seismic signal processing, Proc. IEEE 63 (1975) 649661.
[9] T.S. Rappaport, Wireless Communications: Principles and Practice, Prentice Hall PTR, 1996.
[10] E. Cands, M. Wakin, An introduction to compressive sampling, IEEE Signal Process. Mag. 25 (2) (2008) 2130.
[11] W.F. Schreiber, Advanced television systems for terrestrial broadcasting: Some problems and some proposed solutions, Proc. IEEE 83 (1995) 958981.
[12] S. Ariyavisitakul, N.R. Sollenberger, L.J. Greenstein, Tap selectable decision-feedback equalization, IEEE Trans. Commun. 45 (1997) 14971500.
[13] M. Kocic, et al., Sparse equalization for real-time digital underwater acoustic communications, in: Proc. OCEANS95, San Diego, CA, 1995, pp. 14171422.
[14] R. Baraniuk, Compressed sensing, IEEE Signal Process. Mag. 24 (4) (2007) 118121.
[15] W. Qiu, Y. Hua, K.A. Meraim, A subspace method for the computation of the GCD of polynomials, Automatica 33 (4) (1997) 741743.
[16] E. Cands, T. Tao, The Dantzig selector: Statistical estimation when p is much larger than n, Ann. Stat. 35 (2005) 23132351.
[17] E.T. Hale, W. Yin, Y. Zhang, A xed-point continuation method for
1
-regularized minimization with applications to compressed sensing, CAAM tech-
nical report TR07-07, Rice University, July 2007.
[18] M. Grant, S. Boyd, CVX: Matlab software for disciplined convex programming (web page and software), http://stanford.edu/~boyd/cvx, February 2009.
Wanzhi Qiu received his B.E. and M.E. from University of Electronic Science and Technology of China and Ph.D. from the University
of Melbourne, all in Electrical and Electronic Engineering. He then held senior research and engineering positions with Motorola, NEC
and Bandspeed, working on GSM, WCDMA, Bluetooth, ADSL and 3G mobile projects. He is currently a senior researcher with National
ICT Australia (NICTA). His research has primarily been in the areas of blind system identication, wireless sensor networks and signal
processing.
Syed Khusro Saleem received his B.S.E.E. from Northwestern University in 1992. He received his M.Eng.Sc. and Ph.D. in 1994 and
1997, respectively, both in Electrical and Electronic Engineering from the University of Melbourne, Australia. Dr. Saleem has had various
roles ranging from Research Scientist to Project Leader at the Australian Defence Science & Technology Organisation, NEC Australia and
Bandspeed. He is currently employed by National ICT Australia Ltd as a Project Leader where he oversees the Water Information Net-
works research program. Dr. Saleems interests range from software engineering and digital design to large scale system engineering and
management.
Minh Pham received the Bachelor and the Ph.D. degrees in Electrical Engineering from the University of Melbourne in 1994 and 2000,
respectively. He had worked in industry and academia organizations including the Australias Commonwealth Scientic and Industrial
Research Organisation (CSIRO). He was a research fellow at the University of Lancaster in the United Kingdom and the University of
Melbourne. He is presently a senior researcher at the National ICT Australia. His current research interests are in the following areas:
sensor network technology, electromagnetic eld theory, and digital signal processing.