Documenti di Didattica
Documenti di Professioni
Documenti di Cultura
Nicolae Cleju
Faculty of Electronics, Telecommunications and Information Technology
Gheorghe Asachi Technical University of Iasi, Romania
and
Institute of Computer Science, Romanian Academy, Iasi, Romania
Email: {nikcleju}@etti.tuiasi.ro
AbstractThis paper presents a novel algorithm for sparse algorithm in this paper is based on a similar greedy approach,
signal recovery, named Least Squares Pursuit, based on least- but with different prioritization of the internal constraints
squares minimization followed by choosing the atom with the compared to OMP.
largest resulting coefficient, in a greedy one-by-one fashion. It
follows a similar approach to that of Orthogonal Matching This paper is organized as follows. Section II presents
Pursuit, but with different prioritization of the constraints. We the Least Squares Pursuit algorithm and its relation with
propose an efficient implementation for Least Squares Pursuit, Orthogonal Matching Pursuit, and Section III presents an
derive theoretical guarantees for signal recovery, and finally efficient implementation procedure. Guarantees for successful
present promising simulation results. recovery are derived in Section IV, and simulation results
I. I NTRODUCTION are presented in Section V. Finally, conclusions are drawn in
Section VI.
Sparse signal recovery and its applications has been an
active research field in signal processing for more than a II. T HE L EAST S QUARES P URSUIT ALGORITHM
decade. It is now well known [1] that a signal x which has A. Background
a sparse decomposition ? in some basis or overcomplete
dictionary Directly solving (eq. 2) is not feasible because the `0 norm
x = D ? , with k? k0 = k, is making the problem NP-hard. The class of greedy pursuit
algorithms, featuring well-known algorithms like Matching
can be recovered from a reduced set of linear measurements Pursuit and Orthogonal Matching Pursuit, try to estimate
forming an acquisition matrix M , possibly contaminated with the support in a one-by-one fashion. Specifically, Orthogonal
noise z of energy = kzk22 : Matching Pursuit (OMP) selects a new atom at step k as the
atom number i most correlated with the current residual:
y = M x + z.
i = arg max kDT r(k) | = kDT (y D (k) )k,
The recovery requires solving the NP-complete optimization
problem and then updates the current solution estimate (k) by pro-
jecting y onto the set of all selected atoms until this moment,
x = D arg min kk0 with ky M Dk22 < . (1)
T (k) . In other words, y is approximated as best as possible
This is known as compressed sensing [2], and it has been using only the atoms from T (k) , while the remaining atoms,
(k)
extensively researched in the last 10 - 15 years. which make up the complement set Tc , are not used at all.
In the noiseless case, which we consider in this paper, = 0 This can be expressed as follows:
and the problem reduces to that of finding the sparsest soution (k) = arg min ky Dk22 subject to T (k) = 0, (3)
(in the `0 sense) of an undetermined equation system. We c
therefore consider our canonical problem defined as: where T (k) indicates the restriction of to the indices in
c
(k)
arg min kk0 subject to y = D. (2) Tc .
The idea of our algorithm is to swap the data-fidelity term
where, for brevity we renamed the product M D from (eq. 1) ky Dk22 with sparsity term T (k) = 0 in (eq. 3). Note
c
as simply D. This paper presents a novel greedy algorithm for that neither of the two terms is a priori more important than
solving this problem. the other, since the true solution ? with the true support
The optimization problem (eq. 2) is known to be NP-hard, T satisfies both terms perfectly. Their relative importance
therefore one must settle with suboptimal solutions. One of the is just a choice of algorithm design, and OMP chooses to
most common algorithms for (approximately) solving (eq. 2) strictly enforce the sparsity term T (k) = 0, while accepting
c
is Orthogonal Matching Pursuit [3][4] (OMP). OMP iteratively approximate data-fidelity and working to improve it with each
estimates the support of the solution by progressively adding, new atom. We propose to take the opposite route: always
at every iteration, the most likely atom, i.e. the one most enforce perfect data-fidelity, ky Dk22 = 0, while accepting
similar to the current approximation residual. Our proposed non-exact sparsity T (k) , working to improve it with each new
c
Algorithm 1 Least Squares Pursuit The name Least Squares Pursuit (LSP) is justified by the fact
1: (0)
Initialize: T ,
(0)
{1, 2, ...N }
Tc that the algorithm follows the same one-by-one atom selection
2: repeat as OMP (hence the name Pursuit), but the orthogonality
3: Update current solution: induced by OMPs projection step is replaced with a least
squares problem.
(k) = arg min kIT (k) k22 subject to y = D
c
III. E FFICIENT IMPLEMENTATION
4: Update current support: A naive implementation would solve (eq. 4) anew for (k) at
i = arg max kT (k) k every iteration. In the following, we describe a more efficient
c
implementation that starts from the solution at the previous
(k+1) (k)
T =T {i} step (k 1) and updates it. This is analogue to the update
step in OMP.
Tc(k+1) = Tc(k) \ {i}
Consider the initial least-squares solution:
5: until stop criterion
(0) = arg min kk22 subject to y = D.
Since T (0) is empty, this is the full least-squares solution of
selected atom. Thus our update step, given the estimate support the system, and thus it can be represented compactly using the
T (k) at step k, is: Moore-Penrose pseudoinverse of D, denoted here as D :
(k) = arg min T (k) subject to y = D. (4) (0) = D y
c
of (eq. 6). Moreover, it is embedded in the row space of Considering that there were no wrong atom choices until the
the original dictionary D, since the columns of D and current iteration k, the original co-support Tc is completely
(k) (k)
the rows of D always span the same subspace. inside Tc . Therefore, we can split the set Tc into the
(k)
These considerations show that the subspace spanned by remaining support we still want to find, T? , and the whole
(k)
D DT is the additional null space brought by the decreasing true co-support Tc . Splitting the previous equation in this
(k) manner leads to:
range of Pc in addition to the original null space of D.
How does (k) differ from the previous estimate (k1) ? (k) (k)
NTc Tc = NT (k) (k) . (8)
They are both least-squares solutions of their respective ? T?
systems (eq. 6), and thus orthogonal to their null spaces The difference between the true sparse solution ? and
(k1) (k) (k) (k1)
D DT and D DT . But DT consists of DT together (k)
lives in the null space, since both solutions produce y.
with the last selected atom, henceforth denoted dk , and there- Therefore
fore we can obtain (k) by merely projecting and removing (k) ? = N T v,
from (k1) the part which lays in this additional null space
spanned by D dk : for some vector v. Restricting the above only to the rows in
the true co-support Tc , and considering that ? is zero for
(k) = (I D dk {D dk } ) (k1) . these atoms, leaves
Since dk is a vector, D dk is also a vector, and in this case (k)
Tc = NTTc v,
its pseudoinverse {D dk } is simply the transposed vector
(k)
divided by its squared norm. It follows that: therefore Tc lives in the span of NTTc . As such, it holds that:
1 T
(k) = (k1) T
D dk dTk D (k1) . (k)
Tc = NTc NTc Tc .
(k)
dTk D D dk
Replacing eq. 8 in the above leads to:
We can use the above equation for updating the estimate
(k) (k)
solution after every new atom is added. Tc = NTc NT (k) (k) .
? T?
Alternatively, we can consider the whole set DT (k) of atoms
selected until step k, and compute the solution at time k Thus, having
starting every time from the initial solution (0) = D y: kNTc NT (k) k, < 1
?