Sparse Reconstruction with Compressed Sensing

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search

Author: Ngoc Ly (SysEn 5800 Fall 2021)

Compressed Sensing (CS)

Compressed Sensing summary here

Compression is synonymous with sparsity. So when we talk about compression we are actually referring to the sparsity.

Introduction

$ x \in \mathbb{R}^N $often not really sparse but approximately sparse

$ \Phi \in \mathbb{R}^{M \times N} $for i$ M \ll N $s a Random Gaussian or Bernoulli matrix

$ y \in \mathbb{R}^M $are the observed y samples

$ e \in \mathbb{R}^M $noise vector $ \| e \|_2 \leq \eta $

s.t.

How can we reconstruct x from $ y = \Phi x + e $

The goal is to reconstruct $ x \in \mathbb{R}^N $given $ y $ and $ \Phi $

Sensing matrix $ \Phi $must satisfy RIP i.e. Random Gaussian or Bernoulli matrixies satisfies (cite)

let $ [ N ] = \{ 1, \dots , N \} $be an index set $ [N] $ enumerates the columns of $ \Phi $ and $ x $ $ \Phi $ is an under determined systems with infinite solutions since $ M \ll N $. Why $ l_2 $ norm does not work


The problem formulation is to recover sparse data $ \mathbf{x} \in \mathbb{R}^N $


The support of $ \mathbf{x} $ is $ supp(\mathbf{x}) = \{i \in [N] : \mathbf{x}_i \neq 0 \} $ we say $ \mathbf{x} $ is $ k $ sparse when $ |supp(x)| \leq k $

We are interested in the smallest $ supp(x) $ , i.e. $ min(supp(x)) $


Before we get into RIP lets talk about RIC

Restricted Isometry Constant (RIC) is the smallest $ \delta_{|s} \ s.t. \ s \subseteq [N] $that satisfies the RIP condition introduced by Candes, Tao

Random Gaussian and Bernoulli satisfies RIP

RIP defined as

$ (1 - \delta_s) \| x \|_2 ^2 \leq \| \Phi x \|_2^2 \leq (1 + \delta_s) \| x \|_2 ^2 $

Let $ \Phi \in \mathbb{R}^{M \times N} $ satisfy RIP, Let $ [N] $ be an index set For $ s $ is a restriction on $ \mathbf{x} $ denoted by $ x_{|s} $ $ x \in \mathbb{R}^N $ to $ s $ k-sparse $ \mathbf{x} $ s.t. RIP is satisfied the $ s = supp(\mathbf{x}) $ i.e. $ s \subseteq [N] $and $ \Phi_{|s} \subseteq \Phi $ where the columns of $ \Phi_{|s} $ is indexed by $ i \in S $

In search for a unique solution we have the following $ l_0 = |supp(x)| $ optimization problem. $ \mathbf{\hat{s}} = \underset{s}{arg min} \| \mathbf{s}\|_0 \quad s.t. \quad \mathbf{y} = \Phi \mathbf{s} $, which is an NP-Hard.

From Results of Candes, Romberg, Tao, and Donoho

If $ \Phi $ satisfies RIP and $ \mathbf{y} $ is sparse the $ l_0 $ has a unique solution. The equivalent $ l_1 $ convex program to the $ l_0 $ program. $ \mathbf{\hat{s}} = \underset{s}{arg min} \| \mathbf{s}\|_1 \quad s.t. \quad \mathbf{y} = \Phi \mathbf{s} $


x i , i f i ∈ S

( x | S ) i =

0 otherwise

RIP defined as

( 1 − δ s )k x k 22 ≤ k Φx k 22 ≤ ( 1 + δ s )k x k 22

3 Lemmas Page 267 Blumensath Davies IHT for CS

Lemma 1(Blumensath, Davis 2009 Iterative hard thresholding for compressed

sensing), For all index sets Γ and all Φ for which RIP holds with s = | Γ | that is

s = supp ( x )

1k Φ Γ T k 2 ≤

q

1 + δ | Γ | k y k 2

( 1 − δ | Γ | )k x Γ k 22 ≤ k Φ Γ T Φ Γ x Γ k 22 ≤ ( 1 + δ | Γ | )k x Γ k 22

and

k( I − Φ Γ T Φ Γ )k 2 ≤ δ | Γ | k x Γ k 2

SupposeΓ ∩ Λ = ∅

k Φ Γ T Φ Λ ) x Λ k 2 ≤ δ s k x Λ k 2

Lemma 2 (Needell Tropp, Prop 3.5 in CoSaMP: Iterative signal recovery

from incomplete and inaccurate √ samples)

If Φ satisfies RIP k Φx s k 2 ≤ 1 + δ s k x s k 2 , ∀ x s : k x s k 0 ≤ s, Then ∀ x

k Φx k 2 ≤

p

1 + δ s k x k 2 +

p

1 + δ s

k x k 1

sqrts

Lemma 3 (Needell Tropp, Prop 3.5 in CoSaMP: Iterative signal recovery

from incomplete and inaccurate samples)

Let x s be the best s-term approximation to x. Let x r = x − x s Let

y = Φx + e = Φx s + Φx r + e = Φx s + ẽ

If Φ satisfies RIP for sparsity s, then the norm of error ẽ is bounded by

k ẽ k 2 ≤

p

1 + δ s k x − x s k 2 +

p

1 + δ s

k x − x s k 1

+ k e k 2

s

∀ x

Theory

Verification of the Sensing matrix

Gel’fand n-width

Errors E ( S, Φ, D )

Definition Mutual Coherence

Let $ \Phi \in R^{M \times N} $, the mutual coherence $ \mu_\Phi $ is defined by:</math>

$ \mu_{\Phi} = \underset{i \neq j} {\frac{| \langle a_i, a_j \rangle |}{ \| a_i \| \| a_j \|}} $


We want a small $ \mu_{\Phi} $ because it will be close to the normal matrix, which satisfies RIP


Algorithm IHT

$ z_v^{(n)} = \nabla f_v(x^{(n)}) = - \Phi_v^T( \mathbf{y} - \Phi \mathbf{x}) $ Then $ x^{n+1} = \mathcal{H}\left( \mathbf{x}^{(n)} - \tau \sum_{j \in N}^{N} z_v^{(n)}\right) $

Define the threashholding operators as: $ \mathcal{H}_s[\mathbf{x}] = \underset{z \in \sum_s}{argmin} \| x - \Phi \mathbf{x}\|_2 $ selects the best-k term approximation for some k


Stopping criterion is $ \| y - \Phi \mathbf{x}^{(n)}\|_2 \leq \epsilon $

  • Initialize $ \Phi, \mathbf{y}, \mathbf{e} \ \mbox{with} \ \mathbf{y} = \mathbf{\Phi} \mathbf{x} | \mathbf{e} and \mathfrak{M} $
  • output $ IHT(\mathbf{y}, \mathbf{\Phi}, \mathfrak{M}) $
  • Set $ x^{(0)} = \mathbf{0} $
  • While halting criterion false do
    • $ x^{(n+1)} \leftarrow \mathcal{H}_{s} \left[ x^{(n) + \Phi^T (\mathbf{y} - \mathbf{\Phi x}^{(n)})}\right]} $
    • $ n \leftarrow n + 1 $
    • end while
  • return: $ IHT(\mathbf{y}, \mathbf{\Phi}, \mathfrak{M}) \leftarrow \mathbf{x}^{(n)} $

Numerical Example

Check if $ \Phi $ satisfies RIP


Iterative Hard Thresholding IHT

Applications

Distributed Systems peer-to-peer network

Collaborative sensor networks for energy savings.

Distributed Systems where Energy needs to be preserved.

Conclusion

References

[1]


[2]

[3]

[4]


[5]


[6]


[7]

[8]

[9]

  1. 2. Emmanuel J. Candès and Terence Tao. Decoding by linear programming. IEEE Trans. Inf. Theory, 51(12):4203–4215, 2005.
  2. 5. Stephen A. Vavasis. Elementary proof of the spherical section property for random matrices. Univer-sity of Waterloo, Waterloo,Technical report, 2009.
  3. 6. Angshul Majumdar. Compressed sensing for engineers. Devices, circuits, and systems. CRC Press, Taylor & Francis Group, Boca Raton, FL, 2019. Includes bibliographical references and index.
  4. 7. Simon Foucart and Holger Rauhut. A mathematical introduction to compressive sens- ing. Applied and numerical harmonic analysis. Birkhäuser, New York [u.a.], 2013.
  5. 8. D. L. Donoho. Compressed sensing. 52:1289–1306, 2006.
  6. 12. E. J. Candes, J. Romberg, and T. Tao. Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information. 52:489–509, 2006.
  7. 16. Giulio Coluccia, Chiara Ravazzi, and Enrico Magli. Compressed sensing for dis- tributed systems, 2015.
  8. 17. Mohammed Rostami. Compressed sensing with side information on feasible re- gion, 2013.
  9. 18. Thomas Blumensath and Mike E. Davies. Iterative hard thresholding for com- pressed sensing. May 2008.