Sparse Reconstruction with Compressed Sensing: Difference between revisions
Molokaicat (talk | contribs) No edit summary |
Molokaicat (talk | contribs) No edit summary |
||
Line 171: | Line 171: | ||
* output <math>IHT(\mathbf{y}, \mathbf{\Phi}, \mathfrak{M}) </math> | * output <math>IHT(\mathbf{y}, \mathbf{\Phi}, \mathfrak{M}) </math> | ||
* While halting criterion false do | |||
* <math>x^{(n+1)} \leftarrow \P_{\mathfrak{M} \left[ x^{(n) + \Phi^* (\mathbf{y} - \mathbf{\Phi x}^{(n)})}\right]}</math> | |||
* <math>n \leftarrow n + 1 </math> | |||
end while | |||
return: <math>IHT(\mathbf{y}, \mathbf{\Phi}, \mathfrak{M}) \leftarrow \mathbf{x}^{(n)}</math> | |||
Revision as of 03:41, 28 November 2021
Author: Ngoc Ly (SysEn 5800 Fall 2021)
Introduction
often not really sparse but aproximently sparse
for is a Random Gaussian or Bernoulli matrix
are the observed y samples
noise vector k e k 2 ≤ η
s.t.
How can we reconstruct x from
The goal is to reconstruct x 0 ∈ R N given and
Sensing matrix must satisfy RIP i.e. Random Gaussian or Bernoulli ma-
traxies satsifies (cite)
let be an index set
enumerates the colums of and
is an underdetermined systems with infinite solutions. Why norm does
not work
What is compression is sunomunus with to the sparcity.
The problem formulation is to recover sparse data x 0 ∈ R N
The support of is
we say is sparse when
We are interested in the smallest , i.e.
Restricted isometry property (RIP) Introduced by Candes, Tao
Random Gausian and Bernoulli satsifies RIP
Let Φ ∈ R M × N satsify RIP, Let be an index set
Before we get into RIP lets talk about RIC
Resiricted Isometry Constant (RIC) is the smallest δ | s s.t. s ⊆ [ N ] that satsi-
fies the RIP condition
For is a restrciton on denoted by to k-sparse s.t. RIP is
satisified the s = | Γ | i.e. s ⊆ [ N ] and Φ | s ⊆ Φ where the columns of Φ | s is
indexed by i ∈ S
�
x i , i f i ∈ S
( x | S ) i =
0 otherwise
RIP defined as
( 1 − δ s )k x k 22 ≤ k Φx k 22 ≤ ( 1 + δ s )k x k 22
3 Lemmas Page 267 Blumensath Davies IHT for CS
Lemma 1(Blumensath, Davis 2009 Iterative hard thresholding for compressed
sensing), For all index sets Γ and all Φ for which RIP holds with s = | Γ | that is
s = supp ( x )
1k Φ Γ T k 2 ≤
q
1 + δ | Γ | k y k 2
( 1 − δ | Γ | )k x Γ k 22 ≤ k Φ Γ T Φ Γ x Γ k 22 ≤ ( 1 + δ | Γ | )k x Γ k 22
and
k( I − Φ Γ T Φ Γ )k 2 ≤ δ | Γ | k x Γ k 2
SupposeΓ ∩ Λ = ∅
k Φ Γ T Φ Λ ) x Λ k 2 ≤ δ s k x Λ k 2
Lemma 2 (Needell Tropp, Prop 3.5 in CoSaMP: Iterative signal recovery
from incomplete and inaccurate √ samples)
If Φ satisfies RIP k Φx s k 2 ≤ 1 + δ s k x s k 2 , ∀ x s : k x s k 0 ≤ s, Then ∀ x
k Φx k 2 ≤
p
1 + δ s k x k 2 +
p
1 + δ s
k x k 1
sqrts
Lemma 3 (Needell Tropp, Prop 3.5 in CoSaMP: Iterative signal recovery
from incomplete and inaccurate samples)
Let x s be the best s-term approximation to x. Let x r = x − x s Let
y = Φx + e = Φx s + Φx r + e = Φx s + ẽ
If Φ statisfies RIP for sparcity s, then the norm of error ẽ is bounded by
k ẽ k 2 ≤
p
1 + δ s k x − x s k 2 +
p
1 + δ s
k x − x s k 1
√
+ k e k 2
s
∀ x
Theory
Gel’fend n-width
Errors E ( S, Φ, D )
Definition Mutual Coherence
LetA ∈ R M × N , themutualcoherenceµ A isde f inedby :
µ A =
|h a i , a j i|
k a i kk a j k
i 6 = j
We want a small µ A because it will be close ot the normal matrix, which
will satisfie RIP
Algorithm IHT
- Initalize
- output
- While halting criterion false do
end while return:
Numerical Example
Basis Persuit
Iterative Hard Thresholding IHT
Solver
2