Sparse Reconstruction with Compressed Sensing: Difference between revisions
Molokaicat (talk | contribs) |
Molokaicat (talk | contribs) |
||
Line 320: | Line 320: | ||
Check if error is less then <math>10 ^{-5} </math> | Check if error is less then <math>10 ^{-5} </math> | ||
[[File:Reconstruction plot.png|thumb|right]] | |||
===Stopping condition meat=== | ===Stopping condition meat=== | ||
Revision as of 09:43, 20 December 2021
Author: Ngoc Ly (SysEn 5800 Fall 2021)
Compressed Sensing (CS)
Compressed Sensing summary here
Compression is synonymous with sparsity. So when we talk about compression we are actually referring to the sparsity. We introduce Compressed Sensing and then focus on reconstruction.
Introduction
sub module goal
The goal of compressed sensing is to interact with an underdetermined linear system in which the number of variables is much greater than the number of observations, resulting in an infinite number of signal coefficient vectors for the same set of compressive measurements . As a result, additional information is necessary for to recover from . The objective is to reconstruct a vector in a given of measurements and a sensing matrix A. Instead of taking a large number of high-resolution measurements and discarding the majority of them, consider taking way fewer random measurements and reconstructing the original with high probability from its sparse representation.
sub modual
Begin with a linear equation , where is a sensing matrix that must be obtained and will result in either exact or approximated optimum solution depending on how it is chosen, is a signal vector with at most -sparse entries, which means has non-zero entries, be an index set, is a compressed measurement vector, , is a noise vector and assumed to be bounded if it exists, and .
The goal of compressed sensing is to being with the under determined linear system
, Where
for
How can we reconstruct x from
The goal is to reconstruct given and
Considerably fewer random measurements and reconstruct the original with high probability from its sparse representation instead of taking a large number of high-resolution measurements and discarding the majority of them. being a random matrix
let be an index set enumerates the columns of and . is an under determined systems with infinite solutions since . Why norm won't give sparse solutions, where asl norm will return a sparse solution.
Notation =
often not really sparse but approximately sparse
for Sensing matrix a Random Gaussian or Bernoulli matrix
are the observed y samples
noise vector
put defn of p norm here
where is the sparsifying matrix and are coeficients
sub module sparsity
A vector is said to be -sparse in if it has at most nonzero coefficients. The support of is , and is a -sparse signal when . The set of -sparse vectors is denoted by . Consequently, there are different subsets of -sparse vectors. If a random -sparse is drawn uniformly from , its entropy is approximately equivalent to bits are required for compression of ~cite(Measurements vs Bits).
The idea is to search for the sparsest from the measurement vector and a sensing matrix with . If the number of linear measurements is at least twice as its sparsity , i.e., , then there exists at most one signal that satisfies the constraint and produce the correct result for any [coluccia2015 book7]. Hence, the reconstruction problem can be formulated as an "norm" program.
This optimization problem minimizes the number of nonzero entries of subject to the constraint , that is to find the sparsest element in the affine space [2019 book33]. It turns out to be a combinatorial optimization problem, which is NP-Hard because it includes all possible sets of -sparse out of . Furthermore, if noise is present, the recovery is unstable [Buraniuk "compressed sensing"].
In other words, only the smallest support set of is of interest, i.e., .
The goal is to search for the sparsest given the meassurment and the constraint matrix .
This is antiquated to find the in the set
This searching problem can be formulated to the following program
Another words we are interested in the smallest , i.e.
sub module
Let satisfy RIP, Let be an index set For is a restriction on denoted by to k-sparse s.t. RIP is satisfied the i.e. and where the columns of is indexed by
In search for a unique solution we have the following optimization problem.
sub module zero norm program
, which is an combinatorial NP-Hard problem. Hence, if noise is presence the recovery is not stable. [Buraniuk "compressed sensing"]
Restricted Isometry Property (RIP)
A matrix is said to satisfy the RIP of order if for all has a . A restricted isometry constant (RIC) of is the smallest satisfying this condition [2019 book38, coluccia2015 book7].
Under projections through matrix , the restricted isometry property allows -sparse vectors to have unique measurement vectors . If meets RIP, then does not send two distinct -sparse to the same measurement vector , indicating that is a unique solution under RIP.
If the matrix satisfies the RIP condition of order and the constant , there are two distinct -sparse vectors in . When they are equal, the restricted isometry property holds. If is a -order RIP matrix, it means that no two -sparse vectors are mapped to the same measurement vector by . In other words, when working with sparse vectors, the RIP ensures that the columns of are nearly orthonormal. Furthermore, is an approximately norm-preserving function, which means that it preserves its distance when mapping for -sparse signals for all or more as approaches zero. [Candes, Romberg, Tao[4]] demonstrate that if is -sparse, and satisfies the RIP of order with RIP-constant , then gives a unique sparse solution. The convex optimization problem is the same as the solution to the program and can solve by the Linear Program [2019 book38, coluccia2015 book7]. Hence, the reconstruction problem is as followed.
If satisfies RIP then doesn't send two distinct k-sparse to the same measurment vector . Anothers words is a unique solution under RIP.
RIP defined as
satisfies RIP of order if for satisfies for inequalaty
- TODO switch s to k
sub module RIP matracies
If a sensing matrix must satisfy RIP, then the number of measurements is required recover with high probability.
sub module RIC
Restricted Isometry Constant (RIC) is the smallest in
if i.e. twise the sparsity, then there exists an unique such that .
for is k-sparse and and satisfies RIP of order RIC then the program can have to relaxed convex form program.
If satisfies RIP and is sparse the gives sparse solutions and is a unique. It is equivalent to the following convex optimization problem and can solve by Linear Program.
sub problem 1 norm program
From Results of Candes, Romberg, Tao, and Donoho
Theory
Two things need to be considered when recovering
- (1) The design of the sensing matrix
- (2) The recovery algorithm
Sensing Matrix
Although the sensing matrix satisfies RIP of order in some situations, confirming a given matrix meets RIP's criteria is NP-hard in general. As a result, designing an efficient sensing matrix is critical. These sensing matrices are responsible for signal compression at the encoder end and accurate or approximate reconstruction at the decoder end. For signal compression, different sensing matrices are utilized in compressed sensing. There are random, deterministic, structural, and optimized sensing matrices are used in compressed sensing [2020_book].
Check if satisfies RIP
Checking satisfies RIP is combinatorial hard in general so it's unreasonable to ask a computer to verify a matrix satisfies RIP. In order to get around this problem, we need an understanding of what matrices satisfy RIP and recover with high probability.
- Random Sensing matrices: Gaussian, Bernoulli, Rademacher
- Deterministic Sensing Matrices: binary, bipolar, ternary, Vandermond
- Structural Sensing Matrices: Toeplitz, Circulant, Hadamard
- Optimized Sensing Matrices: (Parkale, Nalbalwar, Sensing Matrices in Compressed Sensing)
Are some examples. Different sensing matrices are more suited for different problems, but in general, we want to use an alternative to Gaussian because it reduces the computational complexity.
Verification of the Sensing matrix
Definition Mutual Coherence
Let , the mutual coherence is defined by:</math>
- TODO switch s to k
Welch bound > [1] > is the coherence between and We want a small because it will be close to the normal matrix, which satisfies RIP. Also, will be needed for the step size for the following IHT.
Need to make the connection of Coherence to RIP and RIC.
- TODO switch s to k
Algorithms
Three big groups of algorithms are:[2]
- Optimization methods: includes minimization i.e. Basis Pursuit, and quadratically constraint
minimization i.e. basis pursuit denoising.
- Greedy methods: include Orthogonal matching pursuit and Compressive Sampling Matching Pursuit (CoSaMP)
- thresholding-based methods: such as Iterative Hard Thresholding(IHT) and Iterative Soft Thresholding, Approximate IHT or AM-IHT, and many more.
More cutting-edge methods include dynamic programming.
We will cover one, i.e. IHT. WHY IHT THEN? Basis pursuit, matching pursuit type algorithms belong to a more general class of iterative thresholding algorithms. [3] So IHT seems like the ideal place to start. If everything compliment with RIP, then IHT has fast convergence.
Algorithm IHT
The convex program mentioned in introduction has an equivalent nonconstraint optimization program.
(cite IT for sparse approximations) ??? [1]. In statistics we call the regularization LASSO with as the regularization parameter. This is the closest convex relaxation to the first program menttioned in the introduction.[The Benefit of Group Sparsity]
Then
sub modual
Define the threashholding operators as: selects the best-k term approximation for some k
Stopping criterion is iff RIC [4]
- Input
- output
- Set
- While Stopping criterion false do
- end while
- return:
is a Adjoint matrix i.e. the transpost of it's cofactor.
Numerical Example
sub module iteration 1
calculate
Calculate the Error
Calculate the new y
Check if error is less then
sub module iteration 2
calculate
Calculate the Error
Calculate the new y
Check if error is less then
Stopping condition meat
Applications and Motivations
Low-Rank Matrices
The Netflix Prize was accompanied by low-rank matrix recovery or the matrix completion problem. The approach then fills in the missing values in the user's ratings for movies that the user hasn't seen. These estimates are based on ratings from other users, who have similar ratings if a matrix is created with all the users as rows and the movie titles as columns. Because some users' interests will be similar and therefore overlap, it is possible to reduce the degrees of freedom significantly. This low-rank structure is frequently assumed for the problem domain of collaborative filtering ~cited citations.
Dictionary Learning
The goal in dictionary learning is to infer the original dictionary as possible. Instead of using a predefined dictionary, researchers have found that learning the dictionary by obtaining "dynamic features" from training data often yields representation. Biometric features can be taken from video clips of each subject in a dataset and used to populate the dictionary's columns. Using random projections and sparse representations for iris detection for noncontact biometrics-based authentication systems from video samples has been proposed ~cited citations.
Single-pixel cameras
Single-pixel cameras or single detector imaging are used in situations when detectors are either prohibitively expensive or difficult to miniaturize. A microarray is made up of a large number of miniature mirrors that can be individually turned on and off. The mechanism behind the random sampling, which results in low coherence between measurements, is the most important component of the single-pixel camera. This microarray reflects the light from the scene, and a lens combines all of the reflected beams into one sensor, which is the single detector of the camera used to capture the image ~cited citations.
Conclusion
Referencse
- ↑ 1.0 1.1 1.2 Cite error: Invalid
<ref>
tag; no text was provided for refs named:1
- ↑ Cite error: Invalid
<ref>
tag; no text was provided for refs named:0
- ↑ Cite error: Invalid
<ref>
tag; no text was provided for refs named:4
- ↑ Cite error: Invalid
<ref>
tag; no text was provided for refs named:2
- ↑ D. L. Donoho, “Compressed sensing,” vol. 52, pp. 1289–1306, 2006, doi: 10.1109/tit.2006.871582.
- ↑ E. J. Candès and T. Tao, “Decoding by linear programming,” IEEE Trans. Inf. Theory, vol. 51, no. 12, Art. no. 12, 2005, doi: 10.1109/TIT.2005.858979.
- ↑ D. L. Donoho, “Compressed sensing,” vol. 52, pp. 1289–1306, 2006, doi: 10.1109/tit.2006.871582.
- ↑ T. Blumensath and M. E. Davies, “Iterative Hard Thresholding for Compressed Sensing,” May 2008.
- ↑ S. Foucart and H. Rauhut, A mathematical introduction to compressive sensing. New York [u.a.]: Birkhäuser, 2013.
- ↑ R. G. Baraniuk, “Compressive Sensing [Lecture Notes],” IEEE Signal Processing Magazine, vol. 24, no. 4, Art. no. 4, 2007, doi: 10.1109/MSP.2007.4286571.