Compressive Sensing Lecture notes by Richard G. Baraniuk

advertisement
Compressive Sensing
Lecture notes by Richard G. Baraniuk
IEEE Signal processing magazine July 2007
Compressive Sensing Tutorial PART 1
Svetlana Avramov-Zamurovic
January 15, 2009.
Motivation

Practically in the vast majority of applications data acquisition is based on
Shannon/Nyquist sampling theorem that requires sampling rate at least
twice the message signal bandwidth in order to achieve exact recovery.

Source frequency characteristic: This requirement is not practical for video
industry since the signal bandwidth is very wide and the technology is not
feasible to achieve necessary processing rates in order to satisfy the
Shannon/Nyquist sampling theorem. Practical solution bandlimits the
signals and prevents aliasing.

Channel frequency characteristic: There is a significant class of signals
(pictures for example) that are compressible: not all the data is necessary
to transmit in order to get ‘good enough’ representation of the original
message. Practical solution introduces lossy compression processing at the
source level.

Compressed sensing is new method to capture and represent compressible
signals at the rate well below Nyquist’s rate.



Employs nonadaptive linear projections (random measurement matrix)
Preserves the signal structure (length of the sparse vectors is conserved)
Reconstructs the signal from the projections using optimization process (L1 norm)
Classical Approach: Transform coding
PICTURE
N samples
(ALL
measurements
are taken)
ORTHOGONALIZITION
Full set of projections
is found
SORTING
CODING
straightforward
decoding
Only K coefficients
K largest
coefficients selected are coded
N-K coef. dumped
K<<N
EXAUSTIVE SEARCH
Compressed sensing
PICTURE
CODING
Signal Reconstruction
Capture only significant components
Only M (K ≈ M) samples (measurements) are taken
(a) Measurements must be carefully designed
(b) Original signal (picture) must be sparse
(1) Underdetermined system
M<N
(2) Reconstructed signal
must have N components
(3) L1 norm is used to find
sparse representation
Inefficiencies of transform coding
(classical)
 Sampling required by Nyquist rate is high
producing huge number of samples N.
 Finding signal representation in an
orthonormal basis (Fourier, etc.) is
computationally intensive since ALL si must
be found and sorted in order to find K
significant coefficients.
 Along with the magnitude of K coefficients
their location needs to be coded introducing
the overhead.
Definitions
x target signal,
represenation of an image
in time or space domain
x is in N dimensional domain
 x1 
 11
    orthonormal basis  
 xN 
 N 1
 1N 
 
i othonormal vector

 NN 
N
 i1 
  
 iN 
Classical Approach: Find the signal projections on a given basis : x=  si i
i 1
sirepresenation of the image in  domain   iT x
x is K-sparce if only K basis vecotors have s  0 (thoeretical)
i
i
x is compressible if it has just a few large s coeficients and many small (practical)
i
Compressable signals are well represented by K-sparse representations.
Compressed sensing: Measure only significant components (M
 y1 
 11
y measurements
    measurement matrix  
y is in M dimensional domain
 yM 
M 1
y=x=s  Goal y  s
1N 


MN 
N)
Φ
y
Ψ
S
=
y
Θ
=
M
x
N
(a)
(b)
K-sparse
(a) Compressive sensing measurement process with a random Gaussian
measurement matrix and discrete cosine transform (DCT) matrix . The
vector of coefficients s is sparse with K = 4.
Φ (phi, measurement matrix) Ψ (psi, orthonormal basis)
Θ (theta, Compressed Sensing reconstruction matrix)
(b) Measurement process with    There are four columns that
correspond to nonzero si coefficients; the measurement vector y is a
linear combination of these columns.
From IEEE Signal processing magazine July 2007, R. G. Baraniuk, Compressive sensing
S
Compressive sensing solution



Measurement matrix must be stable and produce reconstruction of original signal, x, (length N) from M
measurements
1) Ψ othonormal basis must be selected 2), stable Φ (measurement matrix) must be designed (CS
measures PROJECTIONS of signal onto a basis) and 3) the signal x has to be reconstructed from the
underdetermined system using optimization (norm L1) linear combinations, since the projections( linear
combinations are measured) we are extracting the sparse components using optimization algorithm (L1)
with restrictions ( Rip and incoherence)
The following conditions must hold in order to find unique sparse solution

Restricted isometry property RIP

If x is K sparse and K magnitudes and locations are known than for M>K we can find solution
under the following condition:
Θ=ΦΨ MUST preserve the length of the vectors sharing the same K non-zero coefficients as s
1  


v
v
2
2
 1 
R
a 2   ai 2
i 1
ε is a small number, v is any vector sharing the same K non-zero coefficients as s
Incoherence

Rows of Φ (measurement matrix) CAN NOT sparsely represent the columns of Ψ (othonormal
basis) and vice versa. MUST be DENSE!
Direct construction of
measurement matrix (Φ)

Verifying RIP requires

zero entries of length N
However BOTH RIP and incoherence could be achieved by
simply selecting Φ randomly



N
 
K
all possible combinations of K non-
φij are independent and identically distributed random variables
from Gaussian probability density with zero mean and 1/N
variance.
So measurements y= Θ x are randomly weighted linear
combinations of elements of x.
iid Gaussian Φ (MxN) properties


Φ is incoherent with the basis Ψ=I, Θ= ΦΨ=Φ with high
probability if M≥cKlog(N/K), c is a small constant (log2(N/K+1)
so M<<N.
Φ is universal: Θ= ΦΨ will be iid Gaussian and have RIP with
high probability regardless of choice of the othonormal basis Ψ
Designing a signal reconstruction
algorithm



M measurements, y, random measurements matrix, Φ, and basis Ψ, are
used to reconstruct compressible signal x (length) N or equivalently its
sparse coefficient vector s.
Since M<N there are infinitely many solutions. Since Θ(s+r)= Θ(s)=y for
any vector r in the null space N(Θ). But we have a restriction that the
solution is sparse.
Signal reconstruction algorithm aims to find signal’s sparse coefficient
vector in the (N-M)-dimensional translated null space H=N(Θ)+s.



L2 norm (energy) minimized. Pseudo inverse is closed form to find the
solution but it DOES NOT find sparse solution.
L0 norm counts the number of non-zero elements of s. This optimization can
recover K sparse signal exactly with high probability using only M=K+1
measurements but solving it is unstable and NP-complete requiring
exhaustive enumeration of all (N K) possible locations of the non-zero entries
in s.
L1 norm (adding absolute values of all elements) can exactly recover K
sparse signals and closely approximate compressible signals with high
probability using only M≥cKlog(N/K) iid Gaussian measurements.

Convex optimization reducing to linear programming known as Basis pursuit
with the computation complexity about O(N3)
RN
S
a
b
c
(a) The subspaces containing two sparse vectors in R3 lie close to the
coordinate axes.
(b) Visualization of the L2 minimization (5) that finds the non-sparse
point-of-contact s between the 2 ball (hyper-sphere, in red) and the
translated measurement matrix null space (in green).
(c) Visualization of the L1 minimization solution that finds the sparse
point-of-contact s with high probability thanks to the pointiness of the 1
ball.
From IEEE Signal processing magazine July 2007, R. G. Baraniuk, Compressive sensing
Compressive Imaging Camera Architecture
• Single detector: By time multiplexing a single detector, we can use a less expensive and yet more sensitive
photon detector. A single detector camera can also be adapted to image at wavelengths that are currently
impossible with conventional CCD and CMOS imagers (for example IR, security applications).
• Universality: Random and pseudorandom measurement bases are universal in the sense that they can be
paired with any sparse basis. This allows exactly the same encoding strategy to be applied in a variety of
different sensing environments; knowledge of the nuances of the environment are needed only at the
decoder. Random measurements are also future-proof: if future research in image processing yields a
better sparsity-inducing basis, then the same set of random measurements can be used to reconstruct an
even better quality image.
• Encryption: A pseudorandom basis can be generated using a simple algorithm according to a random seed.
Such encoding effectively implements a form of encryption: the randomized measurements will
themselves resemble noise and be meaningless to an observer who does not know the associated seed.
• Robustness and progressivity: Random coding is robust in that the randomized measurements have equal
priority, unlike the Fourier or wavelet coefficients in current transform coders. Thus they allow a
progressively better reconstruction of the data as more measurements are obtained; one or more
measurements can also be lost without corrupting the entire reconstruction.
• Scalability: We can adaptively select how many measurements to compute in order to trade off the amount of
compression of the acquired image versus acquisition time; in contrast, conventional cameras trade off
resolution versus the number of pixel sensors.
• Computational asymmetry: Finally, CI places most of its computational complexity in the decoder, which will
often have more substantial computational resources than the encoder/imager. The encoder is very
simple; it merely computes incoherent projections and makes no decisions.
From D. Takhar at al.
A new Compressive Imaging Camera Architecture using Optical domain Compression
Compressive imaging test bed








Micro-actuated mirrors => commercially viable MEMS technology for the video/projector display market
as well as laser systems and telescope optics. Texas Instruments (TI) digital micromirror device (DMD).
TI DMD developer’s kit and accessory light modulator package (ALP).
The DMD consists of an array of electrostatically actuated micro-mirrors where each mirror the array is
suspended above an individual SRAM cell. The DMD micro-mirrors form a pixel array of size 1024×768.
Each mirror rotates about a hinge and can be positioned in one of two states (+12 degrees and −12
degrees from horizontal); thus light falling on the DMD may be reflected in two directions depending on
the orientation of the mirrors.
Desired image is formed on the DMD plane with the help of a biconvex lens; this image acts as an
object for the second biconvex lens which focuses the image onto the photodiode.
The light is collected from one out of the two directions in which it is reflected (e.g., the light reflected
by mirrors in the +12 degree state). The light from a given configuration of the DMD mirrors is
summed at the photodiode to yield an absolute voltage that yields a coefficient y(m) for that
configuration. The output is amplified through an op-amp circuit and then digitized by a 12-bit ADC.
User decides how many measurements to take (M).
Object is illuminated by LED light source, 1kHz.
Compressed Sensing:




system directly acquires a reduced set of M incoherent projections of
an N-pixel image x without first acquiring the N pixel values;
random position of the mirror(+12 or -12); programmable so we can
decide on pattern (Radamacher)
Mirrors can programmed to stay in a position longer producing better
resolution. Assumption: Object is stationary!
Summing of the light in the photo diode, each measurement
multiplexes several pixel values (activating DMD in blocks) CS
reconstruction algorithm extracts them.
From D. Takhar at al.
A new Compressive Imaging Camera Architecture using Optical domain Compression
Scene
Image
Photodiode
Bitstream
A/D
Digital Micromirror Device
Reconstruction
Single-pixel, compressive
sensing camera.
Array
Random Number Generator
Major challenges
(1) Acquisition speed
(2) DSP processing speed
The images in (a) and (b) are not meant to be aligned.
(a) Conventional
digital camera image
of a soccer ball.
(b) 64 × 64 black-and-white image x of the same ball (N =
4,096 pixels) recovered from M = 1,600 random
measurements taken by the camera in (a)
From IEEE Signal processing magazine July 2007, R. G. Baraniuk, Compressive sensing
Experimental results
Original
16384 Pixels
1600 Measurements
(10%)
16384 Pixels
3300 Measurements
(20%)
Sources of noise
(1) Nonlinearities in the photodiode
(2) Non-uniform reflectance of the mirrors
through the lens focus onto the photodiode
(changing the weighting of the pattern blocks)
(3) Non-uniform mirror positions
65536 Pixels
1300 Measurements
(2%)
65536 Pixels
3300 Measurements
(5%)
Robustness of the CS reconstruction algorithm
Suppresses quantization noise from ADC
and photodiode circuit noise.
Notes

RB in TI : can use 10 Mega pixel camera to get the resolution of 25 Mega
pixel picture! CS is scalable!

The classical approach takes the whole picture at intervals (or as a stream
in the video). How fast is signal reconstruction in CS? Can it keep up with
the moving target? In the classical approach the records are taken and after
the event the zoom in possibility is open. Cs never takes ALL of the
information and relays on random approach that statistically gives good
results but this method may not be acceptable in some applications.


But in D Takhar at al. possibility of zoom in on a part of the image as opposed to
acquiring the whole image and then extract the piece of interest. Adaptively
sectioned to highlight.
RB : just choose the measurement matrix to have random iid Gaussian
distribution.

Michael Lustig : just choosing samples at random is not good idea, because the
distance between samples are not preserved. Globally, such a sampling scheme
has uniform density but locally you will get high density areas and ‘holes’ these
holes really mess up the receiver array source of redundancy. Using Poisson disk
sampling distribution provides local uniform sampling redundancy are maximally
exploited.
Download