Joint Segmentation and Reconstruction of Hyperspectral Data with Compressed Measurements Qiang Zhang

advertisement
Joint Segmentation and Reconstruction of
Hyperspectral Data with Compressed Measurements
Qiang Zhang1,∗ , Robert Plemmons2 , David Kittle3 , David Brady3 ,
and Sudhakar Prasad4
1
Biostatistical Sciences, Wake Forest School of Medicine, Winston-Salem, NC 27157, USA
2
Computer Science and Mathematics, Wake Forest University, Winston-Salem, NC 27106,
USA
3
4
Electrical and Computer Engineering, Duke University, Durham, NC 27708, USA
Physics and Astronomy, University of New Mexico, Albuquerque, NM 87131, USA
∗
Corresponding author: qizhang@wakehealth.edu
This work describes numerical methods for the joint reconstruction and
segmentation of spectral images taken by compressive sensing coded aperture
snapshot spectral imagers (CASSI). In a snapshot, a CASSI captures a twodimensional (2D) array of measurements that is an encoded representation of
both spectral information and 2D spatial information of a scene, resulting in
significant savings in acquisition time and data storage. The reconstruction
process decodes the 2D measurements to render a three-dimensional spatiospectral estimate of the scene, and is therefore an indispensable component
of the spectral imager. In this study, we seek a particular form of the
compressed sensing solution that assumes spectrally homogeneous segments
in the two spatial dimensions, and greatly reduces the number of unknowns,
often turning the under-determined reconstruction problem into one that is
over-determined. Numerical tests are reported on both simulated and real
c 2011 Optical Society of
data representing compressed measurements. ⃝
America
OCIS codes: 100.3010, 110.3010, 110.4234
1
1.
Introduction
Hyperspectral remote sensing technology allows one to capture images using a range of
spectra from ultraviolet to visible to infrared. Multiple images of a scene or object are created
using light from different parts of the spectrum. These hyperspectral images, forming a data
cube, can be used, e.g. for ground or space object identification [1], astrophysics [2], and
biomedical optics [3].
Since a single digital image can typically have a size of 12 megabytes or more, the size of
hyperspectral data cubes could easily move to the gigabyte level. Such high dimensional data
pose challenges in both data acquisition and reconstruction. Technologies such as tunable
filters [4] or computed tomography [5] measure at least as many elements as there are in a
hyperspectral data cube, and require long acquisition time and vast data storage and transfer,
but relatively little effort in reconstruction. Recently proposed compressive imagers such as
a Coded Aperture Snapshot Spectral Imager (CASSI) [6, 7] need only take a single snapshot
from which to reconstruct a hyperspectral data cube if the latter is assumed to be sparse
in some basis. Clearly compressed measurements need much less acquisition time and data
storage, although they demand powerful algorithms for data reconstruction, which is usually
a highly under-determined problem. For example, if the size of a vectorized hyperspectral
cube f is n1 × n2 × n3 , the double disperser CASSI (DD-CASSI, to be discussed later) [6]
measures only a 2D vectorized image g with size n1 × n2 , from which we need to reconstruct
the original cube. Here we consider first a simple least squares approach to estimating f ,
i.e. fˆ = arg minf ∥Hf − g∥22 , where H is the DD-CASSI system matrix having a size n1 n2 ×
n1 n2 n3 . Since there are fewer rows than columns in H, the problem can possess an infinite
number of solutions, and hence additional constraints are needed to obtain a meaningful,
practical solution.
Mathematical analysis of compressive sensing has drawn a great deal of attention after
important results were obtained by Donoho [8, 9] and Candes, Romberg and Tao [10]. The
problem is generally posed as one of finding the sparsest solution to f , with sparsity measured
by the l1 norm or l0 pseudo-norm, i.e.
min ∥f ∥p , subject to Hf = g,
(1)
where p = 0 or 1. Sometimes though the signals themselves are not sparse, an appropriate
or even optimal basis Φ can be found on which the projections of the signals are sparse, i.e.
min ∥Φf ∥p , subject to Hf = g.
f,Φ
(2)
Though it remains important to develop methods to determine the optimal basis for given
2
classes of signals and measurement systems, in many practical situations the signals to be
reconstructed are composed of relatively homogeneous segments or clusters, e.g. hyperspectral images in remote sensing problems [11], and in computed tomography [12]. The optimal
basis here is the set of segment membership functions on the two spatial dimensions, which
take on values of either 0 or 1 for hard segmentation or within the interval [0, 1] for fuzzy
segmentation [13]. Variational segmentation algorithms are a popular area of research, see
e.g. [12, 14–16], but they are often applied either directly to the original measurements or
after the reconstruction. By contrast, work is just beginning on combining reconstruction
and segmentation in a more general linear inverse problem setting. For instance, Li, Ng and
Plemmons [15] have coupled the segmentation and deblurring/denoising models in order
to simultaneously segment and deblur/denoise degraded hyperspectral images. Ramlau and
Ring [17] jointly reconstructed and segmented Radon transformed tomography data. Xing,
Zhou, Castrodad, Guillermo Sapiro, and Carin [18] divided the hyperspectral imagery into
contiguous blocks and used Bayesian dictionary learning method to estimate both the dictionry and the spectra in each block from limited noisy observations. However, neither example
considered compressive measurements directly from a sensor.
We now formalize our approach in the following. Let H ∈ RN ×n1 n2 n3 represent the hyperspectral imager system matrix, f ∈ Rn1 n2 n3 ×1 the vectorized hyperspectral cube, and
g ∈ RN ×1 the vectorized measurement image(s). Consider the following linear least squares
problem for reconstructing compressive measurements,
min ∥Hf − g∥22 .
f
(3)
By compressive measurements, we mean N < n1 n2 n3 . Here we assume the solution f is
composed of a limited number of segments or materials, each of which essentially has a
homogeneous value at each spectral channel. Thus we seek a decomposed solution described
in a continuous form as
L
∑
f (x, y, λ) =
ui (x, y)si (λ),
(4)
i=1
where ui (x, y) is the ith membership function, whose values can be either 0 or 1 for a hard
segmentation or in the interval [0, 1] for a fuzzy segmentation, and satisfies the constraint
∑L
th
segment or
i=1 ui = 1. Here, si (λ) represents the spectral signature function of the i
material. The support of ui lies only on the two spatial dimensions represented by x and y,
and is thus independent of the spectral dimension represented by λ. The spectral signatures,
s = {si (λ), i = 1, . . . , L}, vary only along the spectral dimension. The discrete version of f
3
can thus be written as,
fˆ =
L
∑
ui sTi ,
(5)
i=1
where fˆ ∈ Rn1 n2 ×n3 is the folded hyperspectral cube, ui ∈ Rn1 n2 ×1 is the vectorized membership function and si ∈ Rn3 ×1 . We identify f with fˆ in the remainder of the paper.
With this decomposed form of f , the savings in the number of signals to reconstruct is
significant, L(n1 n2 + n3 ) unknowns compared to the original n1 n2 n3 , since we can usually
expect n3 ≫ L. As shown later, due to this reduction the reconstruction problem to be
solved can be turned from being highly under-determined to being over-determined. Also,
for each i the membership function ui is expected to be sparse in terms of its gradient, since
only near the boundaries is there significant mixing of members in our applications. Thus a
total variation (TV) regularization is particularly suitable here.
This manuscript is organized as follows. In Section 2, we present an alternating least
squares (ALS) approach to separate the original problem (3) into two subproblems and to
solve for ui and si in an alternating fashion. For the first subproblem, namely to solve for
ui given si , when the system matrix H preserves boundaries, e.g. the DD-CASSI system,
we present a generalized segmentation algorithm based on the Chan-Vese model [14] and
the variational model [15, 16] within an inverse problem setting. This enables us to directly
segment a hyperspectral data cube from a single observed image, given known spectral signatures. For more general cases of H, we present and compare three algorithms with increasing
computation complexity to solve the first sub-problem. The second sub-problem, namely to
solve for si , given ui , is typically a highly overdetermined problem, and simple regularized
pseudoinverse methods suffice here, as we shall see. In Section 3, we present both simulated
and real compressed hyperspectral images to be reconstructed with the proposed method.
By increasing the number of measurements N , we also study the relationship between N
and the probability of successful reconstruction from random initial values, and how this
probability depends on the nature of the reconstruction algorithms, thus shedding light on
certain thresholds on the minimum number of measurements required for a finite probability
of success. By comparing the three algorithms, we were pleasantly surprised to find that even
a simple pseudo-inverse approach can achieve quite satisfying results when N is sufficiently
large. We present our main conclusions and comments in Section 4.
2.
Joint Segmentation and Reconstruction
Using the decomposed form of f ≡ fˆ described in (5), we have essentially turned the original linear problem into a non-linear problem. Nevertheless, the reduction in the number of
unknowns enables us to take advantage of compressed measurements to achieve satisfactory
reconstruction results. Our approach to solve this problem is to optimize for ui and si by
4
alternating iterations, given initial values of ui or si . But before we go into details, some
simple results and notation on matrix-vector multiplications are provided.
We denote the concatenation of all membership vectors ui as u ∈ RLn1 n2 ×1 , i.e. u =
(uT1 , uT2 , . . . , uTL )T , and the concatenation of all spectral signatures si as s ∈ RLn3 ×1 , i.e.
s = (sT1 , sT2 , . . . , sTL )T .
Proposition 1. With the decomposed form of f in (5), we have the following equality
Hf = HSu = HU s.
(6)
Here S = S̃ ⊗ I1 , the column vectors of S̃ are the spectral signature vectors si , and I1 ∈
Rn1 n2 ×n1 n2 is the identity matrix. U = Ũ ⊗I2 , Ũ = (ũij )n1 n2 ×L , where each column corresponds
to the vectorized ui , and I2 ∈ Rn3 ×n3 is another identity matrix. ⊗ denotes the Kronecker
product.
Proposition 2. With the decomposed form of f in (5), we have the following equality
Hf =
L
∑
HSi ui =
i=1
L
∑
HUi si .
(7)
i=1
Here Si = si ⊗ I1 , Ui = ũi ⊗ I2 , and ũi is the ith column of Ũ .
The proofs of both propositions are basically bookkeeping by noting the vectorization of
f is done first by the spectral dimension and then by two spatial dimensions, i.e.
f = (f111 , . . . , f11n3 , f121 , . . . , f12n3 , . . . , fn1 n2 1 , . . . , fn1 n2 n3 )T .
(8)
We denote HS by Hs and HU by Hu to represent the system matrices used for solving for
u and s, respectively.
2.A.
Alternating Least Squares (ALS)
The ALS approach turns the original problem (3) into two subproblems, i.e. given s(n) at
step n, we solve,
u(n+1) = arg min ∥Hs(n) u − g∥22 , subject to Eu = 1
u
(9)
and given u(n+1) at step n, we solve
s(n+1) = arg min ∥Hu(n+1) s − g∥22 , subject to s > 0,
s
5
(10)
(n)
(n+1)
where Hs = HS (n) and Hu
= HU (n) by Proposition 1. The two constraints on u
and s respectively are the sum-to-one constraint for the membership functions expressed as
∑
Eu = Li=1 ui = 1, with E = (I1 , I1 , . . . , I1 )n1 n2 ×n1 n2 L , and the nonnegativity constraint for
the spectral signatures.
Since the second sub-problem (10) has only Ln3 unknowns, it is often over-determined, i.e.
N ≫ Ln3 . Hence, simple approaches such as the pseudo-inverse (PI) method with Tikhonov
regularization for noise can be sufficient, and the nonnegativity constraint can be satisfied
with a projection function onto the nonnegative orthant. But because N ≫ Ln3 , the PI
solution could be well within the nonnegative orthant and this would render the projection
step unnecessary.
The first sub-problem (9) has Ln1 n2 unknowns and hence could be under-determined, but
because often times n3 ≫ L, and when Ln1 n2 ≤ N < n1 n2 n3 , we can expect (9) to be exact
or even over-determined. Even better, if we consider a hard segmentation, the number of
nonzeros in u would not exceed n1 n2 , because at each pixel there is exactly only one ui that
is nonzero. In cases when (9) is under-determined, we seek a sparse solution in its null space
with the sparsity defined as the l1 norm of the boundaries of segments, and rewrite (9) as
min
u
L
∑
∥Gui ∥1 , subject to Hs u = g and Eu = 1,
(11)
i=1
where G is the gradient matrix, the discrete version of the more familiar operator ∇, and
thus the problem above can also be regarded as a total variation (TV) minimization problem,
e.g. [19], with the functional
Fu = ∥Hs u −
g∥22
+ αu
L
∑
∥Gui ∥1 ,
(12)
i=1
where αu is the TV regularization parameter.
As a necessary condition for convergence, consider a simpler situation when both (9)
and (10) are exact or over-determined and we only use L2 functionals (informally, square
integrable functionals) for optimizing u and s. We then have the following theorem that
guarantees a nonincreasing sequence of the L2 functional values through the iterations of the
ALS approach.
Theorem 1. When only using the L2 functionals, the alternating least square approach
results in a nonincreasing sequence of functional values, i.e.
∥Hs(n) u(n+1) − g∥22 ≥ ∥Hu(n+1) s(n+1) − g∥22 ≥ ∥Hs(n+1) u(n+2) − g∥22 ,
6
(13)
for n = 0, 1, 2, . . ..
Proof. The proof becomes obvious after noting that by Proposition 1,
Hs(n) u(n+1) = HS (n) u(n+1) = HU (n+1) s(n) = Hu(n+1) s(n) .
Thus by definition of s(n+1) , we have the first inequality,
∥Hs(n) u(n+1) − g∥22 = ∥Hu(n+1) s(n) − g∥22 ≥ ∥Hu(n+1) s(n+1) − g∥22 .
(n+1) (n+1)
And by Hu
ity,
s
(n+1) (n+1)
= Hs
u
(14)
and the definition of u(n+2) , we have the second inequal-
∥Hu(n+1) s(n+1) − g∥22 = ∥Hs(n+1) u(n+1) − g∥22 ≥ ∥Hs(n+1) u(n+2) − g∥22 .
(15)
In the more general case, when (9) is under-determined and when noise is present, we seek
to minimize a combined functional of (10) and (12) with added Tikhonov regularization for
si ,
L
L
∑
∑
F = ∥Hf − g∥22 + αu
∥Gui ∥1 + αs
∥si ∥22 ,
(16)
i=1
i=1
where αs is the Tikhonov regularization parameter. For (16), we have a similar theorem to
guarantee a nonincreasing sequence of F (n) with the ALS approach.
Theorem 2. Using the functional defined in (16), the alternating least square approach
results in a nonincreasing sequence of functional values.
Proof. First we define the solutions of two subproblems as,
u
(n+1)
=
arg min ∥Hs(n) u
u
−
g∥22
+ αu
L
∑
∥Gui ∥1 , subject to Eu(n+1) = 1,
and
s
(n+1)
=
(17)
i=1
arg min ∥Hu(n+1) s
s
−
g∥22
+ αs
L
∑
∥si ∥22 , subject to s > 0.
(18)
i=1
Then the functional value at step n after optimizing for u becomes
Fu(n)
=
∥Hs(n) u(n+1)
−
g∥22
+ αu
L
∑
i=1
7
(n+1)
∥Gui
∥1
+ αs
L
∑
i=1
(n)
∥si ∥22 ,
(19)
and after optimizing for s at step n, it becomes
Fs(n) = ∥Hu(n+1) s(n+1) − g∥22 + αu
L
∑
(n+1)
∥Gui
∥1 + αs
i=1
L
∑
(n+1) 2
∥2 .
∥si
(20)
i=1
By the inequality in (14) and by the definition of s(n+1) , we have
Fu(n) ≥ Fs(n) .
(21)
Similarly, by the inequality in (15) and the definition of u(n+2) , we can see
Fs(n) ≥ Fu(n+1) .
(22)
Next we prove that the nonincreasing sequence leads to a minimizer of (16) in the space
A defined as,
L
∑
A = {(u, s)|ui ∈ BV (Ω), ui ≥ 0,
ui = 1, s ≥ 0},
(23)
i=1
where BV (Ω) is a bounded variation space. We rewrite (16) in its continuous form,
∫
F=
[H(f ) − g] dxdy + αu
2
Ω
L ∫
∑
|∇ui |dxdy + αs
Ω
i=1
L ∫
∑
i=1
s2i (λ)dλ,
(24)
Λ
where H is the continuous version of the system operator H.
Theorem 3. In the space A, there exists a minimizer of the functional defined in (24).
∫
Proof. If we take ui = 1/L and si = 1, then F = Ω [H(1) − g]2 dxdy + αs L|Λ| < +∞. Since
F ≥ 0 in A, we know the infimum of the functional would be finite. Let (u(n) , s(n) ) ⊆ A be a
minimizing sequence of the ALS approach, with u(n) and s(n) defined in (17) and (18). Then
there exists a constant M > 0, such that
F(u(n) , s(n) ) ≤ M.
(25)
Hence each term in F(u(n) , s(n) ) is also bounded, i.e.,
αu
L ∫
∑
i=1
(n)
It is also easy to see that ui
(n)
|∇ui |dxdy ≤ M.
(26)
Ω
is bounded in L1 since ∥ui ∥L1 (Ω) =
8
∫
Ω
(n)
ui dxdy < |Ω|, and by
(n)
the compactness of BV space, up to a subsequence also denoted by {ui } after relabeling,
there exists a function u∗i ∈ BV (Ω) such that
(n)
→ u∗i strongly in L1 (Ω),
(n)
→ u∗i a.e. (x, y) ∈ Ω,
ui
ui
(n)
∇ui
→ ∇u∗i in the sense of measure.
(27)
Also, by the lower semi-continuity of total variation,
∫
∫
(n)
∗
|∇ui |dxdy.
|∇ui |dxdy ≤ lim inf
n→∞
Ω
(28)
Ω
Since u(n) satisfies two constraints, by convergence, so would u∗i . For the convergence of s(n) ,
we have at each iteration
s
(n)
= P+
[(
Hu(n)T Hu(n)
+ αs I
)−1 (
Hu(n)T g
)]
.
(29)
Here we use the discrete form of the objective functional in (18) for to avoid introducing more
(n)
(n)T
notations while maintaining the spirit of the proof. Since the eigenvalues of Hu Hu + αs I
(n)
have a lower bound αs , and both Hu and g are bounded, we know s(n) must have an upper
(n)
bound. Furthermore, because u(n) → u∗i , we know Hu → Hu∗ , where Hu∗ = HU ∗ as defined in
(n)
Proposition 1. Hence we can find an upper bound of Hu for all n and thus an upper bound
of s(n) for all n. By the boundedness of the sequence {s(n) }, we can extract a subsequence
also denoted by {s(n) } and a limit s∗ such that
s(n) → s∗ .
(n)
Since ui
∫
∗
→ u∗i a.e. (x, y) ∈ Ω and s(n) → s∗ , by Fatou’s Lemma we know
[H(f ) − g] dxdy +
Ω
(30)
2
∑∫
i
s∗i (λ)dλ
∫
≤ lim inf
n→∞
[H(f
(n)
) − g] dxdy +
Ω
2
∑∫
(n)
si (λ)dλ, (31)
i
∑ (n)
∑
(n)
where f ∗ = i u∗i (x, y)s∗i (λ) and f (n) = i ui (x, y)si (λ). By (28) and (31), the functional
also satisfies the inequality
F(u∗ , s∗ ) ≤ lim inf F(u(n) , s(n) ),
n→∞
(32)
and we can conclude (u∗ , s∗ ) must be a minimizer.
In the next two sections we focus our attention on solving the first subproblem (12), since
9
the second subproblem is highly over-determined and well-posed, and thus a simple pseudoinverse approach would be sufficient. For the first subproblem, we start from a simpler case,
i.e. when the operator Hs preserves the boundaries. This effectively renders ui independent
from uj , when i ̸= j, and this also makes individual entries within ui independent from each
other. For solving this problem, we generalize two existing segmentation approaches, the
Chan-Vese model [14] and a variational model [15, 16]. We then move to the more general
Hs , i.e. where boundaries are not preserved and where all ui are coupled together by Hs .
2.B.
Boundary Preserving Operator H
Clearly any operations on a hyperspectral data cube only along the spectral dimension will
preserve the boundaries in the two spatial dimensions. One simple example is the summation
operator along the spectral dimension to turn a hyperspectral cube into a 2D image. Another
example, the DD-CASSI system, is similar to the summation operator except that by using
a coded aperture, the system effectively first multiplies the hyperspectral cube with a random aperture code, and then sums along the spectral dimension, as shown in the following
equation.
∑
fijk ci,j−k ,
(33)
gij =
k
where cij is the calibrated 2D aperture code without spectral content, f is the original
hyperspectral cube and g is the observed 2D image. A third example is the correlation
operator, e.g. the moving average method [16], which computes the correlations between
spectral signatures. Here we formalize the definition as following.
Definition 1. We define a system matrix H as boundary preserving if it satisfies
HSi ui = Λi ui ,
(34)
where H ∈ Rn1 n2 ×n1 n2 n3 , Si ∈ Rn1 n2 n3 ×n1 n2 is defined in Proposition 2, and Λi ∈ Rn1 n2 ×n1 n2
is a diagonal matrix.
Notice that the boundary preserving operators have a fixed number of rows, i.e. n1 n2 , for
the apparent reason that the boundary of an object could be exactly the boundary of the
scene. Again using the summation operator as an example, we have H = eT ⊗ I1 , where
e ∈ Rn3 ×1 is a constant vector with all entries 1. It is not hard to verify that HSi is a
diagonal matrix.
Theorem 4. The boundary preserving operator effectively renders each ui independent from
each other, that is to say, we can optimize for each ui separately.
∑
∑
∑
Proof. Because i ui = 1, we have g = i g ⊙ ui = i Diag(g)ui , where ⊙ represents the
element wise product and Diag(g) is a diagonal matrix with elements of g on the diagonal.
10
Hence by Proposition 2 and by the definition of the boundary preserving operator, we have
∥Hs u−g∥22 = ∥
L
∑
L
L
L
∑
∑
∑
HSi ui −
Diag(g)ui ∥22 = ∥
(Λi −Diag(g))ui ∥22 =
∥Λi −Diag(g)∥22 u2i .
i=1
i=1
i=1
i=1
(35)
Clearly ui is decoupled from uj when i ̸= j. Additionally, uij1 is also independent from uij2
for j1 ̸= j2 . Note that the last equality in (35) is due to the hard segmentation assumption,
∑
i.e. the binary ui , and hence the cross terms disappear after expanding ( Li=1 aij uij )2 , where
aij is the j th element on the diagonal of Λi − Diag(g).
Due to the independence of the ui , we can solve for each ui separately through, for example,
by using the popular active contour PDE model, also called the Chan-Vese (C-V) model [14],
which can be described using the following functional,
∫
∫
∫
∫
2
H(ϕ) + α1 (u0 − c1 ) H(ϕ) + α2 (u0 − c2 )2 (1 − H(ϕ)),
F (ϕ, c1 , c2 ) = µ δ(ϕ)|∇ϕ| + ν
Ω
Ω
Ω
Ω
(36)
where ϕ(x, y, t) is the function whose zero level set represents the evolving curve C, and is
chosen to be positive inside C and negative outside C. H(ϕ) is the Heavyside function of ϕ,
which is defined as
{
1 ϕ ≥ 0,
H(ϕ) =
(37)
0 ϕ < 0.
δ(ϕ) is the derivative of H(ϕ), u0 is the observed image, c1 is the mean intensity within C,
and c2 is the mean intensity outside C. Also, µ, ν, α1 , α2 are weighting parameters of the
model. The last two terms on the right are the force terms that either expand or shrink the
initial contour C0 . We refer readers to [14] for further details of the model.
For a hard segmentation, the membership function ui (x, y) is equivalent to the Heavyside
function of ϕ, and g(x, y) is the observed image. The modification only involves slightly
changing two force terms of the original model, i.e.
∫
∫
∫
∫
2
F (ϕ, c1 , c2 ) = µ δ(ϕ)|∇ϕ| + ν
ui + α1 (g̃i − g) ui + α2 (g̃i − g − c2 )2 (1 − ui ), (38)
Ω
Ω
Ω
Ω
where g̃i (x, y) is the image spectrally coded by the ith spectral signature. The discrete form
of g̃i (x, y) is derived by taking the diagonal of HSi .
We can see the only difference from the original C-V model are the force terms. The
modified C-V model in (38) is a generalization of the original model in the sense that if
g̃i (x, y) = c1 , (38) is the same as the original C-V model. It is equivalent to say that we are
segmenting the zero value segment of the image g̃i − g, rather than c1 − g. This becomes clear
if we replace f with its decomposed form and apply the boundary preserving assumption in
11
the following equation,
(
H(f ) = H
∑
)
ui (x, y)si (λ)
=
∑
i
H(si (λ))ui (x, y),
(39)
i
and the difference between H(f ) and g now becomes,
H(f ) − g =
∑
[H(si (λ)) − g(x, y)] ui (x, y).
(40)
i
Here g̃ would simply be H(si (λ)) and its difference from g is the driving force for ui . If
H(si (λ)) = ci , the equation above becomes the regular C-V model.
This modification is crucial since a spectrally homogeneous segment in f could result in
inhomogeneous intensities in the same segment of H(f ) or g. One example would be to
multiply f with a random cube and then sum along the spectral dimension. The variation of
intensities in the original segments would give us wrong segmentation results if we directly
applied the C-V model. An example of such is provided in Figure 1 (a) in Section 3.
The active contour model evolves each individual initial contour according to the given
force terms, and thus it depends on the initial contours and has to be implemented separately
for each segment. The variational model proposed in [15,16] is able to segment all L segments
at the same time and is also formulated for the fuzzy segmentation, i.e. ui ∈ [0, 1]. The model
also accounts for the sum-to-one constraint ui and the nonnegativity constraints on si . Briefly,
the original formulation includes a total variation (TV) regularization term for ui and the
intensity difference between values in the segment and the mean of the area, i.e.
∑∫
∑∫
|∇ui |dxdy + αu
(g − ci )2 u2i dxdy.
(41)
i
Ω
i
Ω
The modification again comes by replacing the force term with the difference between images
g̃ and g.
∑∫
∑∫
|∇ui |dxdy + αu
[g̃i − g]2 u2i dxdy.
(42)
i
Ω
i
Ω
Again, (42) can be seen as a generalization of (41) because when g̃i = ci , we have exactly
the same model. From Theorem 4, we know that after discretization (42) is the same as the
functional Fu in (12). One example is shown by Figure 1 (b) in Section 3, where we were
able to correctly segment and reconstruct a simulated Hubble Satellite Telescope (HST)
hyperspectral cube while only using a single DD-CASSI image.
12
2.C.
General Operator H
In many applications, such as in using the Radon Rransform [20] in computed tomography
or in using the single disperser CASSI (SD-CASSI) system [7], the boundaries cannot be
preserved by the operator. Hence, we cannot always rely on modifying existing segmentation
algorithms. We solve the constrained subproblem (9) with approaches such as pseudo-inverse
methods, methods with total variation minimization, or suboptimal methods for sparse signal
recovery, e.g. the matching pursuit [21] or the orthogonal matching pursuit [22]. In this
section, we will discuss these three options.
As stated before, due to the reduction in the number of unknowns, the subproblem (9) can
be exact or even over-determined when Ln1 n2 ≤ N , while measurements are still compressed
when N < n1 n2 n3 . The pseudo-inverse solution in this case would simply be:
u = PZ2
[(
HsT ∗ Hs
)−1 (
HsT g
)]
,
(43)
where PZ2 is the projection operator onto the space Z2 = {0, 1} for a hard segmentation.
The projection operator onto the space [0, 1] for a fuzzy segmentation is
{
{(
)−1 ( T ) } }
u = min max HsT ∗ Hs
Hs g , 0 , 1 .
(44)
The sum-to-one constraint can be satisfied or closely satisfied by adding a regularization
term to the least square functional, i.e.
α∥Eu − 1∥22 .
(45)
The pseudo-inverse solution often suffers from noise, but this can be effectively reduced
with total variation regularization. Li, Ng and Plemmons [15] proposed the following functional with an auxiliary variable v to jointly segment and deblur/denoise hyperspectral images.
1
αu
FT V = ∥Hs u − g∥22 + ∥v − u∥22 + ∥Gv∥1 ,
(46)
2
2
where v is the auxiliary variable for smoothing u. The following equations, similarly derived
as in [15, 16, 19], can be used to alternatively solve for u and v through the iterations,
p(n) + ϕ∇(div p(n) − αu u(n) )
,
1 + ϕ∇(div p(n) − αu u(n) )
1
= u(n) − div p(n+1) ,
αu
T
= (Hs Hs + αu I)−1 (HsT g + αu v (n+1) ),
p(n+1) =
v (n+1)
u(n+1)
13
(47)
where p(n+1) serves as an intermediate variable. See [15,19] for details such as the satisfaction
of constraints. We call this method the TV regularization method.
The matching pursuit (MP) algorithm finds the “best matching” projections of multidimensional data onto an over-complete dictionary. It iteratively generates for any signal u and
any dictionary Hs a sorted list of indices and scalars which constitute a sub-optimal solution
to the problem of sparse signal representation. The orthogonal matching pursuit algorithm
(OMP) is a modification to MP that maintains full backward orthogonality of the residual
at every step. In each iteration, OMP calculates a new signal approximation u(n) . The approximation error r(n) = u − u(n) is then used in the next iteration to determine which new
element is to be selected. In particular, the selection is based on the inner products between
the current residual r(n) and the column vectors of Hs . The complete algorithm is described
in [23].
Here we modify OMP slightly by introducing a gradient operator, because though the
membership function ui (x, y) is not necessarily sparse, its gradient, being supported over the
segment boundaries, often is. Let û = Gu. We first solve
min ∥û∥1 , subject to Hs G−1 û = g,
û
(48)
through OMP and then set u = G−1 û. We call this approach the gradient orthogonal matching pursuit (GOMP). A similar approach is given in [24] to restore images from subsets of
Fourier transforms. With GOMP, we only consider the hard segmentation, i.e. after solving
for all ui , we search for the maximum ui for each x and y,
{
ui (x, y) =
3.
1 ui (x, y) ≥ uj (x, y), ∀j ̸= i
0 otherwise.
(49)
Numerical Examples
Our experiments used to illustrate the effectiveness of the proposed methods are divided
into two parts, one for the boundary preserving system operator and the other for more
general system operators. Though the methods can also be applied to other compressive
hyperspectral sensing systems, the systems we consider here are the DD-CASSI and SDCASSI.
3.A.
Boundary Preserving Operator
The operator of interest here is the DD-CASSI system, whose forward model has been
described by (33). The details on the optics can be found in [6]. In the three examples
presented in this section, we move progressively from completely simulated data to completely
real data. In the first example, through the forward model (33), we simualte a DD-CASSI
14
image from from a simulated hyperspectral cube of the Hubble Space Telescope (HST)
[25] at size 177 × 193 × 33, then we simulate a DD-CASSI image from a recently acquired
hyperspectral dataset on a urban setting [26] at size 320×360×31, and finally we reconstruct
a hyperspectral cube from a real DD-CASSI image of fluorescent beads [27] at size 600 ×
800 × 59.
We start by testing the generalized C-V model (38) and the generalized variation model
(42) with known spectral signatures in the HST scene. Figures 1 and 2 compare the two
modified models with the original ones and we clearly see the advantages offered by the two
modified models. In Figure 1, an initial square contour is selected in the bottom cylinder of
the satellite, while the modified C-V model progressed to the correct bound and stabilized,
the original C-V model easily broke out of the cylinder boundary and moved to quite arbitrary
places, though it does stabilize in the end. This is due to the highly varying intensities within
each segment in the observed image g. Figure 2 is a similar comparision between the two
total variation (TV) models and clearly the original TV model (41) cannot segment out the
correct areas, while the modified model (42) can.
Next we jointly segmented and reconstructed the original hyperspectral cube from a single
DD-CASSI image. The ALS approach starts from random values of s, i.e. without any prior
knowledge of spectral signatures, and uses the modified variation model (42) to estimate
u(n+1) and the pseudo-inverse method with the Tikhonov regularization to estimate s(n+1) .
The reconstructed tensor is compared with the original tensor, using the l2 norm error,
ϵ=
∥f − f0 ∥22
,
∥f0 ∥22
(50)
where f and f0 are the reconstructed and original hyperspectral cubes, respectively.
Figure 3 shows the first 25 iterations of the estimated membership functions u shown in a
false color map, with one color assigned to each segment. Even without any prior knowledge
on the spectral signatures, we can correctly reconstruct and segment the original HST cube
and the solution u apparently converges after 20 iterations. Figure 4 compares the estimated
spectral signatures with the true ones. Only the fifth spectral signature in the middle differs
from the true one, due to the rather small prevalence of that particular material in the
scene. We also directly ran the general purpose TwIST algorithm [28] without assuming the
decomposed form of solution f , for which the norm error was 0.205 as compared to .012 for
the new approach. The new approach has dramatically reduced the norm error by taking
advantage on the solution form of f .
To test the robustness of the proposed method against noise, we polluted the simulated
DD-CASSI image with white noise having a standard deviation of .3, which effectively results
15
in an SNR of 21 dB, with SNR in dB defined as,
(
SN R = 20 log10
σsignal
σnoise
)
.
(51)
Figure 5 shows the estimated u and s, both of which closely resemble the true ones, though
noisier. The norm error is .05.
In the second example, we considered a real hyperspectral dataset of size 320 × 360 × 29,
with bands from .453µm to .719µm, taken by OpTech (OpTech International, Inc. Kiln,
MS) on the campus of University of Southern Mississippi, in Gulfport, Mississippi. The
data were collected as part of a project led by Professor Paul Gader at the University
of Florida Department of Computer and Information Science and Engineering [26]. The
original dataset had 72 bands ranging from .4µm to 1.0µm, but because the data cube of
DD-CASSI, h(x, y, λ), are only calibrated from .45µm to .72µm, and after matching the
calibrated wavelengths of DD-CASSI with those actually measured bands by OpTech, we
chose 31 of the 72 bands and ran it through a DD-CASSI forward model for a simulated
DD-CASSI image. The left image in Figure 6 shows the Google map of the area and the
right image shows the simulated DD-CASSI image. Here we ran the segmentation algorithm
directly on the simulated DD-CASSI image with seven known spectra taken from the original
hyperspectral cube, shown in Figure 7, and the result is shown in Figure 8. The algorithm
clearly separated out the areas of trees, water/shadow, grass and pavement with a relatively
high resolution. For example, we observe sharp boundaries between trees and grass, and
thin lines of dirt splitting the grass area into four parts in the middle slightly to the left.
Unfortunately due to the chosen bands, the two road strips at the bottom were recognized
as grass, but this can be fixed once we have a system cube covering more long wavelengths.
In terms of target identification, the reconstruction/segmentation clearly identifies three
targets purposely placed on tables just above the ground near the center of the scene. These
consist of colored cloths placed on tables, and we mark these targets by the yellow circle in
Figure 8. This is quite encouraging considering we are only using one snapshot, or 3.45%
of the original data. The only missed target has a similar spectra as grass from .453µm to
.719µm, but if our DD-CASSI system cube is calibrated to longer wavelengths, we will be
able to identify that target as well. The norm error between the reconstructed hyperspectral
data cube and the original cube is .033.
In the third example, we used a real DD-CASSI image of a biomedical scene with fluorescent beads of different colors [27]. The size of the image, 600 × 800, plus the number of
spectral channels, 59, would result in more than 28 million unknowns if we were not using
the decomposed form, and that might render the reconstruction impossible by general purpose algorithms such as TwIST. However, with our decomposition approach we can estimate
16
the spectral signatures of those beads quite close to the estimates given in [27]. Figure 9(a)
shows the original DD-CASSI image and Figure 9(b) shows the reconstructed membership
functions, where we try to match the bead colors with false colors as close as possible, and
they are labeled in the same way as in Figure 9(d). Figure 9(c) and 9(d) are directly taken
from [27] for comparison purposes. Notice that because the signatures of CA and R are quite
close, we cannot quite differentiate between them in Figure 9(b). Also, several long wavelength beads identified in [27], namely C1,S1,S2 and S3, are missing here because pixels of
these beads in the observed DD-CASSI image have weak intensities in the order of 0.01, as
hardly seen in Figure 9(a), while intensities at other beads are between .15 to 1. Hence the
algorithm recognizes them as background. Our reconstruction does show that the boundaries
of beads appear to have different colors from the center parts due to the geometry, while the
reconstruction in [27] treats the whole area of each bead region as uniform in color.
3.B.
General Operator
The system operator of interest here is the SD-CASSI system [7], which can be characterized
using subscript notation as follows:
gij =
n3
∑
ci,j+k fi,j+k,k ,
(52)
k=1
where c is the 2D calibrated aperture code, and the only difference from the DD-CASSI
system is the multiplexing in both the second spatial dimension and the spectral dimension,
and hence the image taken has size n1 × (n2 + n3 − 1). Because of the multiplexing in
the second spatial dimension, this operator does not preserve boundaries. The details of the
optical setup can be found in [7]. Three hyperspectral cubes were used to simulate SD-CASSI
images, the first having only an isolated square object in the image of size 128 × 128 × 33, the
second having two side-by-side rectangular objects in the image to test the method’s ability
to identify sharp boundaries, at size 128 × 128 × 33, and the third being the HST cube at
size 177 × 193 × 33. The forward model (52) was used to generate three SD-CASSI images.
We jointly segmented and reconstructed the original hyperspectral cube from one or more
frames of SD-CASSI images. The ALS approach again starts from random values of s, i.e.
without any prior knowledge, and uses three different algorithms to estimate u(n+1) , the
simple regularized pseudo-inverse (PI) method, the TV regularization model (TV) and the
gradient orthogonal matching pursuit (GOMP) model. The pseudo-inverse method with
Tikhonov regularization was used to estimate s(n+1) .
For the first scene with an isolated object in the scene, we were able to reconstruct exactly
as shown in Figure 10, with only one frame of SD-CASSI image. This tells us that one highly
compressive SD-CASSI image may be sufficient for reconstructing hyperspectral cubes with
17
only isolated objects.
For the second scene with two rectangular objects side by side, using random starting
values of s, we were able to reconstruct exactly from 8 simulated frames of SD-CASSI images
using the simple PI approach, as shown in Figure 11. We also compared the three algorithms
for estimating u with prior known spectral signatures from a single SD-CASSI image. Figure
12 shows the false color images of estimated membership functions by all three algorithms,
where we see an increasing reconstruction quality from top to bottom. Here, GOMP results
on the third row match almost exactly with the true ones, while the TV results are smoother
and closer to truth than those of PI. Clearly, GOMP may be the best candidate among the
three when measurements are most compressed.
Our experience indicates that testing all three algorithms without any prior knowledge
of s, we found a zero probability to reconstruct successfully when the number of frames is
below 4. With the pseudo-inverse approach and with 8 frames of measurements, we were
able to reconstruct exactly to a high probability, and the results are shown in Figure 11. We
further expanded this experiment by running all three algorithms 100 times with different
random starting values, and with different numbers of frames of measurements. A parallel
computation was set up on an eight-core machine and was finished in a week. The probabilities of successful reconstructions are shown in Figure 13(a) and the computation time per
iteration of ALS is shown in Figure 13(b). First we observe that when the number of frames
of observations is below 4, there is no chance to reconstruct successfully for any algorithm.
With 4 frames, the PI approach still could not reconstruct successfully while the TV and
GOMP algorithms have a finite probability, 7% and 8%, respectively. The GOMP has a 40%
success rate at 5 frames compared to 27% for PI and 28% for TV. This informs us that the
GOMP could do reasonably well with more compressive measurements, but given a certain
scene with a fixed number of nonzero projections, there should be a threshold on the minimum number of frames for a successful reconstruction. The TV approach turns out to be
the least successful beyond a sufficiently high number of frames, while we observe an almost
equivalent success rate by PI and GOMP, which is quite encouraging since PI is much faster
as shown in Figure 13(b).
Next, we tried to reconstruct the complex HST scene with a number of frames of SDCASSI images. Here we only tested the PI approach, since the other two would take a
prohibitively long computation time. With random starting values of s, when the number of
frames reach 20, we had satisfactory reconstruction results, as shown in Figure 14, though
in the beginning, the membership function map looks quite messy. Figure 14(b) compares
the estimated six most dominant spectral signatures with the corresponding true ones, and
they all agree quite well.
Finally, the real SD-CASSI images of a color chart image shown in Figure 15(a) were used
18
for reconstructing both the color segments and their spectra. Since the spectral signatures
within each segment are fairly uniform, the scene served as a good candidate for testing
the algorithm. For brevity, we only show the reconstructed colors in the second row as
indicated by the red box in Figure 15(b). The six colors on the second row from left to
right are respectively orange, purplish blue, moderate red, purple, yellow green and orange
yellow, and the spectra of five colors excluding the purple were measured by an Ocean
Optics (Dunedin, Florida) spectrometer. A total of 12 frames of SD-CASSI images were used
for reconstructing a hyperspectral cube in 44 wavelength channels. Figure 15(c) shows the
identified six segments. Clearly we are able to identify the sharp boundaries of the segments.
For the accuracy of reconstructed spectra, we compared them with both references by the
spectrometer in Figure 16(a) and the reconstruction by TwIST in Figure 16(b). Our results
match almost perfectly those obtained by TwIST, and also show good agreements with the
spectrometer references.
4.
Conclusions
By assuming a low-rank representation, we were able to jointly segment and reconstruct
hyperspectral data cubes from compressed measurements using the algorithms proposed in
this study. The assumed form of the solution dramatically reduced the number of unknowns
in the reconstruction, thus allowing better accuracy and faster computation. In our tests, we
obtained promising results after applying the algorithms to a sequence of simulated and real
DD-CASSI and SD-CASSI images.
For boundary preserving measurement operators, such as the DD-CASSI system, we generalized two existing segmentation algorithms, the Chan-Vese model and a total variation
model, to directly segment the compressed measurements along the spatial and spectral dimensions without first reconstructing the data into hyperspectral cubes. In our studies, we
show poor segmentation results after applying the original two models directly on simulated
DD-CASSI images, but very good results after applying the generalized two models. For more
general operators, such as that of SD-CASSI, we proposed three methods, i.e. the pseudoinverse method, the total variation method and the gradient orthogonal matching pursuit
(GOMP) method for estimating the membership functions with given spectral signatures. A
more extensive simulation study shows that several frames of measurements are often needed
for a successful reconstruction, though the theoretical threshold on the minimum number
of frames needs further analysis. With enough observations, the probabilities of successful
reconstruction by different algorithms tend to converge, and even the simple pseudo-inverse
approach can provide successful reconstruction. But when the number of observations is
limited, the GOMP method performs best. For real SD-CASSI data taken on a color chart
scene, we have shown good reconstruction results with the GOMP method.
19
In using a low-rank representation, the algorithms proposed here would better suit situations where distinct spectra lie in some spatially homogeneous areas (segments), e.g.,
cartoon-like scenes. As the size of segments gets smaller and the number of spectra becomes
greater, we would expect a performance reduction given the same number of compressed
measurements. In these situations, we recommend adding measurements as in [29] but when
the scene in study becomes too complex, more general algorithms such as TwIST may be
more reliable. However, we would caution that in these non-sparse situations, the compressive
sensing approach may not be that appealing after all.
5.
Acknowledgement
This work was sponsored in part by the AFOSR under grant number FA9550-08-1-0151.
References
1. K. Jorgensen, J. Africano, K. Hamada, E. Stansbery, P. Sydney, and P. Kervin, “Physical
properties of orbital debris from spectroscopic observations,” Advances in Space Research
34, 1021–1025 (2004).
2. R. Lin, B. Dennis, G. Hurford, D. Smith, A. Zehnder, P. Harvey, D. Curtis, D. Pankow,
P. Turin, M. Bester, and Others, “The Reuven Ramaty high-energy solar spectroscopic
imager (RHESSI),” Solar Physics 210, 3–32 (2002).
3. T. Pham, F. Bevilacqua, T. Spott, J. Dam, B. Tromberg, and S. Andersson-Engels,
“Quantifying the absorption and reduced scattering coefficients of tissuelike turbid media
over a broad spectral range with noncontact Fourier-transform hyperspectral imaging,”
Appl. Optics 39, 6487–6497 (2000).
4. H. Morris, C. Hoyt, and P. Treado, “Imaging spectrometers for fluorescence and Raman
microscopy: acousto-optic and liquid crystal tunable filters,” Applied spectroscopy 48,
857–866 (1994).
5. J. Mooney, V. Vickers, M. An, and A. Brodzik, “High-throughput hyperspectral infrared
camera,” Journal of the Optical Society of America A 14, 2951–2961 (1997).
6. M. Gehm, R. John, D. Brady, R. Willett, and T. Schulz, “Single-shot compressive spectral
imaging with a dual-disperser architecture,” Optics Express 15, 14013–14027 (2007).
7. A. Wagadarikar, R. John, R. Willett, and D. Brady, “Single disperser design for coded
aperture snapshot spectral imaging,” Applied Optics 47, B44–B51 (2008).
8. D. Donoho, “Compressed sensing,” IEEE Transactions on Information Theory 52, 1289–
1306 (2006).
9. D. Donoho and M. Elad, “Optimally sparse representation in general (nonorthogonal)
dictionaries via l1 minimization,” Proceedings of the National Academy of Sciences of
the United States of America 100, 2197 (2003).
20
10. E. Candès, J. Romberg, and T. Tao, “Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information,” IEEE Transactions on information theory 52, 489 (2006).
11. J. Ball and L. Bruce, “Level set segmentation of remotely sensed hyperspectral images,” in “2005 IEEE International Geoscience and Remote Sensing Symposium, 2005.
IGARSS’05. Proceedings,” (2005), pp. 5638–5642.
12. R. Maksimovic, S. Stankovic, and D. Milovanovic, “Computed tomography image analyzer: 3D reconstruction and segmentation applying active contour models–[] snakes’,”
International journal of medical informatics 58, 29–37 (2000).
13. J. Kent and K. Mardia, “Spatial classification using fuzzy membership models,” IEEE
Transactions on Pattern Analysis and Machine Intelligence pp. 659–671 (1988).
14. T. Chan and L. Vese, “Active contours without edges,” IEEE Transactions on image
processing 10, 266–277 (2001).
15. F. Li, M. Ng, and R. Plemmons, “Coupled Segmentation and Denoising/Deblurring
Models for Hyperspectral Material Identification,” Numerical Linear Algebra with Applications (2010).
16. F. Li, M. Ng, R. Plemmons, S. Prasad, and Q. Zhang, “Hyperspectral image segmentation, deblurring, and spectral analysis for material identification,” in “SPIE Symposium
on Defense, Security, and Sensing,” , vol. 7701 (2010), vol. 7701.
17. R. Ramlau and W. Ring, “A Mumford-Shah level-set approach for the inversion and
segmentation of X-ray tomography data,” Journal of Computational Physics 221, 539–
557 (2007).
18. Z. Xing, M. Zhou, A. Castrodad, G. Sapiro, and L. Carin, “Dictionary learning for noisy
and incomplete hyperspectral images,” IEEE Transactions on Geoscience and Remote
Sensing (submitted).
19. A. Chambolle, “An algorithm for total variation minimization and applications,” Journal
of Mathematical Imaging and Vision 20, 89–97 (2004).
20. S. Deans, The Radon transform and some of its applications (Wiley New York, 1983).
21. S. Mallat and Z. Zhang, “Matching pursuits with time-frequency dictionaries,” IEEE
Transactions on signal processing 41, 3397–3415 (1993).
22. J. Tropp, “Greed is good: Algorithmic results for sparse approximation,” IEEE Transactions on Information Theory 50, 2231–2242 (2004).
23. Y. Pati, R. Rezaiifar, and P. Krishnaprasad, “Orthogonal matching pursuit: recursive
function approximation withapplications to wavelet decomposition,” in “1993 Conference
Record of The Twenty-Seventh Asilomar Conference on Signals, Systems and Computers,
1993,” (1993), pp. 40–44.
24. R. Maleh, A. Gilbert, and M. Strauss, “Sparse gradient image reconstruction done
21
faster,” in “IEEE International Conference on Image Processing, 2007. ICIP 2007,” ,
vol. 2 (2007), vol. 2.
25. Q. Zhang, H. Wang, R. Plemmons, and V. Pauca, “Tensor methods for hyperspectral
data analysis: a space object material identification study,” Journal of the Optical Society
of America A 25, 3001–3012 (2008).
26. P. Gader, A. Zare, R. Close, and G. Tuell, “Co-registered hyperspectral and LiDAR Long
Beach, Mississippi data collection,” (2010). University of Florida, University of Missouri,
and Optech International.
27. C. Cull, K. Choi, D. Brady, and T. Oliver, “Identification of fluorescent beads using a
coded aperture snapshot spectral imager,” Applied Optics 49, 59–70 (2010).
28. J. Bioucas-Dias and M. Figueiredo, “A new TwIST: two-step iterative shrinkage/thresholding algorithms for image restoration,” IEEE Transactions on Image processing 16, 2992–3004 (2007).
29. D. Kittle, K. Choi, A. Wagadarikar, and D. Brady, “Multiframe image estimation for
coded aperture snapshot spectral imagers,” Applied Optics 49, 6824–6833 (2010).
22
List of Figure Captions
Fig. 1. (a) Segmented bottom cylinder area using the original Chan-Vese model; (b)
Segmented same area using the generalized C-V.
Fig. 2. Segmentation of a DD-CASSI image (a) using the original TV algorithm and (b)
using the modified TV with known si (λ).
Fig. 3. The first 25 iterations of the estimation of membership functions from left to right
and top to bottom.
Fig. 4. The estimated spectral signatures in blue are compared with the true ones in red.
Fig. 5. The joint estimation of membership functions ui in (a) and spectral signatures, si in
(b) in the presence of noise (SNR = 21 dB).
Fig. 6. The left image shows a Google map of the area while the right image shows one band
of the simulated DD-CASSI image.
Fig. 7. The estimated spectra of the segmented areas.
Fig. 8. The segmentation map estimated from the DD-CASSI image.
Fig. 9. (a) The DD-CASSI image of the beads. (b) The estimated membership function
map. (c) The beads spectral signatures from [27]. (d) The pseudo-colored reconstructed
image using TV minimization [27], with the spectral signatures of the marked beads in (c).
Fig. 10. The top left image is a slice of the simulated hyperspectral cube; the top middle is
the simulated SD-CASSI image; the top right is the spectral signature applied to all pixels
in the square to the left; the bottom left is a slice of the reconstructed cube and the bottom
middle is the reconstructed spectral signature.
Fig. 11. The top left image is a slice of the simulated hyperspectral cube; the top middle is
the simulated SD-CASSI image; the top right are two spectral signatures applied to pixels
in two rectangles to the left; the bottom left is a slice of the reconstructed cube and the
bottom middle are the reconstructed spectral signatures.
Fig. 12. Comparison of three reconstruction methods using a single frame observation and
known si (λ). Only the middle 40 rows of the original scene is used here.
Fig. 13. (a) Relationship between the probability of a successful reconstruction and the
number of observed frames for the three reconstruction methods. (b) Computation time of
each method per iteration of ALS.
Fig. 14. (a) The estimated membership function map in 20 iterations. (b) The estimated
spectral signatures in blue, compared to the original in red.
Fig. 15. (a) The imaged color chart object. (b) The original SD-CASSI image with the
reconstructed area in the red rectangle box. (c) The reconstructed segments.
Fig. 16. (a) The reconstructed spectra compared with reference by an Ocean Optics
spectrometer. (b) The reconstructed spectra compared with reconstruction using TwIST.
1500 Iterations
1500 Iterations
20
20
40
40
60
60
80
80
100
100
120
120
140
140
160
160
20
40
60
80
100
(a)
120
140
160
180
20
40
60
80
100
120
140
160
180
(b)
Fig. 1. (a) Segmented bottom cylinder area using the original Chan-Vese
model; (b) Segmented same area using the generalized C-V.
20
20
40
40
60
60
80
80
100
100
120
120
140
140
160
160
20
40
60
80
100
(a)
120
140
160
180
20
40
60
80
100
120
140
160
180
(b)
Fig. 2. Segmentation of a DD-CASSI image (a) using the original TV algorithm
and (b) using the modified TV with known si (λ).
50
50
50
50
100
100
100
100
150
100
150
150
50
100
150
150
50
50
100
100
150
150
50
100
150
150
50
100
150
100
150
50
50
50
50
100
100
100
150
50
100
150
150
50
100
150
100
150
50
50
50
50
100
100
100
150
50
100
150
150
50
100
150
100
150
50
50
50
50
100
100
100
150
100
150
150
50
100
150
150
50
100
150
50
100
150
50
100
150
50
100
150
150
50
100
50
100
150
50
100
150
50
150
50
100
150
150
100
50
150
100
150
50
100
150
100
150
50
150
100
100
100
150
100
150
50
150
50
100
150
Fig. 3. The first 25 iterations of the estimation of membership functions from
left to right and top to bottom.
Glue
Black Rubber Edge
0.8
0.8
0.8
0.8
0.6
0.4
0.6
0.4
0.2
0
400
0
400
600
800
Wavelength (nm)
Honeycomb side/Bolt
Reflectance
1
Reflectance
1
0.2
0.6
0.4
0.2
0
400
600
800
Wavelength (nm)
Copper Stripping
0.6
0.4
0.2
0
400
600
800
Wavelength (nm)
Honeycomb Top
1
0.8
0.8
0.8
0.8
0.4
Reconstructed
Honeycomb side
0.2
0
400
Bolt
600
800
Wavelength (nm)
0.6
0.4
0.2
0
400
Reflectance
1
Reflectance
1
0.6
0.6
0.4
0.2
600
800
Wavelength (nm)
0
400
600
800
Wavelength (nm)
Background
1
Reflectance
Reflectance
Solar Cell
1
Reflectance
Reflectance
Aluminum
1
0.6
0.4
0.2
600
800
Wavelength (nm)
0
400
600
800
Wavelength (nm)
Fig. 4. The estimated spectral signatures in blue are compared with the true
ones in red.
20
40
60
80
100
120
140
160
20
40
60
80
100
120
140
160
180
(a)
Glue
Black Rubber Edge
0.8
0.8
0.8
0.8
0.6
0.4
0.6
0.4
0.2
0
400
0
400
500
600
700
Wavelength (nm)
Honeycomb side/Bolt
Reflectance
1
Reflectance
1
0.2
0.6
0.4
0.2
0
400
500
600
700
Wavelength (nm)
Copper Stripping
0.6
0.4
0.2
0
400
500
600
700
Wavelength (nm)
Honeycomb Top
1
0.8
0.8
0.8
0.8
0.4
0.2
0
400
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0
400
Reflectance
1
Reflectance
1
0.6
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0
400
500
600
700
Wavelength (nm)
Background
1
Reflectance
Reflectance
Solar Cell
1
Reflectance
Reflectance
Aluminum
1
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0
400
500
600
700
Wavelength (nm)
(b)
Fig. 5. The joint estimation of membership functions ui in (a) and spectral
signatures, si in (b) in the presence of noise (SNR = 21 dB).
50
50
100
100
150
200
150
250
200
300
250
350
300
400
50
100
200
300
100
150
200
250
300
350
400
Fig. 6. The left image shows a Google map of the area while the right image
shows one band of the simulated DD-CASSI image.
Pavement
Tree
0.8
0.8
0.8
0.8
0.6
0.4
0
400
0.6
0.4
0.2
600
800
Wavelength (nm)
0
400
Sand
0.6
0.4
0.2
600
800
Wavelength (nm)
0
400
Rooftop
0.8
0.8
0.8
0.4
0.2
0
400
0.4
Reflectance
1
Reflectance
1
0.6
0.2
600
800
Wavelength (nm)
0
400
0.6
0.4
0.2
600
800
Wavelength (nm)
0
400
600
800
Wavelength (nm)
Water/Shadow
1
0.6
Reflectance
1
Reflectance
1
0.2
Reflectance
Grass
1
Reflectance
Reflectance
Dirt
1
0.6
0.4
0.2
600
800
Wavelength (nm)
0
400
600
800
Wavelength (nm)
Fig. 7. The estimated spectra of the segmented areas.
Water/Shadow
50
Tree
100
Sand
150
Rooftop
200
Grass
250
Pavement
300
50
100
150
200
250
300
350
Dirt
Fig. 8. The segmentation map estimated from the DD-CASSI image.
CA3
R3
G2
RO1
100
O1
R1
R4
Y1
BG1
CA4
200
Y3
Y2
100
CA1
CA2
300
200
Y4
Y7
300
YG1
400
R5
400
Y5
R2
G1
500
Y6
500
600
100
200
300
400
500
600
700
800
600
100
200
300
400
500
(a)
600
700
800
(b)
(c)
(d)
Fig. 9. (a) The DD-CASSI image of the beads. (b) The estimated membership
function map. (c) The beads spectral signatures from [27]. (d) The pseudocolored reconstructed image using TV minimization [27], with the spectral
signatures of the marked beads in (c).
1
20
40
40
60
60
80
80
100
100
120
120
20 40 60 80 100 120
0.8
Reflectance
20
0.6
0.4
0.2
50
100
150
0
450
500
550
600
Wavelength (nm)
650
1
20
0.8
Reflectance
40
60
80
100
0.6
0.4
0.2
120
20 40 60 80 100 120
0
450
500
550
600
Wavelength (nm)
650
Fig. 10. The top left image is a slice of the simulated hyperspectral cube; the
top middle is the simulated SD-CASSI image; the top right is the spectral
signature applied to all pixels in the square to the left; the bottom left is a
slice of the reconstructed cube and the bottom middle is the reconstructed
spectral signature.
1
20
40
40
60
60
80
80
100
100
120
120
0.8
Reflectance
20
0.4
0.2
50
20 40 60 80 100120
0.6
100
150
0
450
500
550
600
Wavelength (nm)
650
1
20
Reflectance
0.8
40
60
80
100
0.6
0.4
0.2
120
20 40 60 80 100120
0
450
500
550
600
Wavelength (nm)
650
Fig. 11. The top left image is a slice of the simulated hyperspectral cube; the
top middle is the simulated SD-CASSI image; the top right are two spectral
signatures applied to pixels in two rectangles to the left; the bottom left is a
slice of the reconstructed cube and the bottom middle are the reconstructed
spectral signatures.
Pseudo−Inverse
10
20
30
40
20
40
60
80
100
120
100
120
Total Variation
10
20
30
40
20
40
60
80
Orthogonal Matching Pursuit
10
20
30
40
20
40
60
80
100
120
Fig. 12. Comparison of three reconstruction methods using a single frame
observation and known si (λ). Only the middle 40 rows of the original scene is
used here.
(a)
(b)
Fig. 13. (a) Relationship between the probability of a successful reconstruction
and the number of observed frames for the three reconstruction methods. (b)
Computation time of each method per iteration of ALS.
20
20
20
20
20
40
40
40
40
40
60
60
60
60
60
80
80
80
80
20
40
60 80
20
40
60
80
20
40
60
80
80
20
40
60 80
20
20
20
20
20
40
40
40
40
40
60
60
60
60
80
80
20
40
60 80
80
20
40
60
80
40
60
80
40
60 80
20
20
20
20
40
40
40
40
40
60
60
60
60
60
80
80
80
80
40
60 80
20
40
60
80
20
40
60
80
40
60 80
20
20
20
20
40
40
40
40
40
60
60
60
60
60
80
80
80
80
40
60 80
20
40
60
80
20
40
60
80
80
20
40
60
80
20
40
60
80
20
40
60
80
80
20
20
20
60
80
20
20
20
40
60
80
20
20
80
20
40
60 80
1
0.8
0.8
0.6
0.4
0.4
0.2
500
550
600
Wavelength (nm)
0
450
650
1
1
0.8
0.9
Reflectance
Reflectance
0
450
0.6
0.6
0.4
0.2
450
500
550
600
Wavelength (nm)
650
500
550
600
Wavelength (nm)
0.4
0.2
500
550
600
Wavelength (nm)
650
500
550
600
Wavelength (nm)
650
1.5
0.8
0.7
450
0.6
0
450
650
Reflectance
0.2
0.8
Reflectance
1
Reflectance
Reflectance
(a)
500
550
600
Wavelength (nm)
650
1
0.5
0
450
(b)
Fig. 14. (a) The estimated membership function map in 20 iterations. (b) The
estimated spectral signatures in blue, compared to the original in red.
100
200
300
400
500
600
700
800
900
200
400
(a)
600
800
1000
1200
1400
(b)
10
20
30
40
50
200
400
600
800
1000
1200
1400
(c)
Fig. 15. (a) The imaged color chart object. (b) The original SD-CASSI image
with the reconstructed area in the red rectangle box. (c) The reconstructed
segments.
Purplish Blue
Orange
1
0.4
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0.2
400
800
Yellow Green
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0
400
800
500
600
700
Wavelength (nm)
800
Orange Yellow
1
1
0.8
0.8
Reflectance
Reflectance
0.8
Reflectance
0.6
0.6
0.4
0.2
0
400
1
0.8
Reflectance
Reflectance
0.8
0
400
Moderate Red
1
Ocean Optics
PI
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0
400
800
500
600
700
Wavelength (nm)
800
(a)
Purplish Blue
Orange
Reflectance
0.6
0.4
500
600
700
Wavelength (nm)
0.6
0.2
400
800
Purple
0.6
0.4
0.2
500
600
700
Wavelength (nm)
0
400
800
Yellow Green
1
1
0.8
0.8
0.8
0.6
0.4
0.2
0
400
0.6
0.4
0.2
500
600
700
Wavelength (nm)
800
0
400
500
600
700
Wavelength (nm)
800
Orange Yellow
1
Reflectance
Reflectance
0.8
0.8
0.4
0.2
TwIST
PI
1
Reflectance
Reflectance
0.8
0
400
Moderate Red
1
Reflectance
1
0.6
0.4
0.2
500
600
700
Wavelength (nm)
800
0
400
500
600
700
Wavelength (nm)
800
(b)
Fig. 16. (a) The reconstructed spectra compared with reference by an Ocean
Optics spectrometer. (b) The reconstructed spectra compared with reconstruction using TwIST.
Download