Scaling Limits in Computational Bayesian Inversion Claudia Schillings, Christoph Schwab

advertisement
Scaling Limits in Computational Bayesian Inversion
Claudia Schillings, Christoph Schwab
Warwick Centre for Predictive Modelling - 23. April 2015
research supported by the European Research Council under FP7 Grant AdG247277
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
1 / 18
Outline
1
Bayesian Inversion of Parametric Operator Equations
2
Sparsity of the Forward Solution
3
Sparsity of the Posterior
4
Sparse Quadrature
5
Numerical Results
6
Small Observation Noise Covariance
7
Summary
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
2 / 18
Inverse Problem
Physical Model
G(u) → δ
u parameter vector / parameter function
G the forward map modelling the physical process
δ result / observations
Forward Problem
Inverse Problem
Find the output δ for given
parameters u
Find the parameters u from
(noisy) observations δ
→ well-posed
→ ill-posed
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Physical Model
G(u) → δ
u parameter vector / parameter function
G the forward map modelling the physical process
δ result / observations
G forward response operator
Forward Problem
Inverse Problem
Find the output δ for given
parameters u
Find the parameters u from
(noisy) observations δ
→ well-posed
→ ill-posed
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Deterministic optimization problem
1
min kδ − G(u)k2 + R(u)
u 2
kδ − G(u)k potential / data misfit
R regularization term
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Deterministic optimization problem
1
min kδ − G(u)k2 + R(u)
u 2
Large-scale, deterministic optimization problem
No quantification of the uncertainty in the unknown u
Proper choice of the regularization term R
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Bayesian inverse problem
u, η, δ random variables / fields
Goal of computation: moments of system quantities under the posterior w.r. to
noisy data δ
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Bayesian inverse problem
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Bayesian inverse problem
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Bayesian inverse problem
Quantification of uncertainty in u and system quantities
Well-posedness of the inverse problem
Incorporation of prior knowledge on the uncertain data u
Need of efficient approximations of the posterior
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Inverse Problem
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η
Bayesian inverse problem
MCMC methods
MAP estimators
Ad hoc algorithms (EnKF, SMC, GP,...)
Sparse deterministic approximations of posterior expectations
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
3 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
Goal: Efficient estimation of system quantities from noisy observations
Infinite-dimensional parameter space
Fast convergence by exploiting sparsity
of the underlying forward problem
Suitable for application to a broad class
of forward problems
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
Goal: Efficient estimation of system quantities from noisy observations
UQ in Nano Optics
Infinite-dimensional parameter space
Fast convergence by exploiting sparsity
of the underlying forward problem
Suitable for application to a broad class
of forward problems
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
Goal: Efficient estimation of system quantities from noisy observations
UQ in Biochemical Networks
Infinite-dimensional parameter space
Fast convergence by exploiting sparsity
of the underlying forward problem
Suitable for application to a broad class
of forward problems
Source: Chen et al.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
Goal: Efficient estimation of system quantities from noisy observations
UQ in Groundwater Flow
Infinite-dimensional parameter space
1
0.9
Fast convergence by exploiting sparsity
of the underlying forward problem
0.8
0.7
0.6
0.5
Suitable for application to a broad class
of forward problems
0.4
0.3
0.2
y
0.1
1
C. Schillings (UoW)
Bayesian Inversion
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0x
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
X separable Banach space
G : X 7→ X the forward map
Abstract Operator Equation
Given u ∈ X, find q ∈ X :
A(u; q) = F(u) in Y 0
with A(u; ·) ∈ L(X , Y 0 ), F : X 7→ Y 0 , X , Y reflexive Banach spaces,
a(v, w) := Y hw, AviY 0 ∀v ∈ X , w ∈ Y corresponding bilinear form
O : X 7→ RK bounded, linear observation operator
G : X 7→ RK uncertainty-to-observation map, G = O ◦ G
η ∈ RK the observational noise (η ∼ N (0, Γ))
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
X separable Banach space
G : X 7→ X the forward map
Model Parametric Elliptic Problem
Given u ∈ X, find q ∈ X :
− ∇ · u∇q = f
in D,
q = 0 in ∂D
R
with weak formulation D u(x)∇q(x) · ∇w(x)dx = X hw, f iX 0 for all w ∈ X ,
X = Y = H01 (D), D ⊂ Rd bounded Lipschitz domain with Lipschitz boundary ∂D
O : X 7→ RK bounded, linear observation operator
G : X 7→ RK uncertainty-to-observation map, G = O ◦ G
η ∈ RK the observational noise (η ∼ N (0, Γ))
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Find the unknown data u ∈ X from noisy observations
δ = G(u) + η,
X separable Banach space
G : X 7→ X the forward map
O : X 7→ RK bounded, linear observation operator
G : X 7→ RK uncertainty-to-observation map, G = O ◦ G
η ∈ RK the observational noise (η ∼ N (0, Γ))
Least squares potential Φ : X × RK → R
1
Φ(u; δ) :=
(δ − G(u))> Γ−1 (δ − G(u))
2
Reformulation of the forward problem with unknown stochastic input
data as an infinite dimensional, parametric deterministic problem
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problems (Stuart 2010)
Parametric representation of the unknown u
X
u = u(y) := hui +
yj ψj ∈ X
j∈J
y = (yj )j∈J iid sequence of real-valued random variables yj ∼ U[−1, 1]
hui, ψj ∈ X
J finite or countably infinite index set
Prior measure on the uncertain input data
µ0 (dy) :=
O1
j∈J
2
λ1 (dyj ) .
N
1
(U, B) = [−1, 1]J ,
j∈J B [−1, 1] measurable space
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
4 / 18
Bayesian Inverse Problem
Theorem (Schwab and Stuart 2011)
Assume that G(u)
is bounded and continuous.
P
u=hui+ j∈J yj ψj
Then µδ (dy), the distribution of y ∈ U given δ, is absolutely continuous
with respect to µ0 (dy), and
dµδ
1
(y) = Θ(y)
dµ0
Z
with the parametric Bayesian posterior Θ given by
Θ(y) = exp −Φ(u; δ) ,
P
u=hui+
j∈J yj ψj
and the normalization constant
Z
Z=
Θ(y)µ0 (dy) .
U
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
5 / 18
Bayesian Inverse Problem
Expectation of a Quantity of Interest φ : X → S
Z
µδ
−1
E [φ(u)] = Z
exp −Φ(u; δ) φ(u)
P
u=hui+
U
with Z =
R
U
j∈J yj ψj
µ0 (dy) =: Z 0 /Z
exp(− 21 (δ − G(u))> Γ−1 (δ − G(u)) )µ0 (dy).
Reformulation of the forward problem with unknown stochastic input data
as an infinite dimensional, parametric deterministic problem
Parametric, deterministic representation of the derivative of the posterior
measure with respect to the prior µ0
Approximation of Z 0 and Z to compute the expectation of QoI under the
posterior given data δ
Efficient algorithm to approximate the conditional expectations given
the data with dimension-independent rates of convergence
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
5 / 18
Bayesian Inverse Problem
Expectation of a Quantity of Interest φ : X → S
Z
µδ
−1
E [φ(u)] = Z
exp −Φ(u; δ) φ(u)
P
u=hui+
U
with Z =
R
U
j∈J yj ψj
µ0 (dy) =: Z 0 /Z
exp(− 21 (δ − G(u))> Γ−1 (δ − G(u)) )µ0 (dy).
Exploiting sparsity in the parametric operator equation
Parameters belonging to a specified sparsity class
Analytic regularity of the parametric, deterministic Bayesian posterior
Parametric, deterministic Bayesian posterior belongs to the same
sparsity class
→ Sparsity of generalized pce + dimension-independent convergence
rates for Smolyak integration algorithms
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
5 / 18
Model Parametric Elliptic Problem
Z
u(x)∇q(x) · ∇w(x)dx =
X hw, f iX 0
for all w ∈ X
D
with D ⊂ Rd bounded Lipschitz domain with Lipschitz boundary ∂D, X = H01 (D), f ∈ L2 (D).
Structural Assumptions
u(x, y) = hui(x) +
X
yj ψj (x),
x∈X
j∈J
with hui ∈ L∞ (D) and (ψj )j∈J ⊂ L∞ (D), y = (y1 , y2 , . . . ) ∈ U = [−1, 1]J
Assumption 1 (UEA)
0 < umin ≤ u(x, y) ≤ umax < ∞
∀x ∈ D, y ∈ U
with 0 < umin ≤ umax < ∞.
Assumption 2
X
kψj kL∞ (D) ≤
j∈J
κ
hui
1 + κ min
with huimin = minx∈D hui(x) > 0 and κ > 0
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
6 / 18
Sparse Polynomial Chaos Approximations
Sparsity in the unknown coefficient function u,
i.e. if
P∞
j=1
kψj kpL∞ (D) < ∞ for 0 < p < 1,
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
7 / 18
Sparse Polynomial Chaos Approximations
Sparsity in the unknown coefficient function u,
P∞
j=1
kψj kpL∞ (D) < ∞ for 0 < p < 1,
(||ψ j||) j
(||ψ j||) j
(||ψ j||) j
(||ψ j||) j
1
0.9
0.8
summable, p=1
summable, p=1/2
summable, p=1/3
summable, p=1/4
0.7
0.6
||ψ j||
i.e. if
0.5
0.4
0.3
0.2
0.1
0
0
2
4
6
8
10
12
j
C. Schillings (UoW)
14
16
Bayesian Inversion
18
20
WCPM - 23. Apr. 2015
7 / 18
Sparse Polynomial Chaos Approximations
Sparsity in the unknown coefficient function u,
i.e. if
P∞
j=1
kψj kpL∞ (D) < ∞ for 0 < p < 1,
implies p-sparsity in the solution’s Taylor expansion,
∀y ∈ U :
q(y) =
P
C. Schillings (UoW)
ν∈F
τν yν ,
τν :=
1 ν
∂ q(y) |y=0
ν! y
Bayesian Inversion
,
F = {ν ∈ NN0 : |ν| < ∞}:
WCPM - 23. Apr. 2015
7 / 18
Sparse Polynomial Chaos Approximations
Sparsity in the unknown coefficient function u,
i.e. if
P∞
j=1
kψj kpL∞ (D) < ∞ for 0 < p < 1,
implies p-sparsity in the solution’s Taylor expansion,
∀y ∈ U :
q(y) =
P
ν∈F
τν yν ,
τν :=
1 ν
∂ q(y) |y=0
ν! y
,
F = {ν ∈ NN0 : |ν| < ∞}:
There exists a p-summable, monotone envelope t = {tν }ν∈F
i.e. tν := supµ≥ν kτν kX with C(p, q) := ktk`p (F ) < ∞ .
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
7 / 18
Sparse Polynomial Chaos Approximations
Sparsity in the unknown coefficient function u,
i.e. if
P∞
j=1
kψj kpL∞ (D) < ∞ for 0 < p < 1,
1
implies p-sparsity
in the solution’s Taylor expansion,
taylor coeff. (||τj||) j, 1D
∀y ∈ U :
0.9
P
q(y) =0.8 ν∈F τν yν ,
τν :=
1 ν
∂ q(y) |y=0
ν! y
monotone envelope (tj)j
F = {ν ∈ NN0 : |ν| < ∞}:
,
0.7
||τj||, t j
There exists a0.6
p-summable, monotone envelope t = {tν }ν∈F
0.5
i.e. tν := supµ≥ν 0.4
kτν kX with C(p, q) := ktk`p (F ) < ∞ .
0.3
0.2
0.1
0
0
2
4
6
8
10
12
14
j
C. Schillings (UoW)
Bayesian Inversion
16
18
20
WCPM - 23. Apr. 2015
7 / 18
Sparse Polynomial Chaos Approximations
Sparsity in the unknown coefficient function u,
i.e. if
P∞
j=1
kψj kpL∞ (D) < ∞ for 0 < p < 1,
implies p-sparsity in the solution’s Taylor expansion,
∀y ∈ U :
q(y) =
P
ν∈F
τν yν ,
τν :=
1 ν
∂ q(y) |y=0
ν! y
,
F = {ν ∈ NN0 : |ν| < ∞}:
There exists a p-summable, monotone envelope t = {tν }ν∈F
i.e. tν := supµ≥ν kτν kX with C(p, q) := ktk`p (F ) < ∞ .
and monotone ΛTN ⊂ F corresponding to the N largest terms of t with
X
sup q(y) −
τν yν ≤ C(p, t)N −(1/p−1) .
y∈U
ν∈ΛTN
X
A. Cohen, R. DeVore and Ch. Schwab. Analytic regularity and polynomial approximation of
parametric and stochastic elliptic PDEs, Analysis and Applications, 2010.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
7 / 18
Sparse Polynomial Chaos Approximations
Idea of Proof
Holomorphic Extension
z 7→ q(z) ,
z ∈ Uρ := ⊗j≥1 {zj ≤ ρj }
holomorph with bound kq(z)kX ≤ Cδ for any positive sequence ρ = {ρj }j≥1 such that
P
j≥1 ρj |ψj (x)| ≤ hui(x) − δ, x ∈ D for some δ > 0.
Estimate on the Taylor Coefficients
kτν kX ≤ Cδ inf{ρ−ν :
X
ρj |ψj (x)| ≤ hui(x) − δ, x ∈ D}
j≥1
by recursive application of Cauchy’s formula
Construction of a particular ρ
{kψj kL∞ }j≥1 ∈ `p (N) ⇒ {tν }ν∈F ∈ `p (F )
Stechkin
X
sup q(y) −
τν yν y∈U
C. Schillings (UoW)
ν∈ΛTN
X
Bayesian Inversion
≤
X
ktν kX ≤ CN
− p1 +1
ν ∈Λ
/ TN
WCPM - 23. Apr. 2015
7 / 18
Sparsity of the Posterior
Theorem (ClS and Schwab 2013)
Assume that the unknown coefficient function u is p-sparse for
0 < p < 1. Then, the posterior density’s Taylor expansion is p-sparse
with the same p.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
8 / 18
Sparsity of the Posterior
Theorem (ClS and Schwab 2013)
Assume that the unknown coefficient function u is p-sparse for
0 < p < 1. Then, the posterior density’s Taylor expansion is p-sparse
with the same p.
N-term Approximation Results
X
1
P
sup Θ(y) −
Θν Pν (y) ≤ N −s kθ P k`pm (F ) , s := − 1 .
p
X
y∈U
P
ν∈ΛN
Adaptive Smolyak quadrature algorithm with convergence rates
depending only on the summability of the parametric operator
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
8 / 18
Sparsity of the Posterior
Theorem (ClS and Schwab 2013)
Assume that the unknown coefficient function u is p-sparse for
0 < p < 1. Then, the posterior density’s Taylor expansion is p-sparse
with the same p.
Examples
Parametric initial value ODEs (Hansen & Schwab; Vietnam J. Math. 2013)
Affine-parametric, linear operator equations (ClS & Schwab; 2013)
Semilinear elliptic PDEs (Hansen & Schwab; Math. Nachr. 2013)
Elliptic multiscale problems (Hoang & Schwab; Analysis and Applications 2012)
Nonaffine holomorphic-parametric, nonlinear problems (Cohen, Chkifa &
Schwab; 2013)
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
8 / 18
Sparsity of the Posterior
Theorem (ClS and Schwab 2013)
Assume that the unknown coefficient function u is p-sparse for
0 < p < 1. Then, the posterior density’s Taylor expansion is p-sparse
with the same p.
Goal of computation
Expectation of a Quantity of Interest φ : X → S
Z
δ
Eµ [φ(u)] = Z −1
exp −Φ(u; δ) φ(u)
P
u=hui+
U
with Z =
R
U
j∈J yj ψj
µ0 (dy) =: Z 0 /Z
exp(− 21 (δ − G(u))> Γ−1 (δ − G(u)) )µ0 (dy).
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
8 / 18
Sparse Quadrature Operator
For any finite monotone set Λ ⊂ F, the quadrature operator is defined
by
X
XO
QΛ =
∆ν =
∆νj
ν∈Λ
ν∈Λ j≥1
with associated collocation grid
ZΛ = ∪ν∈Λ Z ν .
(Qk )k≥0 sequence of univariate quadrature formulas
∆j = Qj − Qj−1 , j ≥ 0 univariate quadrature difference operator
N
N
Qν = j≥1 Qνj , ∆ν = j≥1 ∆νj tensorized multivariate operators
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
9 / 18
Sparse Quadrature Operator
For any finite monotone set Λ ⊂ F, the quadrature operator is defined
by
X
XO
QΛ =
∆ν =
∆νj
ν∈Λ
ν∈Λ j≥1
with associated collocation grid
ZΛ = ∪ν∈Λ Z ν .
1
1
0
0
−1
−1
C. Schillings (UoW)
0
1
Bayesian Inversion
−1
−1
0
1
WCPM - 23. Apr. 2015
9 / 18
Convergence Rates for Adaptive Smolyak Integration
Theorem
Assume that the unknown coefficient function u is p-sparse for 0 < p < 1.
Then there exists a sequence (ΛN )N≥1 of monotone index sets ΛN ⊂ F
such that #ΛN ≤ N and
|Z − QΛN [Θ]| ≤ C1 N
R
with Z = U Θ(y)µ0 (dy), Θ(y) = exp −Φ(u; δ) − 1p +1
and,
P
u=hui+
kZ 0 − QΛN [Ψ]kX ≤ C2 N
R
with Z 0 = U Ψ(y)µ0 (dy), Ψ(y) = Θ(y)φ(u)
,
P
u=hui+
yψ
j∈J j j
− p1 +1
,
C1 , C2 > 0 independent of N.
yψ
j∈J j j
Remark: SAME index sets ΛN for BOTH, Z 0 and Z.
ClS and Schwab Sparsity in Bayesian Inversion of Parametric Operator Equations, 2013.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
10 / 18
Uncertainty Quantification in Nano Optics
Goal: Quantification of the influence of defects in fabrication process
on the optical response of nano structures
Propagation of plane wave and its interaction with
scatterer described by Helmholtz equation (2D).
Stochastic shape of the scatterer
0 < ρmin ≤ ρ(ω, φ) ≤ ρmax ,
ω ∈ Ω, φ ∈ [0, 2π)
Collaboration with Ralf Hiptmair, Laura Scarabosio
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
11 / 18
Uncertainty Quantification in Nano Optics
Forward Problem
−∆u − k12 u = 0
−∆u −
k22 u
=0
in D1 (ω)
in D2 (ω)
u|+ − u|− = 0 on Γ(ω)
∂u ∂u −µ
= 0 on Γ(ω)
∂n +
∂n −
p
∂
lim
|x|
− ik1 (u(ω) − ui )(x) = 0,
|x|→∞
∂ |x|
ui (x) = eik1 ·x
Parametrization of the shape
r(ω, ϕ) = r0 (ϕ) +
C. Schillings (UoW)
64
X
1
1
y (ω) cos(kϕ) + ζ y2k+1 (ω) sin(kϕ),
ζ 2k
k
k
k=1
Bayesian Inversion
ϕ ∈ [0, 2π)
WCPM - 23. Apr. 2015
11 / 18
Numerical Results of Sparse Interpolation
Interpolation, dimension 64
1
10
Estimated error CC, nonadaptive strategy, zeta=2
Estimated error CC, nonadaptive strategy, zeta=3
Estimated error CC, nonadaptive strategy, zeta=4
0
10
−1
10
−2
10
−3
10
−4
10
−5
10
−6
10
0
10
1
2
10
10
3
10
#Λ
Figure: Estimated error curves of the interpolation error w.r. to the cardinality of the
index set ΛN based on the sequences CC with nonadaptive refinement, uniform
distribution of the parameters, dimension of the parameter space 64.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
12 / 18
Concentration Effect of the Posterior for 0 < Γ 1
Model parametric elliptic problem
−div(u∇q) = f
in D := [0, 1] , q = 0
in ∂D ,
with f (x) = 100 · x and
u(x, y) = 0.15 + y1 ψ1 + y2 ψ2 ,
with J = 2, J = {1, 2}, ψ1 (x) = 0.1 sin(πx), ψ2 (x) = 0.025 cos(2πx) and with
yj ∼ U [−1, 1], j ∈ J .
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
13 / 18
Concentration Effect of the Posterior for 0 < Γ 1
For given (noisy) data δ,
δ = G(u) + η ,
we are interested in the behavior of the posterior
Θ(y) = exp −ΦΓobs (u; δ) P2
u=
with
ΦΓobs (u; δ) =
j=1 yj ψj
,
1
(δ − G(u))> Γ−1 (δ − G(u)) .
2
and variance of the noise Γ = λ · id
QoI φ is the solution of the forward problem at x = 0.5.
Observation operator O consists of 2 system responses at x = 0.25 and x = 0.75.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
13 / 18
Concentration Effect of the Posterior for 0 < Γ 1
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, 6=0.25
1
2
reference parameter
posterior
0.8
0.6
0.4
y2
0.2
0
-0.2
-0.4
-0.6
-0.8
-1
-1
-0.5
0
0.5
1
y1
Figure: Contour plot of the posterior with observational noise λ = 0.252 .
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
13 / 18
Concentration Effect of the Posterior for 0 < Γ 1
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, 6=0.5
1
2
reference parameter
posterior
0.8
0.6
0.4
y2
0.2
0
-0.2
-0.4
-0.6
-0.8
-1
-1
-0.5
0
0.5
1
y1
Figure: Contour plot of the posterior with observational noise λ = 0.52 .
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
13 / 18
Concentration Effect of the Posterior for 0 < Γ 1
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, 6=0.05
1
2
reference parameter
posterior
0.8
0.6
0.4
y2
0.2
0
-0.2
-0.4
-0.6
-0.8
-1
-1
-0.5
0
0.5
1
y1
Figure: Contour plot of the posterior with observational noise λ = 0.052 .
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
13 / 18
Asymptotic Analysis
Theorem (ClS and Schwab 2014)
Assume that the parameter space is finite (possibly after dimension truncation) and G(·), δ are such that the assumptions of Laplace’s
method hold; in particular, the minimum y0 of
S(y) =
1
(δ − G(u))> (δ − G(u))
2
is nondegenerate.
Then, as Γ ↓ 0, the Bayesian estimate admits an asymptotic expansion
δ
Eµ [φ] =
ZΓ0
∼ a0 + a1 λ + a2 λ2 + ....
ZΓ
where a0 = φ(y0 ).
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
14 / 18
Curvature Rescaling Regularization
Removing the degeneracy in the integrand function
The maximizer y0 of the posterior measure Θ(y) is computed by
minimizing the potential
1
(δ − G(u))> (δ − G(u)) P2
.
2
u= j=1 yj ψj
using a trust-region Quasi-Newton approach with SR1 updates.
Diagonalize the approximated Hessian HSR1 = QMQ> and regularize the
integrand by the curvature rescaling transformation
y0 + Γ1/2 QM −1/2 z ,
C. Schillings (UoW)
Bayesian Inversion
z ∈ RJ .
WCPM - 23. Apr. 2015
15 / 18
Numerical Experiments
Curvature rescaling
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, 6=0.25 2
1
0.8
transformed posterior
0.8
0.6
0.6
0.4
0.4
0.2
0.2
y2
y2
Contour plot of the transf. posterior, 2 meas. at 0.25 and 0.75, 6=0.25 2
1
reference parameter
posterior
optimal point
0
0
-0.2
-0.2
-0.4
-0.4
-0.6
-0.6
-0.8
-1
-1
-0.8
-0.5
0
y1
0.5
1
-1
-1
-0.5
0
0.5
1
y1
Figure: Contour plot of the posterior with observational noise Γobs = 0.252 · Id (left)
and contour plot of the transformed posterior (right).
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
16 / 18
Numerical Experiments
Curvature rescaling
Truncated domain, 2 meas. at 0.25 and 0.75, 6=0.25 2
1
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, 6=0.25 2
1
reference parameter
posterior
optimal point
0.6
0.6
0.4
0.4
0.2
0.2
0
0
-0.2
-0.2
-0.4
-0.4
-0.6
-0.6
-0.8
-1
-1
transformed posterior
truncated domain
0.8
y2
y2
0.8
-0.8
-0.5
0
y1
0.5
1
-1
-1
-0.5
0
0.5
1
y1
Figure: Contour plot of the posterior with observational noise Γobs = 0.252 · Id (left)
and and truncated domain of integration of the rescaled Smolyak approach in the
original coordinate system (right).
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
16 / 18
Numerical Experiments
Curvature rescaling
2
2
10
orig. integrand
trans. integrand
0
10
−2
−2
10
10
−4
10
−6
10
−8
10
−10
10
−4
10
−6
10
−8
10
−10
10
−12
−12
10
10
−14
10
orig. integrand
trans. integrand
0
10
Estimated (absolute) error (Z’)
Estimated (absolute) error (normalization constant Z)
10
−14
0
10
1
2
10
10
3
10
10
0
10
#Λ
1
2
10
10
3
10
#Λ
Figure: Comparison of the estimated (absolute) error curves using the Smolyak
approach for the original integrand (gray) and the transformed integrand (black) for the
computation of ZΓobs (left) and ZΓ0 obs (right) with observational noise Γobs = 0.252 · Id.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
16 / 18
Numerical Experiments
Curvature rescaling
Extension to lognormal case
ln(u(x, y)) = 0.15 + y1 ψ1 + y2 ψ2 ,
with J = 2, J = {1, 2}, ψ1 (x) = 0.1 sin(πx), ψ2 (x) = 0.025 cos(2πx) and with
yj ∼ N (0, 1), j ∈ J.
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
16 / 18
Numerical Experiments
Curvature rescaling
Extension to lognormal case
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, , 6=0.01 2
3
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, 6=1.0001
3
reference parameter
posterior
optimal point
2
1
y2
y2
1
0
0
-1
-1
-2
-2
-3
-3
reference parameter
posterior
optimal point
2
-2
-1
0
1
2
3
-3
-3
-2
y1
-1
0
1
2
3
y1
Figure: Contour plot of the posterior density with observational noise Γobs = 0.012 · Id
(left), and contour plot of the transformed posterior (right).
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
16 / 18
Numerical Experiments
Curvature rescaling
Extension to lognormal case
Contour plot of the posterior, 2 meas. at 0.25 and 0.75, , 6=0.01 2
3
Contour plot of the transformed posterior, 2 meas. at 0.25 and 0.75, 6=0.0001
6
reference parameter
posterior
optimal point
2
transformed posterior
quadrature points
4
2
y2
y2
1
0
0
-1
-2
-2
-4
-3
-3
-2
-1
0
1
2
3
-6
-6
-4
y1
-2
0
2
4
6
y1
Figure: Contour plot of the posterior density with observational noise Γobs = 0.012 · Id
(left), and contour plot of the transformed posterior (right).
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
16 / 18
Conclusions and Outlook
New class of sparse, adaptive quadrature methods for Bayesian
inverse problems for a broad class of operator equations
Dimension-independent convergence rates depending only on the
summability of the parametric operator
Numerical confirmation of the predicted convergence rates
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
17 / 18
Conclusions and Outlook
New class of sparse, adaptive quadrature methods for Bayesian
inverse problems for a broad class of operator equations
Dimension-independent convergence rates depending only on the
summability of the parametric operator
Numerical confirmation of the predicted convergence rates
Efficient treatment of small observation noise variance Γ
Development of preconditioning techniques to overcome the
convergence problems in the case Γ ↓ 0
Combination of optimization and sampling techniques
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
17 / 18
References
Chkifa A 2013 On the Lebesgue constant of Leja sequences for the unit disk Journal of Approximation Theory 166
176-200
Chkifa A, Cohen A and Schwab C 2013 High-dimensional adaptive sparse polynomial interpolation and applications to
parametric PDEs Foundations of Computational Mathematics 1-33
Cohen A, DeVore R and Schwab C 2011 Analytic regularity and polynomial approximation of parametric and stochastic
elliptic PDEs Analysis and Applications 9 1-37
Gerstner T and Griebel M 2003 Dimension-adaptive tensor-product quadrature Computing 71 65-87
Hansen M and Schwab C 2013 Sparse adaptive approximation of high dimensional parametric initial value problems
Vietnam J. Math. 1 1-35
Hansen M, Schillings C and Schwab C 2012 Sparse approximation algorithms for high dimensional parametric initial value
problems Proceedings 5th Conf. High Performance Computing, Hanoi
Schillings C and Schwab C 2013 Sparse, adaptive Smolyak algorithms for Bayesian inverse problems Inverse Problems
29 065011
Schillings C and Schwab C 2014 Sparsity in Bayesian inversion of parametric operator equations Inverse Problems 30
065007
Schillings C 2013 A Note on Sparse, Adaptive Smolyak Quadratures for Bayesian Inverse Problems Oberwolfach Report
10/1 239-293
Schillings C and Schwab C 2014 Scaling Limits in Computational Bayesian Inversion SAM Report 2014-26
Schwab C and Stuart A M 2012 Sparse deterministic approximation of Bayesian inverse problems Inverse Problems 28
045003
C. Schillings (UoW)
Bayesian Inversion
WCPM - 23. Apr. 2015
18 / 18
Download