Predictive Coarse-Graining M. Sch¨ oberl , N. Zabaras

advertisement
Predictive Coarse-Graining
M. Schöberl1 , N. Zabaras2,3 , P.S. Koutsourelakis1
1
Continuum Mechanics Group
Technische Universität München
3
2
Institute for Advanced Study
Technische Universität München
Warwick Centre of Predictive Modeling
University of Warwick
November 27, 2015
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
1 / 56
Motivation Coarse-Graining
Atomistic simulation for obtaining insights of
chemical and physical process of complex systems.
Difficulty
I
Complex interactions
I
Long-range interactions
I
Small time- and length-scales
→ Exceeding computational tractability
A coarse description allows us
I
to evaluate larger systems during larger time
intervals
I
to gain understanding of physics of the system.
WCPM Seminar
Predictive Coarse-Graining
Figure :
Coarse-graining water
November 27, 2015
2 / 56
Motivation
Approach
I
Describing system with less degrees of freedom
I
Determining optimal parameter set for given parametrization of a
coarse-grained potential
I
Leading to point estimates of its parametrization and thus also in predictions
I
Predictions performed on coarse scale
How can we quantify the uncertainty induced by a coarse
description and the loss of information?
How can we reconstruct fine configurations given a coarse
configuration?
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
3 / 56
General coarse-graining problem
Fine-scale degrees of freedom x ∈ M with M ⊂ Rn , n 1 in equilibrium
described by a Boltzmann-type PDF:
Fine-scale description
pf (x|β) =
I
Potential U(x)
I
Inverse Temperature β =
1
kb T
Partition function Z (β) =
R
I
WCPM Seminar
exp {−βU(x)}
Z (β)
, with temperature T
M
exp {−βU(x)} dx
Predictive Coarse-Graining
November 27, 2015
4 / 56
General coarse-graining problem
Coarse-scale description
pc (X|β) =
exp {−βUc (X, θ c )}
Zc (θ c , β)
I
Coarse variables X ∈ Mc , Mc ⊂ Rnc , nc n
I
Coarse-grained potential selected Ûc selected out of candidate potentials
Ûc ∈ Uc
I
Parametrization and shape selected by assuming Uc (X, θ c ) but any potential
possible
How are fine and coarse configurations connected?
Connecting fine variables x with coarse variables X with coarse-graining map
ξ : M −→ Mc
X = ξ(x)
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
5 / 56
How to determine the effective coarse potential Uc ?
Analytical solution for the effective coarse-grained potential:
Uc (X) = −β −1 ln
Z
exp {−βU(x)} δ(ξ(x) − X)dx
M
Potential of mean force with respect to coarse-grained coordinates ξ(x). →
intractable!
Numerical coarse-graining strategies:
I
iterative Boltzmann inversion [5]
I
inverse Monte Carlo [8]
I
force matching [6]
I
variational mean-field theory [9]
I
relative entropy [10]
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
6 / 56
The relative entropy method[10]
→ minimize a distance metric between pf (X|β) and pc (X|β) with the PDF
pf (X|β) sampling configuration x that maps to configuration X
Z
pf (X|β) ∝
pf (x|β)δ(ξ(x) − X)dx
M
Relative entropy method[10, 4, 2]
(or KL-divergence)
Z
KL(pf ||pc ) =
pf (X|β) ln
Mc
pf (X|β)
pc (X|β)
dX
Formulation in M:
Z
KL(pf ||pc ) =
pf (x|β) ln
M
I
pf (x|β)
pc (ξ(x)|β)
dx + Smap (ξ(x))
R
with Smap = M δ(ξ(x) − X)dx p (x|β) measures the information loss
f
induced by mapping ξ(x).
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
7 / 56
Relative entropy method
Smap (ξ(x)) is independent of Ûc , the minimization of the KL-divergence is
equivalent to minimization of
pf (x|β)
F(Ûc ) = ln
pc (ξ(x)|β) pf (x|β)
For any given m-dimensional parametrization θ c ∈ Θc ⊂ Rm of the potential, the
optimization problem follows:
θ ∗c = arg min F(Uc (X, θ c ))
θ c ∈Θc
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
8 / 56
Relative entropy method
Model Restrictions
I
Relative entropy method has no capability to predict fine-scale configurations
(no mapping coarse to fine)
I
I
Eg. for predicting correlations within fine configurations
Predicting quantities of interest (QoI) by evaluating on fine-scale
Z
hf (x)ipf (x|β) =
f (x)pf (x|β)dx
M
I
With coarse-grained description pc (X|β) predictions in Mc but not in M
Difficulty:
I
I
I
How to define observable g (X) for desired QoI in Mc ? R
Are you able to predict same QoI with hg (X)ipc (X|β) = Mc g (X)pc (X|β)dX
No quantification of epistemic uncertainty: determine Smap (ξ(x))
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
9 / 56
Proposed Model
Framework following generative probabilistic models and builds upon:
1
Coarse-scale PDF pc (X|θ c , β) ∝ exp {−βUc (X, θ c )}
I
I
2
Probabilistic mapping from coarse to fine pcf (x|X, θ cf )
I
Conditional PDF of x given the coarse variables X
Parametrization of the probabilistic mapping by θ cf ∈ Θcf ⊂ Rmcf
Deterministic mapping ξ(x) is not invertible (many to one map)
I
→ Probabilistic relation necessary
Smap not fixed for per definition. → Optimization with respect to θ cf possible.
I
I
3
Describing statistics of coarse variables X
Parametrized by θ c , no restrictions in shape of pc
Prior PDF for model parametrization
I
I
p(θ c )
p(θ cf )
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
10 / 56
Proposed Model
I
Proposed model describes a Bayesian perspective of coarse-graining
equilibrium atomistic ensembles
I
Model is data driven by N data points x(i) in the fine-scale, denoted as
x(1:N) = {x(1) , · · · , x(N) }
Given the data x(1:N) a posterior distribution on the model parameters θ cf , θ c and
the latent variables X(1:N) can be defined.
I
Coarse variables seen as latent
I
To each observable x(i) one latent Variable X(i) is assigned
I
X(i) represents pre-image of the fine-scale observation x(i)
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
11 / 56
Bayesian coarse-graining - posterior
Posterior distribution
p(θ cf , θ cf , X(1:N) |x(1:N) ) ∝ p(x(1:N) |θ cf , θ c , X(1:N) )p(θ cf , θ c , X(1:N) )
!
N
Y
(i)
(i)
(i)
pcf (x |θ cf , X )pc (X |θ c ) p(θ cf )p(θ c )
=
i=1
Use given data x(1:N) for inferring the the posterior
distribution or for determining MAP estimates of the
model parameters θ ∗cf , θ ∗c .
I
θc
θ cf
X(n)
x(n)
Bayesian formulation answers the questions of
I
I
model validation and
prediction
N
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
12 / 56
Bayesian coarse-graining
The posterior distribution leads to the predictive distribution for fine-scale
configurations x:
Predictive distribution
p(x|x(1:N) ) =
Z
pcf (x|X, θ cf )pc (X|θ c )p(θ c , θ cf |x(1:N) )dXdθ cf dθ cf
Simulate fine-scale system
2
Draw sample θ ∗c , θ ∗cf ∼ p(θ c , θ cf |x(1:N) )
Draw sample of coarse-scale description X∗ ∼ pc (X|θ ∗c )
3
Draw sample of fine-scale description x∗ ∼ pcf (x∗ |X∗ , θ ∗cf )
1
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
13 / 56
Bayesian coarse-graining: ensemble averages
Approximating ensemble averages by
Z
hf (x)ipf (x) =
f (x)pf (x)dx
M
= E [f (x)]≈E [f (x)|x(1:N) ]
Z
=
f (x)p(x|x(1:N) )dx
M
Z
= f (x)pcf (x|X, θ cf )pc (X|θ c )p(θ c , θ cf |x(1:N) )dxdXdθ c dθ cf
I
Error arising from discrepancy between E [f (x)] and E [f (x)|x(1:N) ]
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
14 / 56
Bayesian coarse-graining: coarse-scale model pc
Coarse-scale model
pc (X|θ cf ) =
exp {−βUc (X, θ c )}
Zc (θ c )
Requirements
I
Allow sufficient flexibility of pc (X)
I
Uc built from high-order interactions affording flexibility in pc (X)
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
15 / 56
Bayesian coarse-graining: reconstruction map pcf
Map from coarse to fine: pcf (x|X, θ cf )
I
Several fine-configurations map to same coarse-configuration
I
I
Probabilistic relation between a given coarse-configuration and a
fine-configuration takes account of it. [7]
Fast reconstruction ([7, 1]) of fine-scale configurations x given a
coarse-configuration X desired.
coarse coordinate X
Influence of pcf
I
What is expected from the coarse variables X in
terms of predicting the given data x(1:N) ?
I
Adjusting pc (X|θ c ) so that X agrees best with data
x(1:N) connected with probabilistic mapping pcf .
WCPM Seminar
Predictive Coarse-Graining
pcf (xi|X)
(...)
possible set of fine coordinates xi for given coarse coordinate X
Figure : Configurations x
given X
November 27, 2015
16 / 56
Optimization
For learning the model parameters θ c and θ cf the expectation maximization
algorithm is applied.
I
Point estimates for maximizing the likelihood or posterior distribution are
obtained.
I
Approximation of posterior on θ cf by Laplace (validated by full posterior
sampling)
MAP estimate
h
i
θ MAP = arg max log p(θ c , θ cf , x(1:N) )
θ
h
i
= arg max log p(x(1:N) |θ cf , θ cf ) + log p(θ c ) + log p(θ cf )
θ
I
In the following the ML-estimate is shown and will be extended by prior
distributions p(θ) in a next step by adding log-priors log p(θ c ) + log p(θ cf ).
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
17 / 56
Optimization
I
I
To obtain point estimates with MLE for the model parameters θc and θcf we
maximize the log-likelihood by using the EM algorithm.
The likelihood is defined as
p(x
(1:N)
Z
|θ c , θ cf ) =
p(x
(1:N)
Z
=
pcf (x
Z Y
N
=
,X
(1:N)
(1:N)
|X
|θ c , θ cf )dX
(1:N)
(i)
, θ cf )pc (X
(i)
(1:N)
(1:N)
|θ c )dX
(i)
pcf (x |X , θ cf )pc (X |θ c )dX
(1:N)
(1:N)
i=1
I
augmenting the log-likelihood with an arbitrary density q(X(1:N) )
log p(x
(1:N)
Z
|θ c , θ cf ) = log
q(X
Z
≥
q(X
= L(q(X
WCPM Seminar
(1:N)
(1:N)
(1:N)
)
) log
pcf (x(1:N) |X(1:N) , θ cf )pc (X(1:N) |θ c )
q(X(1:N) )
dX
pcf (x(1:N) |X(1:N) , θ cf )pc (X(1:N) |θ c )
q(X(1:N) )
(1:N)
dX
(1:N)
), θ c , θ cf )
Predictive Coarse-Graining
November 27, 2015
18 / 56
Expectation maximization
It holds the following decomposition for the lower bound:
L(q(X
(1:N)
Z
), θ c , θ cf ) =
q(X
(1:N)
= − KL(q(X
) log
(1:N)
p(x(1:N) |X(1:N) , θ cf )p(X(1:N) |θ c )
)||p(X
q(X(1:N) )
(1:N)
(1:N)
|x
dX
, θ cf , θ cf )) + log p(x
(1:N)
)
Decomposition of the log likelihood
log p(x
(1:N)
|θ c , θ cf ) = L(q(X
(1:N)
), θ c , θ cf ) + KL(q(X
(1:N)
)||p(X
(1:N)
|x
(1:N)
, θ c , θ cf )).
Maximizing L(q(X(1:N) ), θ c , θ cf ) is equal to minimizing
KL(q(X(1:N) )||p(X(1:N) |x(1:N) , θ cf , θ cf ))
p(X(1:N) |x(1:N) , θ cf , θ cf ) PDF over pre-images for X for the given data x(1:N) and
parameters θ.
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
19 / 56
Expectation maximization
For given parameters θ cf , θ c the lower bound is maximized if we choose
q(X(1:N) ) ∝ pcf (x(1:N) |X(1:N) , θ cf )pc (X(1:N) |θ c ).
Therefore,
⇒ KL(q(X(1:N) )||p(X(1:N) |x(1:N) , θ c , θ cf )) = 0.
Expectation maximization
1
2
Initial parameter setting for iteration t = 0: θ 0cf and θ 0c
E-step: q (t+1) = arg maxq L(q, θ t )
Involves MCMC to obtain samples X ∼ q(X(1:N) |x(1:N) , θ tcf , θ tcf )
3
M-step: θ t+1 = arg maxθ L(q t+1 , θ t )
4
t →t +1
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
20 / 56
Derivatives
First order derivatives


+
*
N
X
∂L
∂Uc (X, θ c )
∂Uc (X(i) , θ c )
= β N
−
∂θ c
∂θ c
∂θ c
pc (X(i) |θ t )
i=1
∂L
=
∂θ cf
*
1
pcf
(x(1:N) |X, θ
∂pcf (x(1:N) |X, θ cf )
∂θ cf
cf )

q(X(i) |x(i) ,θ t ,θ tc )
cf
c
+
q(X(1:N) |x(1:N) ,θ t ,θ tc )
cf
Second order derivatives
N
X
∂2L
=−β
∂θk ∂θl
i=1
*
∂ 2 Uc (X(i) , θ c )
∂θk ∂θl
+
*
+ βN
q(X(i) |x(i) ,θ 0 ,θ 0c )
cf
∂ 2 Uc (X, θ c )
∂θk ∂θl
+
pc (X|θ c )
∂Uc (X, θ c ) ∂Uc (X, θ c )
2
−β N
∂θk
∂θl
pc (X|θ c )
∂Uc (X, θ c )
∂Uc (X, θ c )
2
+β N
∂θk
∂θ
l
pc (X|θ c )
pc (X|θ c )
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
21 / 56
Derivatives
Second order derivatives
N
X
∂ 2 L(q(X(1:N) ), θ c , θ cf )
=
∂θ 2cf
i=1
WCPM Seminar
*
∂ 2 pcf (x(i) |X(i) , θ cf )
1
−
∂θ 2cf
pcf (x(i) |X(i) , θ cf )
pcf (x(i) |X(i) , θ cf )2
1
Predictive Coarse-Graining
∂pcf (x(i) |X(i) , θ cf )
∂θ cf
November 27, 2015
!
22 / 56
Robbins-Monro Optimization
Gradients are sample averages → noise afflicted.
Robbins-Monro Algorithm
θ t+1 = θ t + αt ∇θ L(θ t )
with
αt =
α
1
, with < ρ ≤ 1.
ρ
(t + A)
2
I
Convergence is guaranteed to the true optimum θ ∗ if infinite steps performed.
I
αt is a sequence of real numbers and has to fulfill:
∞
X
t=1
WCPM Seminar
αt = ∞ and
∞
X
αt2 < ∞
t=1
Predictive Coarse-Graining
November 27, 2015
23 / 56
Laplace Approximation
Approximating a density function p(z) =
1
Zf
(z) by
q(z) = N (z|z0 , Σ)
−1
with Σ = (−∇∇f (z)|z=0 )
and ∇f (z)|z=z0 = 0.
Figure : Laplace approximation[3]
For the given problem using a noniformative prior it follows
Laplace approximation
N
X
∂ 2 log p(θ cf |x(1:N) )
=
2
∂θ cf
i
WCPM Seminar
*
∂ 2 log pcf (x(i) |X(i) , θ cf )
∂θ 2cf
Predictive Coarse-Graining
+
q(X(i) )
November 27, 2015
24 / 56
Predictive uncertainty
Propagate uncertainty induced by mapping
1
−1
2
∂ log p(θ cf |x(1:N) )
θ icf ∼ N (θ MAP
,
Σ)
with
Σ
=
−
2
cf
∂θ
2
Xij ∼ pc (X|θ MAP
)
c
3
xijk ∼ pcf (x|Xij , θ icf )
cf
Predictice uncertainty of properties:
D
WCPM Seminar
E
f |θc ,θicf = =
Z
f (x)pcf (x|X, θ icf )pc (X|θ MAP
)dXdx
c
Predictive Coarse-Graining
November 27, 2015
25 / 56
Example Problem: Coarse-graining 1D Ising Model
I
Applying proposed method (’predCg’) for coarse-graining a one dimensional
Ising Model
I
Comparison between ’predCg’ and the deterministic formulation of
minimizing the KL-divergence (’relEntr’) by Shell (2008)
I
Assessed by predictive capabilities for magnetization and correlation
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
26 / 56
Ising Model - System Setting
The Ising-model regarded in the following discussion is one dimensional. It has the
properties:
I
System size nf in fine- and nc in coarse-scale
I
Level of coarse-graining lc =
I
Inverse temperature β
I
I
External field µ
Regarded interaction length Lf in fine- and Lc in coarse scale, including all
interactions k ≤ L
I
J0 = 1 overall interaction strength
WCPM Seminar
nf
nc
Predictive Coarse-Graining
November 27, 2015
27 / 56
Ising Model - Potential in Fine-Scale
Fine variables take the values xi ∈ {−1, 1} following
pf (x|β) ∝ exp(−βU(x, Jk , µ)):
Fine-scale potential
U(x, Jk , µ) = −
Lf
nf
X
X
1X
Jk
xi xj − µ
xi
2
k=1
|i−j|=k
i=1
with i, j ∈ {1, . . . , nf } having nf lattice sites.
I Maximal interactions of Lf sites apart are regarded in the potential.
I |i − j| = k interpreted as summation over neighbors k over all sites i sites
apart
I Jk , strength of the k-th interaction.
with Jk following a power law for a given overall strength J0 and exponent a,
Jk = LkK a with,
L
X
K = J0 L1−a
k −a
k=1
in order to normalize the interaction strength [1].
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
28 / 56
Ising Model - Potential in Coarse-Scale
Coarse variables take the values Xi ∈ {−1, 1}, pc ∝ exp{−βUc (X, θ c , µ)}
Coarse-scale potential
n
Uc (X, θ c , µ) = −
+
c
1 lin X
(θ
Xi
2
i=1
Lc
X
twop
θk
nc
X
trip
θmn
nc
X
XX
xi xi±m xi±m±n )
m=1 n=1
i=1
−µ
Xi Xj
|i−j|=k
k=1
+
X
Xi
i=1
with i, j ∈ {1, . . . , nc } and nc nf lattice sites.
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
29 / 56
Ising Model - Mapping Fine to Coarse (for relative entropy)
Each coarse variable Xr , r = {1, . . . , R} describes S fine-scale variables xr ,s with
s = {1, . . . , S}. The mapping from fine-scale variables xr ,s to coarse variable Xr is
defined as,
Fine-to-coarse mapping

PS
1

+1,
S Ps xr ,s > 0
S
1
Xr = −1,
s xr ,s < 0 .
S


U(−1, +1), otherwise
Figure : Mapping from fine-scale x to coarse-scale X
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
30 / 56
Ising Model - Mapping Coarse to Fine
Coarse-to-fine mapping pcf (x|X, θcf )
1+xr ,s Xr
2
pcf (xr ,s |Xr ) = θcf
(1 − θcf )
1+xr ,s Xr
2
Assumption: xi conditionally
independent for given X:
p(x
(1:N)
|X
(1:N)
, θ cf ) =
N
Y
(t)
(t)
pcf (x |X , θ cf )
i=1
=
N Y
R Y
S
Y

1+xri ,s Xri
2
cf
θ
(1 − θcf )
1−xri ,s Xri
2


i=1 r =1 s=1
1+xri ,s Xri
PN PR PS
r =1
s=1
i=1
2
= θcf
(1 − θcf )
WCPM Seminar
1−xri ,s Xri
PN PR PS
r =1
s=1
i=1
2
Figure : Probabilistic mapping form coarse to
fine: pcf (x|X, θ cf )
Predictive Coarse-Graining
November 27, 2015
31 / 56
Which properties to predict?
The magnetization m is given:
Magnetization in the fine-scale
Z
mfine =
n
1X
xi pf (x)dx
n
i
Magnetization calculated with proposed method (’predcg ’)
Z
mML,fine =
n
1X
xi pcf (x|X, θ cf )pc (X|θ c )p(θ cf , θ c |x(1:N) ) dθ cf dθ c dXdx
n
i
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
32 / 56
Quantity of Interest: Magnetization
Magnetization calculated with relative entropy method ‘relEntr ’
(deterministic mapping fine to coarse: X = ξ(x))
I
I
For each µ calculate optimal model parameter θ ∗cg, det
Since we want to compare the magnetization on the same (fine) scale a
mapping is introduced
I
I
I
if Xr = −1, select randomly a possible configuration xr ,s where either all
xr ,s = −1 or one is xr ,s = −1 and the other xr ,s = 1 (for given r : s ∈ {1 . . . lc })
if Xr = +1, select randomly a possible configuration xr ,s where either all
xr ,s = +1 or one is xr ,s = +1 and the other xr ,s = −1 (for given r : s ∈ 1, 2)
For a given coarse-state the selected fine-states xdet are used for calculating
the magnetization
Relative error in magnetization,
errmag =
WCPM Seminar
km − mtruth k
.
kmtruth k
Predictive Coarse-Graining
November 27, 2015
33 / 56
Quantity of Interest: Correlation
The correlation Rl of all sites being separated by l sites is measured as,
Rl =
n
1 X
< xi xj > .
n
|i−j|=l
Relative correlation error,
errcorr =
WCPM Seminar
kR − Rtruth k
.
kRtruth k
Predictive Coarse-Graining
November 27, 2015
34 / 56
Ising Model - ML Algorithm
1:
2:
3:
4:
5:
6:
7:
7:
8:
8:
9:
10:
11:
12:
13:
14:
14:
14:
for µ = µmin to µmax do
set initial fine-scale parameter: J0 , nf , µ
create data set: x(1:Sf ) ∼ p(x(1:Sf ) |J0 , µ, β) with Sf samples
n
set initial coarse-graining parameter: θ c 0 , θ cf0 , nc = f , µ
lc
θ tc ← θ c 0
t
θ cf ← θ cf0
while !(Convergence check passed) do
function E-Step
X ∼ q(X|x(1:N) , θ tc , θ tcf , µ) ∝ pcf (x(1:N) |X, θ tcf )pc (X|θ tc , µ) with Sq per data point x(i)
function M-Step
∂Uc (X,θ c )
∂L =−β ∂Uc (X,θ c )
+β
∂θ c
∂θ c
∂θ c
q(X|x(1:N) ,θ t ,θ ct )
pc (X|θ ct )
cf
*
+
∂pcf (x(1:N) |X,θ cf )
1
∂L =
∂θ cf
∂θ cf
pcf (x(1:N) |X,θ cf )
q(X|x,θ t ,θ ct )
cf
t +α
∂L
θ t+1
←
θ
θ c ∂θ c
c
c
t+1
t
∂L
θ
← θ cf + αθ
cf
cf ∂θ cf
end while
t
t
xpred ∼ p̃(xnew |θ c , θ cf ) with Spred ∗ Sc samples
function EvalMag(xpred , x(1:N) )
function EvalCorr(xpred , x(1:N) )
15: end for=0
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
35 / 56
Ising Model - Posterior and Prediction
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
36 / 56
Overview
1
Behavior of parameters with respect to µ
2
Influence of available data Sf for training the model
3
Level of coarse-graining
4
Aspects of coarse-grained models
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
37 / 56
Behavior of parameters with respect to µ
Constant attributes:
I
Sc = 200 (samples X ∼ pc (X|θ c ))
I
Sq = 50 (samples X ∼ q(X(i) |x(i) , θ c , θ cf ))
I
J0 = 1 (overall interaction strength)
I
Spred = 100Sc (samples for prediction step after learning the model)
Furthermore:
I
nf = 32, system size fine-scale
I
Lf = Lc = 1, in fine- and coarse scale nearest-neighbor interactions are
regarded
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
38 / 56
Behavior of parametrization with respect to µ
The potentials follows to,
nf
X
1 X
xi
U(J, µ) = − J
xi xj − µ
2
|i−j|=1
i=1
nc
X
X
1
Uc (θ c , µ) = − θc
Xi Xj − µ
Xi .
2
|i−j|=1
i=1
I
Learn two parameters θ c and θ cf of the mapping pcf
I
Learning for every evaluated µ: µ = [−4.5, 4.5] with a step size of 0.6
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
39 / 56
1
θc,predCg
θc,relEntr
2
θcf
θc
0.9
1
0.8
0
−4
−2
0
2
0.7
4
µ
Figure : θc : Parameter of potential Uc
WCPM Seminar
−4
−2
0
2
4
µ
Figure : θcf : Parameter of the mapping pcf (x|X, θcf )
Predictive Coarse-Graining
November 27, 2015
40 / 56
1
Truth
relEntr
predCg
hm(µ)i
0.5
0
−0.5
−1
−4
−2
0
2
4
µ
Figure : Magnetization
Key aspects
I
True magnetization based on data set and predicted by ’predCg’ coincides.
I
Method ’relEntr’ not able to predict data set due to missing mapping
capabilities from coarse to fine.
I
Parameters θ c and θ cf behave symmetrical with respect to µ = 0.
I
Stronger interaction θ c for bigger |µ| but also higher probability θ cf that the
coarse configuration reflects the fine configuration.
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
41 / 56
Behavior with respect to size of data-set
I
Learn θ c and θ cf at each evaluated µ
I
Given different size of data-set Sf with Sf ∈ {5, 10, 20, 50}.
I
Interaction length in fine-scale Lf = 10 with exponential decay of overall
strength J0 = 1.5
Potentials
Uf (J, µ) = −
Lf
nf
X
X
1X
Jk
xi xj − µ
xi
2
k
X
1
1
Xi Xj − θclin
Uc (θ c , µ) = − θc
2
2
|i−j|=1
WCPM Seminar
i=1
|i−j|=k
Predictive Coarse-Graining
nc
X
i
Xi − µ
nc
X
Xi .
i=1
November 27, 2015
42 / 56
1
1
0.5
0.5
hm(µ)i
hm(µ)i
Behavior with respect to size of data-set
0
−0.5
−1
0
−0.5
−2
0
−1
2
−2
µ
0
2
µ
Figure : Uncertainty in magnetization Sf = 5
99% - 1% conf. interval
95% - 5% conf. interval
Pred. mean
Truth
Data
Figure : Uncertainty in magnetization Sf = 10
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
43 / 56
1
1
0.5
0.5
hm(µ)i
hm(µ)i
Behavior with respect to size of data-set
0
−0.5
−1
0
−0.5
−2
0
−1
2
µ
0
2
µ
Figure : Uncertainty in magnetization Sf = 20
WCPM Seminar
−2
Figure : Uncertainty in magnetization Sf = 50
Predictive Coarse-Graining
November 27, 2015
44 / 56
Behavior with respect to size of data-set
0.8
0.1
errcorr
errmag
0.6
5 · 10−2
0.4
0.2
0
0
20
40
60
80
0
100
Sf
20
40
60
80
100
Sf
Figure : Relative mean error of magnetization
with respect to increasing amount of data Sf ,
Sf ∈ {5, 10, 20, 50, 100}
WCPM Seminar
0
Figure : Relative mean error of correlation with
respect to increasing amount of data Sf ,
Sf ∈ {5, 10, 20, 50, 100}
Predictive Coarse-Graining
November 27, 2015
45 / 56
Behavior with respect to size of data-set
1
I
Uncertainty decreasing with
increasing amount of data available
I
By using Sf = 5 predictions
resembling true magnetization well
Sf = 1
Sf = 5
0.8
Sf = 10
Sf = 20
0.6
Rl
Sf = 50
Sf = 100
0.4
Truth
0.2
0
0
5
10
15
l sites apart
Figure : Correlation Rl for µ = 0 with
Sf ∈ {5, 10, 20, 50, 100}
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
46 / 56
Predictability by various levels of coarse-gaining
I
Choose various level of coarse-graining lc
I
Mapping pcf (x|X, θ cf ) responsible for lc ∈ {2, 4, 8, 16} fine variables
I
Coarse-grained potential Uc (X, θ c ) as given before
lc = 2
lc = 4
lc = 8
lc = 16
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
47 / 56
1
1
0.5
0.5
hm(µ)i
hm(µ)i
Predictability by various levels of coarse gaining
0
−0.5
−1
0
−0.5
−2
0
−1
2
−2
µ
0
2
µ
Figure : Uncertainty in magnetization lc = 16
99% confidence interval
95% confidence interval
Posterior mean
Truth
Data
Figure : Uncertainty in magnetization lc = 8
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
48 / 56
1
1
0.5
0.5
hm(µ)i
hm(µ)i
Predictability by various levels of coarse gaining
0
−0.5
−1
0
−0.5
−2
0
−1
2
µ
0
2
µ
Figure : Uncertainty in magnetization lc = 4
WCPM Seminar
−2
Figure : Uncertainty in magnetization lc = 2
Predictive Coarse-Graining
November 27, 2015
49 / 56
Predictability by various levels of coarse gaining
1
I
Applying lc = 8 leads already to
good predictions in magnetization
I
Predicting correlations needs finer
resolution in coarse-scale
lc = 2
lc = 4
0.8
lc = 8
Rl
lc = 16
Truth
0.6
0.4
0.2
0
5
10
15
l sites apart
Figure : Correlation Rl at µ = 0 with
lc ∈ {2, 4, 8, 16}
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
50 / 56
Model Comparison
I
I
Assumption before: Model is given for potential Uc (X, θ c )
Using different order of interactions and interaction lengths
Linear
2nd order
Interaction length 2
3rd order
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
51 / 56
Model Comparison
errcorr
·10−2
4
0.1
5 · 10−2
0
0
100
010
001
110
011
101
111
121
131
112
2
100
010
001
110
011
101
111
121
131
112
errmag
6
Model
Model
Model [abc]
I
(a) a = 0 no linear term, a = 1 linear term
I
(b) b = 0 no 2nd order term, b = x 2nd order term up to interaction length
Lc = x
I
(c) c = 0 no third order term, c = 1 third order term on (nearest correlation)
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
52 / 56
Model Comparison
I
Linear term necessary to describe data
I
Second order interaction necessary
I
In correlation: errors decreasing for including 2nd order interactions with
higher interaction lengths Lc
I
No further improvement for longer ranged third order interactions
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
53 / 56
Outline
Numerical issues / efficiency
I
Use advanced sampling methods / optimization methods, using curvature
information (in progress)
I
Variational approximations
Coarse-Graining
I
Hierarchical coarse-graining
I
How to select mappings from coarse to fine?
I
Probit- or Logit-classification model, O model for coarse variables
Algorithmic
I
Approach with sparsity priors for model selection (in progress).
WCPM Seminar
Predictive Coarse-Graining
November 27, 2015
54 / 56
Bibliography I
S. Are, M. A. Katsoulakis, P. Plecháč, and L. R. Bellet.
Multibody Interactions in Coarse-Graining Schemes for Extended Systems.
SIAM Journal on Scientific Computing, 31(2):987–1015, 2008.
I. Bilionis and N. Zabaras.
A stochastic optimization approach to coarse-graining using a relative-entropy framework.
The Journal of chemical physics, 138(4):044313, Jan. 2013.
C. M. Bishop.
Pattern Recognition and Machine Learning.
Springer, 2006.
A. Chaimovich and M. S. Shell.
Coarse-graining errors and numerical optimization using a relative entropy framework.
The Journal of chemical physics, 134(9):094112, Mar. 2011.
F. Ercolessi and J. B. Adams.
Interatomic potentials from first-principles calculations: The force-matching method.
EPL (Europhysics Letters), 26(8):583, 1994.
S. Izvekov and G. A. Voth.
Multiscale coarse graining of liquid-state systems.
J. Chem. Phys., 123(13):134105, 2005.
M. Katsoulakis, P. Plecháč, and L. Rey-Bellet.
Numerical and statistical methods for the coarse-graining of many-particle stochastic systems.
Journal of Scientific Computing, 37(1):43–71, 2008.
A. P. Lyubartsev and A. Laaksonen.
Calculation of effective interaction potentials from radial distribution functions: A reverse Monte Carlo approach.
Phys. Rev. E, 52:3730–3737, Oct 1995.
Bibliography II
D. Ming and M. E. Wall.
Allostery in a coarse-grained model of protein dynamics.
Phys. Rev. Lett., 95:198103, Nov 2005.
M. S. Shell.
The relative entropy is fundamental to multiscale and inverse thermodynamic problems.
J. Chem. Phys., 129(14):144108, 2008.
Download