Rong Chen

advertisement
Discussion: Population SMC
SMC or MCMC – is it still a question?
Rong Chen
Department of Statistics
Rutgers University
1
SMC and MCMC
Misconceptions:
• SMC is used only for state-space models, HMM and other
types of dynamic systems
– A sequence of target distributions with increasing dimension
• MCMC is (the only solution) for fixed dimensional problems
– A single distribution
2
We claim:
• SMC is a valid alternative to MCMC
• SMC can be a (possibly more) powerful tool in solving fixdimensional problems
• MCMC is a special case of SMC
3
Change
The dynamic system framework for SMC (Liu and Chen 1998)
• A sequence of distributions π1(x1), . . . , πt(xt), . . .
• Increasing dimension xt = (xt−1, xt)
• Sequential importance sampling:
(j)
– generate xt
(j)
conditioned on xt−1
(j)
(j)
(j)
– append: xt = (xt−1, xt )
– update weight
What do we do with one single target distribution π(x)?
4
Approach (I): The Growth Principle
Decompose a complex problem into a sequence of simpler problems.
• Target distribution π(x) where x = (x1, . . . , xN )
• Let xt = (x1, . . . , xt) = (xt−1, xt)
• Define a sequence of intermediate distributions πt(xt).
• Moving from πt−1(xt−1) to πt(xt) is simple.
• Moving from πt−1(xt−1) to πt(xt) is smooth.
Z
πt(xt−1, xt)dxt ≈ πt−1(xt−1)
• πN (xN ) = π(x)
5
Approach (II): Fixed dimension with augmentation
• Target distribution π(x), for x ∈ Ω
• Let xt = (x1, . . . , xt) ∈ Ωt where xi ∈ Ω.
• Construct a sequence of intermediate (joint) distributions
πt(x1, . . . , xt)
on
Ωt,
often through a transition rule πt−1 → πt.
• The low-dimensional marginal distributions of the giant joint
distribution πn(x1, . . . , xn) is in some way related to the target
distribution π(x).
– e.g. πn(x1) = π(x) or πn(xn) = π(x)
6
Arnaud’s approach:
• starting with a relatively trivial π1(x1), where x1 ∈ Ω
• augment: πt(x1, . . . , xt), t = 1, . . . , n.
• eventually, at t = n, the marginal distribution of the last component πn(xn) is the target distribution π(x)
(j)
• inference using the samples of the last component xn
Chris’ approach:
• starting with true π1(x1) = π(x).
• augment: πt(x1, . . . , xt) so the marginal distribution of the last
component πt(xt) is the target distribution π(x)
• inference using samples of all components.
7
MCMC as a special case of SMC:
MCMC:
• Target distribution π(x) where x ∈ Ω.
• A Markov chain x1, . . . , xt, . . ., with a transition kernel k and
each xt ∈ Ω.
• x1 ∼ π1
• (x1, . . . , xt) ∼ πt(x1, . . . , xt) where
πt(xt) = πt−1(xt−1)k(xt | xt−1)
• (hopefully) after some N , the marginal distribution of the last
component xt, πt(xt) is close to π(x), for t > N .
8
MCMC as a special case of SMC:
• A dynamic system through augmentation
• The sequence of intermediate distributions is built based on a
special type of kernel to ensure that the marginal distributions
of the t-th component converge to target distribution
• Trial distribution is the same as kernel (often through rejection method, as in M-H)
• Does not use importance weights but uses burn-in
9
Open questions:
Caution: flexibility comes with more ways of making mistakes
• Design the sequence of intermediate distributions πt(xt)
• Construct efficient trial distributions
• Take advantage of MCMC-like moves
• Avoid the problems often encountered with MCMC algorithms.
10
Download