Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
A new class of interacting Markov Chain Monte Carlo methods
P. Del Moral, A. Doucet INRIA Bordeaux & UBC Vancouver
Workshop on Numerics and Stochastics, Helsinki, August 2008
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
Outline
1 Introduction
Stochastic sampling problems Some stochastic engineering models Interacting sampling methods
2 Some classical distribution flows Feynman-Kac models
”Static” Boltzmann-Gibbs models
3 Interacting stochastic sampling technology Mean field particle methods
Interacting Markov chain Monte Carlo models (i-MCMC)
4 Functional fluctuation & comparisons
5 Some references
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Stochastic sampling problems
Stochastic sampling problems
”Nonlinear” distribution flow with ↑ level of complexity.
η n (dx n ) = γ n (dx n )
γ n (1) Time index n ∈ N State var. x n ∈ E n
Two objectives :
1
∼ ”Sampling independent ” random variables w.r.t. η n
2Computation of the normalizing constants γ n (1)
(= Z n Partition functions).
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Some stochastic engineering models
Stochastic engineering Conditional & Boltzmann-Gibbs’ measures Filtering: Signal-Observation (X n , Y n ) [Radar, Sonar, GPS, ...]
η n = Law(X n | (Y 0 , . . . , Y n ))
Rare events: [Overflows, ruin processes, epidemic propagations,...]
η n = Law(X n | n intermediate events) & Z n = P (Rare event) Molecular simulation: [ground state energies, directed polymers...]
η n := Feynman-Kac/Boltzmann-Gibbs
∼ Free Markov motion in an absorbing medium Combinatorial counting, Global optimization, HMM
η n = 1 Z n
e
−βnV (x) λ(dx) or η n = 1 Z n
1 A
n(x) λ(dx)
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Interacting sampling methods
Two simple ingredients
Find or Understand the probability mass transformation η n = Φ n (η n−1 )
∼ Cooling schemes, temp. variations, constraints sequences, subset restrictions, observation data, conditional events,...
Natural interacting sampling idea :
Use η n−1 or its empirical approx. to sample w.r.t. η n
1Monte-Carlo/ Mean Field models :
η n = Law(X n ) with Markov : X n−1
∼ηn−1
−−−−− −−→ X n
2Interacting MCMC models :
Use the occupation measures of an MCMC with target η n−1
MCMC target η n
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Feynman-Kac models
Feynman-Kac distribution flows
Weak representation: [f n test funct. on a state space E n ]
η n (f n ) = γ n (f n )
γ n (1) with γ n (f n ) = E
f n (X n ) Y
0≤p<n
G p (X p )
A Key Formula: Z n = E Q
0≤p<n G p (X p )
= Q
0≤p<n η p (G p ) Path space models X n = (X 0
0, . . . , X n
0)
Examples
G n ∈ [0, 1] particle absorption models.
G n = Observation likelihood function Filtering models.
G n = 1 A
nConditional/Restriction models.
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Feynman-Kac models
Nonlinear distribution flows Evolution equation:
η n+1 = Φ n+1 (η n ) = Ψ G
n(η n )M n+1
With the only 2 transformations :
X -Free Markov transport eq. : [M n (x n−1 , dx n ) from E n−1 into E n ] (η n−1 M n )(dx n ) :=
Z
E
n−1η n−1 (dx n−1 ) M n (x n−1 , dx n ) Bayes-Boltzmann-Gibbs transformation :
Ψ G
n(η n )(dx n ) := 1
η n (G n ) G n (x n ) η n (dx n )
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
”Static” Boltzmann-Gibbs models
Boltzmann-Gibbs distribution flows
Target distribution flow : η n (dx) ∝ g n (x) λ(dx) Product hypothesis :
g n = g n−1 × G n−1 = ⇒ η n = Ψ G
n−1(η n−1 ) Running Ex.:
g n = 1 A
nwith A n ↓ ⇒ G n−1 = 1 A
ng n = e
−βnV with β n ↑ ⇒ G n−1 = e
−(βn−βn−1)V
Problem : η n = Ψ G
n−1(η n−1 ) = unstable equation.
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
”Static” Boltzmann-Gibbs models
Feynman-Kac modeling
Choose M
n(x, dy) s.t. local fixed point eq. → η n = η n M n (Metropolis, Gibbs,...)
Stable equation :
g n = g n−1 × G n−1 = ⇒ η n = Ψ G
n−1(η n−1 )
= ⇒ η
n= η
nM n = Ψ
Gn−1(η
n−1)M n = FK-model Feynman-Kac ”dynamical” formulation (X n Markov M n )
Z
f (x) g n (x) λ(dx) ∝ E
f (X n ) Y
0≤p<n
G p (X p )
Interacting Metropolis/Gibbs/... stochastic algorithms.
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Mean field particle methods
Mean field interpretation
Nonlinear Markov models : Always ∃K n,η (x , dy ) Markov s.t.
η n = Φ n (η n−1 ) = η n−1 K n,η
n−1=Law X n
i.e. :
P (X n ∈ dx n | X n−1 ) = K n,η
n−1(X n−1 , dx n ) Mean field particle interpretation
Markov chain ξ
n= (ξ
1n, . . . , ξ
nN) ∈ E
nNs.t.
η n N := 1 N
X
1≤i≤N
δ
ξin
' N↑∞ η n
Particle approximation transitions (∀1 ≤ i ≤ N) ξ n−1 i ξ n i ∼ K n,η
Nn−1
(ξ n−1 i , dx n )
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
Mean field particle methods
Discrete generation mean field particle model
Schematic picture : ξ
n∈ E
nNξ
n+1∈ E
n+1Nξ n 1
K
n+1,ηN−−−−−−−−−−→
n.. .
ξ n i −−−−−−−−−−→
.. .
ξ n N −−−−−−−−−−→
ξ 1 n+1 .. . ξ i n+1
.. . ξ N n+1
Rationale :
η N n ' N↑∞ η n = ⇒ K n+1,η
Nn
' N↑∞ K n+1,η
n= ⇒ ξ i n almost iid copies of X n
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Mean field particle methods
Ex.: Feynman-Kac distribution flows
FK-Nonlinear Markov models :
n = n (η n ) ≥ 0 s.t. η n -a.e. n G n ∈ [0, 1] ( n = 0 not excluded) K n+1,η
n(x, dz) =
Z
S n,η
n(x, dy ) M n+1 (y , dz ) S n,η
n(x, dy) := n G n (x) δ x (dy) + (1 − n G n (x)) Ψ G
n(η n )(dy) Mean field genetic type particle model :
ξ n i ∈ E n
accept/reject/selection
−−−−−−−−− −−→ ξ b n i ∈ E n
proposal/mutation
−−−−−−−−− −−→ ξ n+1 i ∈ E n+1
Examples :
G n = 1 A killing with uniform replacement.
M n -Metropolis/Gibbs moves G n -interaction function
(subsets fitting or change of temperatures)
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
Mean field particle methods
Mean field genetic type particle model :
ξ 1 n .. . ξ i n
.. . ξ n N
S
n,ηN−−−−−−−−−−→
n
ξ b n 1
M
n+1−−−−−−−−−−→
.. .
ξ b n i −−−−−−−−−−→
.. .
ξ b n N −−−−−−−−−−→
ξ n+1 1 .. . ξ n+1 i
.. . ξ n+1 N
Accept/Reject/Selection transition : S n,η
Nn
(ξ n i , dx) := n G n (ξ i n ) δ
ξin
(dx) + 1 − n G n (ξ n i ) P N j=1
G
n(ξ
jn)
PNk=1
G
n(ξ
nk) δ
ξj n(dx)
Ex. : G n = 1 A , n = 1 G n (ξ n i ) = 1 A (ξ n i )
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Mean field particle methods
Path space models
X
n= (X
00, . . . , X
n0) genealogical tree/ancestral lines η n N := 1
N X
1≤i≤N
δ
ξin
= 1
N X
1≤i≤N
δ (ξ
i0,n,ξi1,n,...,ξin,n
) ' N↑∞ η n
Unbias particle approximations : γ n N (1) = Y
0≤p<n
η N p (G p ) ' N↑∞ γ n (1) = Y
0≤p<n
η p (G p )
Ex. G n = 1 A :
⇒ γ N n (1) = Y
0≤p<n
(success % at p)
FK-Mean field particle models = sequential Monte Carlo,
population Monte Carlo, particle filters, pruning, spawning,
reconfiguration, quantum Monte carlo, go with the winner...
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Interacting Markov chain Monte Carlo models (i-MCMC)
Objective
Find a series of MCMC models X (n) := (X k (n) ) k≥0 s.t.
η (n) k = 1 k + 1
X
0≤l≤k
δ X
(n) l' k↑∞ η n
⇒ Use η
(n)k' η
nto define X
(n+1)with target η
n+1Advantages
Using η n the sampling η n+1 is often easier.
Improve the proposition step in any Metropolis type model with target η n+1 ( enters the stability prop. of the flow η
n) Increases the precision at every time step.
But CLT variance often ≥ CLT variance mean field models.
Easy to combine with mean field stochastic algorithms.
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Interacting Markov chain Monte Carlo models (i-MCMC)
Interacting Markov chain Monte Carlo models Find M 0 and a collection of transitions M n,µ s.t.
η 0 = η 0 M 0 and Φ n (µ) = Φ n (µ)M n,µ
(X k (0) ) k≥0 Markov chain ∼ M 0 .
Given X (n) , we let X k (n+1) with Markov transtions M n+1,η
(n) kRationale : η (n) k ' η n = ⇒
( Φ n+1 (η k (n) ) ' Φ n+1 (η n ) = η n+1
M n+1,η
(n) k' M n+1,η
nwith fixed point η n+1
= ⇒ η k (n+1) ' η n+1
Example : M n,µ (x, dy) = Φ n (µ)(dy) X k (n+1) r.v. ∼ Φ n+1
η k (n)
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references Interacting Markov chain Monte Carlo models (i-MCMC)
((n − 1)-th chain) X 0 (n−1)
↓ X 1 (n−1)
↓ .. .
↓ X k (n−1)
ηk(n−1)'ηn−1
−−−−−−−−−−− −−→
↓ .. .
(n-th chain) X 0 (n)
↓ .. . .. .
↓ X k (n)
↓ M n,η
(n−1)k
' M n,η
n−1↓
X k+1 (n)
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
[MEAN FIELD PARTICLE MODEL]Nonlinear semigroup−→Φp,n(ηp) :=ηn
Local fluctuation theorem : WnN:=
√ Nh
ηnN−Φn“ ηNn−1”i
'Wn ⊥Centered Gaussian field Local transport formulation :
η0 → η1= Φ1(η0) → η2= Φ0,2(η0) → · · · → Φ0,n(η0)
⇓
ηN0 → Φ1(ηN0) → Φ0,2(ηN0) → · · · → Φ0,n(ηN0)
⇓
ηN1 → Φ2(ηN1) → · · · → Φ1,n(ηN1)
⇓
η2N → · · · → Φ2,n(ηN2)
⇓
.. . ηn−1N → Φn(ηNn−1)
⇓ ηnN Key decomposition formula :
ηnN−ηn = n X
q=0
[Φq,n(ηqN)−Φq,n(Φq(ηNq−1))]
' 1
√ N
n X
q=0
WqNDq,n←-First order decomp.Φp,n(η)−Φp,n(µ)'(η−µ)Dp,n+ (η−µ)⊗2. . .
⇒ Example Functional CLT :
√ N
h ηNn−ηn
i '
n X
q=0 WqDq,n
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
[i-MCMC] Nonlinear sgΦp,n(ηp) =ηnwith a first order decomp. :
Φp,n(η)−Φp,n(µ)'(η−µ)Dp,n+ (η−µ)⊗2. . .
⇓ Functional CLT for correlated/interacting MCMC models :
√ k h
ηk(n)−ηni '
n X
q=0
p(2(n−q))!
(n−q)! VqDq,n
with` Vq´
q≥0⊥ Centered Gaussian field
E
“ Vq(f)2”
=ηqh
(f−ηq(f))2i + 2P
m≥1ηq
»
(f−ηq(f))Mq,ηm
q−1(f−ηq(f)) –
”Comparisons” :[Mean field case]` Wq´
q≥0⊥ Centered Gaussian field
E
“ Wq(f)2”
=ηq−1n
Kq,ηq−1(f−Kq,ηq−1(f))2o
Case :Kq,η(x,dy) =Mq,η(x,dy) = Φq(η)(dy) =⇒(Vq=Wq) =⇒[Mean field]>[i-MCMC]
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references
Some references
Interacting stochastic simulation algorithms
Mean field and Feynman-Kac particle models :
Feynman-Kac formulae. Genealogical and
interacting particle systems, Springer (2004) ⊕ Refs.
joint work with L. Miclo. A Moran particle system
approximation of Feynman-Kac formulae. Stochastic Processes and their Applications, Vol. 86, 193-216 (2000).
joint work with L. Miclo. Branching and Interacting Particle Systems Approximations of Feynman-Kac Formulae. S´ eminaire de Probabilit´ es XXXIV, Lecture Notes in Mathematics, Springer-Verlag Berlin, Vol. 1729, 1-145 (2000).
Sequential Monte Carlo models :
joint work with Doucet A., Jasra A. Sequential Monte Carlo Samplers. JRSS B (2006).
joint work with A. Doucet. On a class of genealogical and
interacting Metropolis models. S´ em. de Proba. 37 (2003).
Introduction Some classical distribution flows Interacting stochastic sampling technology Functional fluctuation & comparisons Some references