SlideShare une entreprise Scribd logo
1  sur  19
Télécharger pour lire hors ligne
Proximal algorithms in probability spaces
Dr. Marcelo Pereyra
http://www.macs.hw.ac.uk/∼mp71/
Maxwell Institute for Mathematical Sciences, Heriot-Watt University
March 2018, SAMSI workshop, USA.
M. Pereyra (MI — HWU) LMS 17 0 / 18
Outline
1 Background
2 Proximal algorithms in probability spaces
M. Pereyra (MI — HWU) LMS 17 1 / 18
Imaging inverse problems
We are interested in an unknown x ∈ Rd
.
We measure y, related to x by a statistical model p(y x).
The recovery of x from y is ill-posed or ill-conditioned, resulting in
significant uncertainty about x.
For example, in many imaging problems
y = Ax + w,
for some operator A that is rank-deficient, and additive noise w.
x y ˆx
M. Pereyra (MI — HWU) LMS 17 2 / 18
The Bayesian framework
We use priors to reduce uncertainty and deliver accurate results.
Given the prior p(x), the posterior distribution of x given y
p(x y) = p(y x)p(x)/p(y)
models our knowledge about x after observing y.
In this talk we consider that p(x y) is log-concave; i.e.,
p(x y) = exp{−φ(x)}/∫
Rd
exp{−φ(x)}dx ,
where φ(x) is a l.s.c. convex function (often φ ∉ C1
).
M. Pereyra (MI — HWU) LMS 17 3 / 18
Maximum-a-posteriori (MAP) estimation
The predominant Bayesian approach in imaging is MAP estimation
ˆxMAP = argmax
x∈Rd
p(x y),
= argmin
x∈Rd
φ(x),
(1)
that can be computed efficiently by proximal convex optimisation.
Typically φ = f + g, with ∇f Lipchitz and g ∉ C1
and prox-easy.
However, knowledge of ˆxMAP provides little information about p(x y)!!!
M. Pereyra (MI — HWU) LMS 17 4 / 18
Uncertainty quantification in radio-interferometry
3C2888 ˆxMMSE local credible intervals
M31 ˆxMMSE local credible intervals
Figure : Radio galaxies (size 256 × 256 pixels) - computing time approx. 5 min.
M. Pereyra (MI — HWU) LMS 17 5 / 18
Bayesian model selection
p(Mk y) = p(Mk )∫ p(x,y Mk )dx/p(y) with
p(x,y M1) ∝ exp[−( y − H1x 2
/2σ2
) − βTV (x)],
p(x,y M2) ∝ exp[−( y − H2x 2
/2σ2
) − βTV (x)].
Boat image deblurring experiment (comp. time 30 minutes p/model):
observation y
(5 × 5 uniform blur, BSNR 40dB)
ˆxM1 (PSNR 34dB)
p(M1 y) = 0.96
ˆxM2 (PSNR 33dB)
p(M2 y) = 0.04
M. Pereyra (MI — HWU) LMS 17 6 / 18
Sparse image deblurring
Estimation of reg. param. β by marginal maximum likelihood
ˆβ = argmax
β∈R+
p(y β), with p(y β) ∝ ∫ exp(− y − Ax 2
/2σ2
− β x 1)dx
y ˆxMAP Reg. param β
Figure : Maximum marginal likelihood estimation of regularisation parameter β.
Computing time 0.75 secs..
M. Pereyra (MI — HWU) LMS 17 7 / 18
Sparse image deblurring
Bayesian credible region C∗
α = {x p(x y) ≥ γα} with
P[x ∈ Cα y] = 1 − α, and p(x y) ∝ exp(− y − Ax 2
/2σ2
− β x 1)
y ˆxMAP uncertainty estimates
Figure : Live-cell microscopy data (Zhu et al., 2012). Uncertainty analysis
(±78nm × ±125nm) in close agreement with the experimental precision ±80nm.
Computing time 4 minutes. M = 105
iterations.
M. Pereyra (MI — HWU) LMS 17 8 / 18
Inference by Markov chain Monte Carlo integration
Monte Carlo integration
Given a set of samples X1,...,XM distributed according to p(x y), we
approximate posterior expectations and probabilities
1
M
M
∑
m=1
h(Xm) → E{h(x) y}, as M → ∞
Guarantees from CLTs, e.g., 1√
M
∑M
m=1 h(Xm) ∼ N[E{h(x) y},Σ].
Markov chain Monte Carlo:
Construct a Markov kernel Xm+1 Xm ∼ K(⋅ Xm) such that the Markov
chain X1,...,XM has p(x y) as stationary distribution.
MCMC simulation in high-dimensional spaces is very challenging.
M. Pereyra (MI — HWU) LMS 17 9 / 18
Unadjusted Langevin algorithm
Suppose for now that p(x y) ∈ C1
. Then, we can generate samples by
mimicking a Langevin diffusion process that converges to p(x y) as t → ∞,
X dXt =
1
2
∇log p (Xt y)dt + dWt, 0 ≤ t ≤ T, X(0) = x0.
where W is the n-dimensional Brownian motion.
Because solving Xt exactly is generally not possible, we use an Euler
Maruyama approximation and obtain the “unadjusted Langevin algorithm”
ULA Xm+1 = Xm + δ∇log p(Xm y) +
√
2δZm+1, Zm+1 ∼ N(0,In),
which is remarkably efficient when φ(x) = −log p(x y) is l.s.c. convex and
L-Lipchitz differentiable with δ < L−1
.
M. Pereyra (MI — HWU) LMS 17 10 / 18
Proximal ULA
If φ ∉ C1
, we use the (Moreau-Yoshida regularised) proximal ULA
MYULA Xm+1 = (1 − δ
λ )Xm − δ∇f {Xm} + δ
λ proxλ
g {Xm} +
√
2δZm+1,
for any suitable decomposition φ = f + g, λ > 0, and δ < λ/(λLf + 1).
Non-asymptotic estimation error bound:
For any > 0, there exist δ > 0 and M ∈ N such that ∀δ < δ and ∀M ≥ M
δx0 QM
δ − p TV < + λL2
g .
If f + g is strongly convex outside some ball, then M scales as
O(d log(d)). See Durmus et al. (2017) for other convergence results.
M. Pereyra (MI — HWU) LMS 17 11 / 18
Outline
1 Background
2 Proximal algorithms in probability spaces
M. Pereyra (MI — HWU) LMS 17 12 / 18
Wasserstein space
P2(Rd
): set of probability measures with finite second moment.
Wasserstein distance of order 2
W2(µ,ν) = inf
π∈Π(µ,ν)
∫
Rd ×Rd
x − y 2
π(x,y)dxdy
with Π(µ,ν) is the set of measures that have µ and ν as marginals.
Kullbar-Liebler divergence
KL(µ,ν) = ∫
Rd
log [
ν(x)
µ(x)
]ν(x)dx
Note: KL(µ,ν) = 0 iff µ = ν.
M. Pereyra (MI — HWU) LMS 17 13 / 18
A functional F P2(Rd
) → (−∞,∞] is λ-displacement convex iff
F(µt) ≤ F(µ0) + F(µ1) − λt(1 − t)W2(µ0,µ1)
where µt is the point t ∈ [0,1] along the constant speed geodesic
between µ0 and µ1, i.e., W2(µt,µs) = (t − s)W2(µ0,µ1).
If π ∈ P2 is λ-strongly log-concave (i.e., −log π strongly convex), then
KL(µ,π) is λ-displacement convex w.r.t. µ.
M. Pereyra (MI — HWU) LMS 17 14 / 18
Proximal optimisation in the Wasserstein space
Key Result (R. Jordan, D. Kinderlehrer, and F. Otto. 1998)
The sequence of distributions generated by the iterative algorithm
µ(t+1)
= argmin
ν∈P2(Rd )
KL(ν,π) + W2(µ(t)
,ν)
(proximal-point method in P2 with metric W2) is equivalent to that of
dXt =
1
2
∇log p (Xt y)dt + dWt, 0 ≤ t ≤ T, X(0) ∼ µµ0,
evaluated at times t = 1,... (their results use a different notation and
involve the Fokker-Plank representation of the SDE).
MYULA is essentially a sampling approximation of this algorithm for
minimising KL(µ,πλ) where πλ is the Moreau approximation of π.
M. Pereyra (MI — HWU) LMS 17 15 / 18
Open problems:
What do other (e.g., accelerated, primal-dual, etc.) optimisation
algorithms look like in P2 with metric W2?
What are the MCMC methods that we can derive from them?
Do they inherit their good convergence properties?
M. Pereyra (MI — HWU) LMS 17 16 / 18
Thank you!
Bibliography:
Durmus, A., Moulines, E., and Pereyra, M. (2017). Efficient Bayesian computation by
proximal Markov chain Monte Carlo: when Langevin meets Moreau. SIAM J. Imaging
Sci. to appear.
Zhu, L., Zhang, W., Elnatan, D., and Huang, B. (2012). Faster STORM using
compressed sensing. Nat. Meth., 9(7):721–723.
M. Pereyra (MI — HWU) LMS 17 17 / 18
Comparison with PxMALA
M. Pereyra (MI — HWU) LMS 17 18 / 18

Contenu connexe

Tendances

Introduction to MCMC methods
Introduction to MCMC methodsIntroduction to MCMC methods
Introduction to MCMC methodsChristian Robert
 
Poster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conferencePoster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conferenceChristian Robert
 
Non-informative reparametrisation for location-scale mixtures
Non-informative reparametrisation for location-scale mixturesNon-informative reparametrisation for location-scale mixtures
Non-informative reparametrisation for location-scale mixturesChristian Robert
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big DataChristian Robert
 
On Clustering Financial Time Series - Beyond Correlation
On Clustering Financial Time Series - Beyond CorrelationOn Clustering Financial Time Series - Beyond Correlation
On Clustering Financial Time Series - Beyond CorrelationGautier Marti
 
MCMC and likelihood-free methods
MCMC and likelihood-free methodsMCMC and likelihood-free methods
MCMC and likelihood-free methodsChristian Robert
 

Tendances (20)

Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Richard Everitt's slides
Richard Everitt's slidesRichard Everitt's slides
Richard Everitt's slides
 
Chris Sherlock's slides
Chris Sherlock's slidesChris Sherlock's slides
Chris Sherlock's slides
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Introduction to MCMC methods
Introduction to MCMC methodsIntroduction to MCMC methods
Introduction to MCMC methods
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Poster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conferencePoster for Bayesian Statistics in the Big Data Era conference
Poster for Bayesian Statistics in the Big Data Era conference
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Non-informative reparametrisation for location-scale mixtures
Non-informative reparametrisation for location-scale mixturesNon-informative reparametrisation for location-scale mixtures
Non-informative reparametrisation for location-scale mixtures
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...
QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...
QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big Data
 
On Clustering Financial Time Series - Beyond Correlation
On Clustering Financial Time Series - Beyond CorrelationOn Clustering Financial Time Series - Beyond Correlation
On Clustering Financial Time Series - Beyond Correlation
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
MCMC and likelihood-free methods
MCMC and likelihood-free methodsMCMC and likelihood-free methods
MCMC and likelihood-free methods
 
Intro to ABC
Intro to ABCIntro to ABC
Intro to ABC
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 

Similaire à QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces - Marcelo Pereyra, Mar 22, 2018

Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
 
A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...
A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...
A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...A G
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Alexander Litvinenko
 
Tensor train to solve stochastic PDEs
Tensor train to solve stochastic PDEsTensor train to solve stochastic PDEs
Tensor train to solve stochastic PDEsAlexander Litvinenko
 
Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesFrank Nielsen
 
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming ProblemHigher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Probleminventionjournals
 
On learning statistical mixtures maximizing the complete likelihood
On learning statistical mixtures maximizing the complete likelihoodOn learning statistical mixtures maximizing the complete likelihood
On learning statistical mixtures maximizing the complete likelihoodFrank Nielsen
 
Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)Frank Nielsen
 
Low rank tensor approximation of probability density and characteristic funct...
Low rank tensor approximation of probability density and characteristic funct...Low rank tensor approximation of probability density and characteristic funct...
Low rank tensor approximation of probability density and characteristic funct...Alexander Litvinenko
 
Litvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfLitvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfAlexander Litvinenko
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationAlexander Litvinenko
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”IOSRJM
 

Similaire à QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces - Marcelo Pereyra, Mar 22, 2018 (20)

Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...
A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...
A Fast Algorithm for Solving Scalar Wave Scattering Problem by Billions of Pa...
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
 
Tensor train to solve stochastic PDEs
Tensor train to solve stochastic PDEsTensor train to solve stochastic PDEs
Tensor train to solve stochastic PDEs
 
Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective Divergences
 
MUMS Opening Workshop - Panel Discussion: Facts About Some Statisitcal Models...
MUMS Opening Workshop - Panel Discussion: Facts About Some Statisitcal Models...MUMS Opening Workshop - Panel Discussion: Facts About Some Statisitcal Models...
MUMS Opening Workshop - Panel Discussion: Facts About Some Statisitcal Models...
 
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming ProblemHigher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
Higher-Order (F, α, β, ρ, d) –Convexity for Multiobjective Programming Problem
 
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
 
On learning statistical mixtures maximizing the complete likelihood
On learning statistical mixtures maximizing the complete likelihoodOn learning statistical mixtures maximizing the complete likelihood
On learning statistical mixtures maximizing the complete likelihood
 
Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)
 
talk MCMC & SMC 2004
talk MCMC & SMC 2004talk MCMC & SMC 2004
talk MCMC & SMC 2004
 
Metodo gauss_newton.pdf
Metodo gauss_newton.pdfMetodo gauss_newton.pdf
Metodo gauss_newton.pdf
 
Low rank tensor approximation of probability density and characteristic funct...
Low rank tensor approximation of probability density and characteristic funct...Low rank tensor approximation of probability density and characteristic funct...
Low rank tensor approximation of probability density and characteristic funct...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Litvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdfLitvinenko_RWTH_UQ_Seminar_talk.pdf
Litvinenko_RWTH_UQ_Seminar_talk.pdf
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantification
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
 
hone_durham
hone_durhamhone_durham
hone_durham
 
A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”A Note on “   Geraghty contraction type mappings”
A Note on “   Geraghty contraction type mappings”
 

Plus de The Statistical and Applied Mathematical Sciences Institute

Plus de The Statistical and Applied Mathematical Sciences Institute (20)

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
 

Dernier

Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfagholdier
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfNirmal Dwivedi
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxDenish Jangid
 
Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structuredhanjurrannsibayan2
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - Englishneillewis46
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfDr Vijay Vishwakarma
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024Elizabeth Walsh
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...ZurliaSoop
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptxMaritesTamaniVerdade
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSCeline George
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxVishalSingh1417
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentationcamerronhm
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Pooja Bhuva
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxDr. Sarita Anand
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseAnaAcapella
 

Dernier (20)

Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structure
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - English
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
 
Spatium Project Simulation student brief
Spatium Project Simulation student briefSpatium Project Simulation student brief
Spatium Project Simulation student brief
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptx
 
Spellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please PractiseSpellings Wk 3 English CAPS CARES Please Practise
Spellings Wk 3 English CAPS CARES Please Practise
 

QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces - Marcelo Pereyra, Mar 22, 2018

  • 1. Proximal algorithms in probability spaces Dr. Marcelo Pereyra http://www.macs.hw.ac.uk/∼mp71/ Maxwell Institute for Mathematical Sciences, Heriot-Watt University March 2018, SAMSI workshop, USA. M. Pereyra (MI — HWU) LMS 17 0 / 18
  • 2. Outline 1 Background 2 Proximal algorithms in probability spaces M. Pereyra (MI — HWU) LMS 17 1 / 18
  • 3. Imaging inverse problems We are interested in an unknown x ∈ Rd . We measure y, related to x by a statistical model p(y x). The recovery of x from y is ill-posed or ill-conditioned, resulting in significant uncertainty about x. For example, in many imaging problems y = Ax + w, for some operator A that is rank-deficient, and additive noise w. x y ˆx M. Pereyra (MI — HWU) LMS 17 2 / 18
  • 4. The Bayesian framework We use priors to reduce uncertainty and deliver accurate results. Given the prior p(x), the posterior distribution of x given y p(x y) = p(y x)p(x)/p(y) models our knowledge about x after observing y. In this talk we consider that p(x y) is log-concave; i.e., p(x y) = exp{−φ(x)}/∫ Rd exp{−φ(x)}dx , where φ(x) is a l.s.c. convex function (often φ ∉ C1 ). M. Pereyra (MI — HWU) LMS 17 3 / 18
  • 5. Maximum-a-posteriori (MAP) estimation The predominant Bayesian approach in imaging is MAP estimation ˆxMAP = argmax x∈Rd p(x y), = argmin x∈Rd φ(x), (1) that can be computed efficiently by proximal convex optimisation. Typically φ = f + g, with ∇f Lipchitz and g ∉ C1 and prox-easy. However, knowledge of ˆxMAP provides little information about p(x y)!!! M. Pereyra (MI — HWU) LMS 17 4 / 18
  • 6. Uncertainty quantification in radio-interferometry 3C2888 ˆxMMSE local credible intervals M31 ˆxMMSE local credible intervals Figure : Radio galaxies (size 256 × 256 pixels) - computing time approx. 5 min. M. Pereyra (MI — HWU) LMS 17 5 / 18
  • 7. Bayesian model selection p(Mk y) = p(Mk )∫ p(x,y Mk )dx/p(y) with p(x,y M1) ∝ exp[−( y − H1x 2 /2σ2 ) − βTV (x)], p(x,y M2) ∝ exp[−( y − H2x 2 /2σ2 ) − βTV (x)]. Boat image deblurring experiment (comp. time 30 minutes p/model): observation y (5 × 5 uniform blur, BSNR 40dB) ˆxM1 (PSNR 34dB) p(M1 y) = 0.96 ˆxM2 (PSNR 33dB) p(M2 y) = 0.04 M. Pereyra (MI — HWU) LMS 17 6 / 18
  • 8. Sparse image deblurring Estimation of reg. param. β by marginal maximum likelihood ˆβ = argmax β∈R+ p(y β), with p(y β) ∝ ∫ exp(− y − Ax 2 /2σ2 − β x 1)dx y ˆxMAP Reg. param β Figure : Maximum marginal likelihood estimation of regularisation parameter β. Computing time 0.75 secs.. M. Pereyra (MI — HWU) LMS 17 7 / 18
  • 9. Sparse image deblurring Bayesian credible region C∗ α = {x p(x y) ≥ γα} with P[x ∈ Cα y] = 1 − α, and p(x y) ∝ exp(− y − Ax 2 /2σ2 − β x 1) y ˆxMAP uncertainty estimates Figure : Live-cell microscopy data (Zhu et al., 2012). Uncertainty analysis (±78nm × ±125nm) in close agreement with the experimental precision ±80nm. Computing time 4 minutes. M = 105 iterations. M. Pereyra (MI — HWU) LMS 17 8 / 18
  • 10. Inference by Markov chain Monte Carlo integration Monte Carlo integration Given a set of samples X1,...,XM distributed according to p(x y), we approximate posterior expectations and probabilities 1 M M ∑ m=1 h(Xm) → E{h(x) y}, as M → ∞ Guarantees from CLTs, e.g., 1√ M ∑M m=1 h(Xm) ∼ N[E{h(x) y},Σ]. Markov chain Monte Carlo: Construct a Markov kernel Xm+1 Xm ∼ K(⋅ Xm) such that the Markov chain X1,...,XM has p(x y) as stationary distribution. MCMC simulation in high-dimensional spaces is very challenging. M. Pereyra (MI — HWU) LMS 17 9 / 18
  • 11. Unadjusted Langevin algorithm Suppose for now that p(x y) ∈ C1 . Then, we can generate samples by mimicking a Langevin diffusion process that converges to p(x y) as t → ∞, X dXt = 1 2 ∇log p (Xt y)dt + dWt, 0 ≤ t ≤ T, X(0) = x0. where W is the n-dimensional Brownian motion. Because solving Xt exactly is generally not possible, we use an Euler Maruyama approximation and obtain the “unadjusted Langevin algorithm” ULA Xm+1 = Xm + δ∇log p(Xm y) + √ 2δZm+1, Zm+1 ∼ N(0,In), which is remarkably efficient when φ(x) = −log p(x y) is l.s.c. convex and L-Lipchitz differentiable with δ < L−1 . M. Pereyra (MI — HWU) LMS 17 10 / 18
  • 12. Proximal ULA If φ ∉ C1 , we use the (Moreau-Yoshida regularised) proximal ULA MYULA Xm+1 = (1 − δ λ )Xm − δ∇f {Xm} + δ λ proxλ g {Xm} + √ 2δZm+1, for any suitable decomposition φ = f + g, λ > 0, and δ < λ/(λLf + 1). Non-asymptotic estimation error bound: For any > 0, there exist δ > 0 and M ∈ N such that ∀δ < δ and ∀M ≥ M δx0 QM δ − p TV < + λL2 g . If f + g is strongly convex outside some ball, then M scales as O(d log(d)). See Durmus et al. (2017) for other convergence results. M. Pereyra (MI — HWU) LMS 17 11 / 18
  • 13. Outline 1 Background 2 Proximal algorithms in probability spaces M. Pereyra (MI — HWU) LMS 17 12 / 18
  • 14. Wasserstein space P2(Rd ): set of probability measures with finite second moment. Wasserstein distance of order 2 W2(µ,ν) = inf π∈Π(µ,ν) ∫ Rd ×Rd x − y 2 π(x,y)dxdy with Π(µ,ν) is the set of measures that have µ and ν as marginals. Kullbar-Liebler divergence KL(µ,ν) = ∫ Rd log [ ν(x) µ(x) ]ν(x)dx Note: KL(µ,ν) = 0 iff µ = ν. M. Pereyra (MI — HWU) LMS 17 13 / 18
  • 15. A functional F P2(Rd ) → (−∞,∞] is λ-displacement convex iff F(µt) ≤ F(µ0) + F(µ1) − λt(1 − t)W2(µ0,µ1) where µt is the point t ∈ [0,1] along the constant speed geodesic between µ0 and µ1, i.e., W2(µt,µs) = (t − s)W2(µ0,µ1). If π ∈ P2 is λ-strongly log-concave (i.e., −log π strongly convex), then KL(µ,π) is λ-displacement convex w.r.t. µ. M. Pereyra (MI — HWU) LMS 17 14 / 18
  • 16. Proximal optimisation in the Wasserstein space Key Result (R. Jordan, D. Kinderlehrer, and F. Otto. 1998) The sequence of distributions generated by the iterative algorithm µ(t+1) = argmin ν∈P2(Rd ) KL(ν,π) + W2(µ(t) ,ν) (proximal-point method in P2 with metric W2) is equivalent to that of dXt = 1 2 ∇log p (Xt y)dt + dWt, 0 ≤ t ≤ T, X(0) ∼ µµ0, evaluated at times t = 1,... (their results use a different notation and involve the Fokker-Plank representation of the SDE). MYULA is essentially a sampling approximation of this algorithm for minimising KL(µ,πλ) where πλ is the Moreau approximation of π. M. Pereyra (MI — HWU) LMS 17 15 / 18
  • 17. Open problems: What do other (e.g., accelerated, primal-dual, etc.) optimisation algorithms look like in P2 with metric W2? What are the MCMC methods that we can derive from them? Do they inherit their good convergence properties? M. Pereyra (MI — HWU) LMS 17 16 / 18
  • 18. Thank you! Bibliography: Durmus, A., Moulines, E., and Pereyra, M. (2017). Efficient Bayesian computation by proximal Markov chain Monte Carlo: when Langevin meets Moreau. SIAM J. Imaging Sci. to appear. Zhu, L., Zhang, W., Elnatan, D., and Huang, B. (2012). Faster STORM using compressed sensing. Nat. Meth., 9(7):721–723. M. Pereyra (MI — HWU) LMS 17 17 / 18
  • 19. Comparison with PxMALA M. Pereyra (MI — HWU) LMS 17 18 / 18