SlideShare une entreprise Scribd logo
1  sur  1
Télécharger pour lire hors ligne
Non-informative reparametrisations for location-scale mixtures
Kaniav Kamary1, Kate Lee2, Christian P. Robert1,3
1CEREMADE, Université Paris–Dauphine, Paris 2Auckland University of Technology, New Zealand 3Dept. of Statistics, University of
Warwick, and CREST, Paris
Introduction
Traditional definition of mixture density:
f(x θ,p) =
k
∑
i=1
pif(x θi)
k
∑
i=1
pi = 1. (1)
which gives a separate meaning to each component.
For the location-scale Gaussian mixture:
f(x θ,p) =
k
∑
i=1
piN(x µi,σi)
Mengersen and Robert (1996) [2] established that an improper prior on (µ1,σ1) leads to a proper prior
when
µi = µi−1 + σi−1δi and σi = τiσi−1,τi < 1.
Diebolt and Robert (1994) [3] discussed the alternative approach of imposing proper posteriors on
improper priors by banning almost empty components from the likelihood function.
Setting global mean and variance Eθ,p(X) = µ and varθ,p(X) = σ2, imposes natural constraints on the
component parameters;
µ =
k
∑
i=1
piµi; σ2
=
k
∑
i=1
piµ2
i +
k
∑
i=1
piσ2
i − µ2
; Eθ,p(X2
) =
k
∑
i=1
piµ2
i +
k
∑
i=1
piσ2
i
which implies that (µ1,...,µk,σ1,...,σk) belongs to a specific ellipse.
New reparametrisation: Modifying the parameterization of the location-scale mixture in terms of
the global mean and variance of the mixture distribution.
Writing
f(x θ,p) =
k
∑
i=1
pif(x µ + σγi/
√
pi,σηi/
√
pi), (2)
leads a parameter space such that (p1,...,pk,γ1,...,γk,η1,...,ηk) is constrained by
pi,ηi ≥ 0 (1 ≤ i ≤ k)
k
∑
i=1
pi = 1
k
∑
i=1
√
piγi = 0
k
∑
i=1
{η2
i + γ2
i } = 1.
which implies ∀i 0 ≤ pi ≤ 1, 0 ≤ γi ≤ 1, 0 ≤ ηi ≤ 1. The constraints lead that (γ1,...,η) belongs to an
hypersphere of R2k centered at the origin with the radius of r = 1 intersected with an hyperplane of this
space passing the origin that results in a circle centered at the origin with radius 1.
Spherical coordinate representation of γ’s:
Suppose that ∑
k
i=1 γ2
i = ϕ2. The vector γ belongs
both to the hypersphere of radius ϕ and to the
hyperplane orthogonal to
√
pi;i = 1,...,k.
s-th orthogonal base Λs:
̃Λ1,j =
⎧⎪⎪
⎨
⎪⎪⎩
−
√
p2, j = 1
√
p1, j = 2
0, j > 2
s-th vector is given by
̃Λs,j =
⎧⎪⎪⎪⎪⎪⎪
⎨
⎪⎪⎪⎪⎪⎪⎩
−(pjps+1)1/2/(∑
s
l=1
pl)
1/2
, s > 1, j ≤ s
(∑
s
l=1
pl)
1/2
, s > 1, j = s + 1
0, s > 1, j > s + 1
and s-th orthonormal base is Fs = Λs/Λs .
Figure: Image from Robert Osserman.
(γ1,...,γk) can be written as
(γ1,...,γk) = ϕcos( 1)F1 + ϕsin( 1)cos( 2)F2 + ... + ϕsin( 1)⋯sin( k−2)Fk−1
with the angles 1,..., k−3 in [0,π] and k−2 in [0,2π].
Foundational consequences: The restriction is compact and helpful in selecting improper and
non-informative priors over mixtures.
Prior modeling:
Global mean and variance: The posterior distribution associated with the prior π(µ,σ) = 1/σ is proper
when (a) proper distributions are used on the other parameters and (b) there are at least two
observations in the sample.
Component weights: (p1,...,pk) ∼ Dir(α0,...,α0),
Angles ’s: 1,..., k−3 ∼ U[0,π] and k ∼ U[0,2π],
Raduis ϕ and η1,...,ηk: If k is small, (ϕ2,η2
1,...,η2
k) ∼ Dir(α,...,α) while for k more than 3, (η1,...,ηk)
is written through spherical coordinates
ηi =
⎧⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪
⎨
⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎩
√
1 − ϕ2 cos(ξi), i = 1
√
1 − ϕ2
i−1
∏
j=1
sin(ξj)cos(ξi), 1 < i < k
√
1 − ϕ2
i−1
∏
j=1
sin(ξj), i = k
Unlike , the support for all angles ξ1,⋯,ξk−1 is limited to [0,π/2], due to the positivity requirement on
the ηi’s.
(ξ1,⋯,ξk−1) ∼ U([0,π/2]k−1
).
MCMC algorithm
Metropolis-within-Gibbs algorithm for reparameterised mixture model:
1 Generate initial values (µ(0),σ(0),p(0),ϕ(0),ξ
(0)
1 ,...,ξ
(0)
k−1,
(0)
1 ,...,
(0)
k−2).
2 For t = 1,...,T, the update of (µ(t),σ(t),p(t),ϕ(t),ξ
(t)
1 ,...,ξ
(t)
k−1,
(t)
1 ,...,
(t)
k−2)
follows;
2.1 Generate a proposal µ′ ∼ N(µ(t−1),εµ) and update µ(t) against
π(⋅ x,σ(t−1)
,p(t−1)
,ϕ(t−1)
,ξ(t−1)
, (t−1)
).
2.2 Generate a proposal log(σ)′ ∼ N(log(σ(t−1)),εσ) and update σ(t) against
π(⋅ x,µ(t)
,p(t−1)
,ϕ(t−1)
,ξ(t−1)
, (t−1)
).
2.3 Generate a proposal (ϕ2)′ ∼ Beta((ϕ2)(t)εϕ + 1,(1 − (ϕ2)(t))εϕ + 1) and update ϕ(t) against
π(⋅ x,µ(t)
,σ(t)
,p(t−1)
,ξ(t)
, (t)
).
2.4 Generate a proposal p′ ∼ Dir(p
(t−1)
1 εp + 1,...,p
(t−1)
k εp + 1), and update p(t) against
π(⋅ x,µ(t)
,σ(t)
,ϕ(t)
,ξ(t)
, (t)
).
2.5 Generate proposals ξ′
i ∼ U[ξ
(t)
i − εξ,ξ
(t)
i + εξ], i = 1,⋯,k − 1, and update (ξ
(t)
1 ,...,ξ
(t)
k−1) against
π(⋅ x,µ(t)
,σ(t)
,p(t)
,ϕ(t)
, (t)
).
2.6 Generate proposals ′
i ∼ U[
(t)
i − ε ,
(t)
i + ε ], i = 1,⋯,k − 2, and update (
(t)
1 ,...,
(t)
k−2) against
π(⋅ x,µ(t)
,σ(t)
,p(t)
,ϕ(t)
,ξ(t)
).
where p(t) = (p
(t)
1 ,...,p
(t)
k ), x = (x1,...,xn), ξ(t) = (ξ
(t)
1 ,...,ξ
(t)
k−1) and (t) = (
(t)
1 ,...,
(t)
k−2).
Ultimixt package
▸ Implementation of the Metropolis-within-Gibbs algorithm for reparametrized mixture distribution;
▸ Calibrate the scales of the various proposals by aiming an average acceptance rate of either 0.44 or 0.234
depending on the dimension of the simulated parameter;
▸ Accurately estimate the component parameters;
Point estimator of the component parameters in the case of label switching:
▸ K-means clustering algorithm;
▸ Reordering labels towards producing the shortest distance between the current posterior sample and the
(or a) maximum posterior probability (MAP) estimate; [1].
Mixture of two normal distributions
A sample of size 50 simulated from .65N(−8,2) + .35N(−.5,1),
Figure: Empirical densities of 10 sequences of running Metropolis-within-Gibbs algorithm in parallel with 2e + 05 iterations.
▸ Outcomes of 10 parallel chains started
randomly from different starting values,
are indistinguishable;
▸ Chains are well-mixed;
▸ Sampler output covers the entire
sample space;
▸ Estimated densities converge to a
neighborhood of the true values;
▸ Estimated mixture density is remarkably
smooth;
Mixture of three normal distributions
A sample of size 50 is simulated from model .27N(−4.5,1) + .4N(10,1) + .33N(3,1)
Figure: Sequences of µi,σi and pi and estimated mixture density; mixture density estimate based on 104 MCMC iterations
Overfitting case
Extreme valued posterior samples for an overfitted model.
Galaxy dataset: Point estimator of the parameters of a mixture of (Left) 6 components; (Right) 4 components.
References
[1] S. Früwirth. Schnatter. (2001). Markov chain Monte Carlo estimation of classical and dynamic switching
and mixture models. J. American Statist. Assoc., 96 194–209.
[2] K. Mengersen and C. Robert. (1996) Testing for mixtures: A Bayesian entropic approach (with discussion).
In Bayesian Statistics 5 (J. Berger, J. Bernardo, A. Dawid, D. Lindley and A. Smith, eds). Oxford University
Press, Oxford, 255–276.
[3] J. Diebolt and C. Robert. (1994) Estimation of finite mixture distributions by Bayesian sampling. J. Royal
Statist. Society Series B, 56 363–375.
kamary@ceremade.dauphine.fr

Contenu connexe

Tendances

Generating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceGenerating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceCheng-An Yang
 
Harmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningHarmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningSungbin Lim
 
Testing for mixtures by seeking components
Testing for mixtures by seeking componentsTesting for mixtures by seeking components
Testing for mixtures by seeking componentsChristian Robert
 
Deep generative model.pdf
Deep generative model.pdfDeep generative model.pdf
Deep generative model.pdfHyungjoo Cho
 
Coordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like samplerCoordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like samplerChristian Robert
 
Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)Matthew Leingang
 
Lesson 27: Integration by Substitution (Section 041 slides)
Lesson 27: Integration by Substitution (Section 041 slides)Lesson 27: Integration by Substitution (Section 041 slides)
Lesson 27: Integration by Substitution (Section 041 slides)Matthew Leingang
 
Lesson 27: Evaluating Definite Integrals
Lesson 27: Evaluating Definite IntegralsLesson 27: Evaluating Definite Integrals
Lesson 27: Evaluating Definite IntegralsMatthew Leingang
 
Bregman divergences from comparative convexity
Bregman divergences from comparative convexityBregman divergences from comparative convexity
Bregman divergences from comparative convexityFrank Nielsen
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Matthew Leingang
 

Tendances (19)

Generating Chebychev Chaotic Sequence
Generating Chebychev Chaotic SequenceGenerating Chebychev Chaotic Sequence
Generating Chebychev Chaotic Sequence
 
Harmonic Analysis and Deep Learning
Harmonic Analysis and Deep LearningHarmonic Analysis and Deep Learning
Harmonic Analysis and Deep Learning
 
Testing for mixtures by seeking components
Testing for mixtures by seeking componentsTesting for mixtures by seeking components
Testing for mixtures by seeking components
 
Matrix calculus
Matrix calculusMatrix calculus
Matrix calculus
 
Deep generative model.pdf
Deep generative model.pdfDeep generative model.pdf
Deep generative model.pdf
 
Fourier series 2
Fourier series 2Fourier series 2
Fourier series 2
 
Coordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like samplerCoordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like sampler
 
Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)Lesson 25: Evaluating Definite Integrals (slides)
Lesson 25: Evaluating Definite Integrals (slides)
 
Chapter 4 (maths 3)
Chapter 4 (maths 3)Chapter 4 (maths 3)
Chapter 4 (maths 3)
 
Chapter 2 (maths 3)
Chapter 2 (maths 3)Chapter 2 (maths 3)
Chapter 2 (maths 3)
 
Lesson 27: Integration by Substitution (Section 041 slides)
Lesson 27: Integration by Substitution (Section 041 slides)Lesson 27: Integration by Substitution (Section 041 slides)
Lesson 27: Integration by Substitution (Section 041 slides)
 
Fourier series
Fourier seriesFourier series
Fourier series
 
Lesson 27: Evaluating Definite Integrals
Lesson 27: Evaluating Definite IntegralsLesson 27: Evaluating Definite Integrals
Lesson 27: Evaluating Definite Integrals
 
Chris Sherlock's slides
Chris Sherlock's slidesChris Sherlock's slides
Chris Sherlock's slides
 
the ABC of ABC
the ABC of ABCthe ABC of ABC
the ABC of ABC
 
Intro to ABC
Intro to ABCIntro to ABC
Intro to ABC
 
Bregman divergences from comparative convexity
Bregman divergences from comparative convexityBregman divergences from comparative convexity
Bregman divergences from comparative convexity
 
Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)Lesson 26: The Fundamental Theorem of Calculus (slides)
Lesson 26: The Fundamental Theorem of Calculus (slides)
 
Chapter 3 (maths 3)
Chapter 3 (maths 3)Chapter 3 (maths 3)
Chapter 3 (maths 3)
 

Similaire à Non-informative reparametrisation for location-scale mixtures

Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
 
Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesFrank Nielsen
 
ABC with Wasserstein distances
ABC with Wasserstein distancesABC with Wasserstein distances
ABC with Wasserstein distancesChristian Robert
 
On Clustering Histograms with k-Means by Using Mixed α-Divergences
 On Clustering Histograms with k-Means by Using Mixed α-Divergences On Clustering Histograms with k-Means by Using Mixed α-Divergences
On Clustering Histograms with k-Means by Using Mixed α-DivergencesFrank Nielsen
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixturesChristian Robert
 
Integration of biological annotations using hierarchical modeling
Integration of biological annotations using hierarchical modelingIntegration of biological annotations using hierarchical modeling
Integration of biological annotations using hierarchical modelingUSC
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt msFaeco Bot
 
The Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is DiagonalizableThe Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is DiagonalizableJay Liew
 
SMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionSMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionLilyana Vankova
 
prior selection for mixture estimation
prior selection for mixture estimationprior selection for mixture estimation
prior selection for mixture estimationChristian Robert
 
Talk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniquesTalk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniquesPierre Jacob
 
Slides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingSlides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingFrank Nielsen
 
A new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributionsA new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributionsFrank Nielsen
 

Similaire à Non-informative reparametrisation for location-scale mixtures (20)

BAYSM'14, Wien, Austria
BAYSM'14, Wien, AustriaBAYSM'14, Wien, Austria
BAYSM'14, Wien, Austria
 
Vancouver18
Vancouver18Vancouver18
Vancouver18
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
 
Patch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective DivergencesPatch Matching with Polynomial Exponential Families and Projective Divergences
Patch Matching with Polynomial Exponential Families and Projective Divergences
 
Igv2008
Igv2008Igv2008
Igv2008
 
ABC with Wasserstein distances
ABC with Wasserstein distancesABC with Wasserstein distances
ABC with Wasserstein distances
 
On Clustering Histograms with k-Means by Using Mixed α-Divergences
 On Clustering Histograms with k-Means by Using Mixed α-Divergences On Clustering Histograms with k-Means by Using Mixed α-Divergences
On Clustering Histograms with k-Means by Using Mixed α-Divergences
 
Bayesian inference on mixtures
Bayesian inference on mixturesBayesian inference on mixtures
Bayesian inference on mixtures
 
Integration of biological annotations using hierarchical modeling
Integration of biological annotations using hierarchical modelingIntegration of biological annotations using hierarchical modeling
Integration of biological annotations using hierarchical modeling
 
Multivriada ppt ms
Multivriada   ppt msMultivriada   ppt ms
Multivriada ppt ms
 
The Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is DiagonalizableThe Probability that a Matrix of Integers Is Diagonalizable
The Probability that a Matrix of Integers Is Diagonalizable
 
SMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last versionSMB_2012_HR_VAN_ST-last version
SMB_2012_HR_VAN_ST-last version
 
prior selection for mixture estimation
prior selection for mixture estimationprior selection for mixture estimation
prior selection for mixture estimation
 
Talk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniquesTalk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniques
 
Slides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingSlides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processing
 
Quantum chaos of generic systems - Marko Robnik
Quantum chaos of generic systems - Marko RobnikQuantum chaos of generic systems - Marko Robnik
Quantum chaos of generic systems - Marko Robnik
 
Modeling the dynamics of molecular concentration during the diffusion procedure
Modeling the dynamics of molecular concentration during the  diffusion procedureModeling the dynamics of molecular concentration during the  diffusion procedure
Modeling the dynamics of molecular concentration during the diffusion procedure
 
The Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal FunctionThe Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal Function
 
A new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributionsA new implementation of k-MLE for mixture modelling of Wishart distributions
A new implementation of k-MLE for mixture modelling of Wishart distributions
 
Bayes gauss
Bayes gaussBayes gauss
Bayes gauss
 

Plus de Christian Robert

Asymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de FranceAsymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de FranceChristian Robert
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinChristian Robert
 
How many components in a mixture?
How many components in a mixture?How many components in a mixture?
How many components in a mixture?Christian Robert
 
Testing for mixtures at BNP 13
Testing for mixtures at BNP 13Testing for mixtures at BNP 13
Testing for mixtures at BNP 13Christian Robert
 
Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?Christian Robert
 
discussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihooddiscussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihoodChristian Robert
 
NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)Christian Robert
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerChristian Robert
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Christian Robert
 
Likelihood-free Design: a discussion
Likelihood-free Design: a discussionLikelihood-free Design: a discussion
Likelihood-free Design: a discussionChristian Robert
 
CISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergenceCISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergenceChristian Robert
 
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment modelsa discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment modelsChristian Robert
 

Plus de Christian Robert (20)

Asymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de FranceAsymptotics of ABC, lecture, Collège de France
Asymptotics of ABC, lecture, Collège de France
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael Martin
 
discussion of ICML23.pdf
discussion of ICML23.pdfdiscussion of ICML23.pdf
discussion of ICML23.pdf
 
How many components in a mixture?
How many components in a mixture?How many components in a mixture?
How many components in a mixture?
 
restore.pdf
restore.pdfrestore.pdf
restore.pdf
 
Testing for mixtures at BNP 13
Testing for mixtures at BNP 13Testing for mixtures at BNP 13
Testing for mixtures at BNP 13
 
Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?Inferring the number of components: dream or reality?
Inferring the number of components: dream or reality?
 
CDT 22 slides.pdf
CDT 22 slides.pdfCDT 22 slides.pdf
CDT 22 slides.pdf
 
discussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihooddiscussion on Bayesian restricted likelihood
discussion on Bayesian restricted likelihood
 
NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)NCE, GANs & VAEs (and maybe BAC)
NCE, GANs & VAEs (and maybe BAC)
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like sampler
 
eugenics and statistics
eugenics and statisticseugenics and statistics
eugenics and statistics
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
asymptotics of ABC
asymptotics of ABCasymptotics of ABC
asymptotics of ABC
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
Likelihood-free Design: a discussion
Likelihood-free Design: a discussionLikelihood-free Design: a discussion
Likelihood-free Design: a discussion
 
CISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergenceCISEA 2019: ABC consistency and convergence
CISEA 2019: ABC consistency and convergence
 
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment modelsa discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
a discussion of Chib, Shin, and Simoni (2017-8) Bayesian moment models
 

Dernier

Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPirithiRaju
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticssakshisoni2385
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSarthak Sekhar Mondal
 
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsHubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsSérgio Sacani
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
Disentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTDisentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTSérgio Sacani
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfSumit Kumar yadav
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfSumit Kumar yadav
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bSérgio Sacani
 
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptxUnlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptxanandsmhk
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...ssifa0344
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PPRINCE C P
 
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRStunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRDelhi Call girls
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)Areesha Ahmad
 
Formation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksFormation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksSérgio Sacani
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
Biological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdfBiological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdfmuntazimhurra
 
Green chemistry and Sustainable development.pptx
Green chemistry  and Sustainable development.pptxGreen chemistry  and Sustainable development.pptx
Green chemistry and Sustainable development.pptxRajatChauhan518211
 

Dernier (20)

Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdfPests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
Pests of cotton_Borer_Pests_Binomics_Dr.UPR.pdf
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
 
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatidSpermiogenesis or Spermateleosis or metamorphosis of spermatid
Spermiogenesis or Spermateleosis or metamorphosis of spermatid
 
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroidsHubble Asteroid Hunter III. Physical properties of newly found asteroids
Hubble Asteroid Hunter III. Physical properties of newly found asteroids
 
Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
Disentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTDisentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOST
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdf
 
Zoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdfZoology 4th semester series (krishna).pdf
Zoology 4th semester series (krishna).pdf
 
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 bAsymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
Asymmetry in the atmosphere of the ultra-hot Jupiter WASP-76 b
 
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptxUnlocking  the Potential: Deep dive into ocean of Ceramic Magnets.pptx
Unlocking the Potential: Deep dive into ocean of Ceramic Magnets.pptx
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
TEST BANK For Radiologic Science for Technologists, 12th Edition by Stewart C...
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C P
 
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCRStunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
Stunning ➥8448380779▻ Call Girls In Panchshil Enclave Delhi NCR
 
GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)GBSN - Microbiology (Unit 2)
GBSN - Microbiology (Unit 2)
 
Formation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksFormation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disks
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
Biological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdfBiological Classification BioHack (3).pdf
Biological Classification BioHack (3).pdf
 
Green chemistry and Sustainable development.pptx
Green chemistry  and Sustainable development.pptxGreen chemistry  and Sustainable development.pptx
Green chemistry and Sustainable development.pptx
 
CELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdfCELL -Structural and Functional unit of life.pdf
CELL -Structural and Functional unit of life.pdf
 

Non-informative reparametrisation for location-scale mixtures

  • 1. Non-informative reparametrisations for location-scale mixtures Kaniav Kamary1, Kate Lee2, Christian P. Robert1,3 1CEREMADE, Université Paris–Dauphine, Paris 2Auckland University of Technology, New Zealand 3Dept. of Statistics, University of Warwick, and CREST, Paris Introduction Traditional definition of mixture density: f(x θ,p) = k ∑ i=1 pif(x θi) k ∑ i=1 pi = 1. (1) which gives a separate meaning to each component. For the location-scale Gaussian mixture: f(x θ,p) = k ∑ i=1 piN(x µi,σi) Mengersen and Robert (1996) [2] established that an improper prior on (µ1,σ1) leads to a proper prior when µi = µi−1 + σi−1δi and σi = τiσi−1,τi < 1. Diebolt and Robert (1994) [3] discussed the alternative approach of imposing proper posteriors on improper priors by banning almost empty components from the likelihood function. Setting global mean and variance Eθ,p(X) = µ and varθ,p(X) = σ2, imposes natural constraints on the component parameters; µ = k ∑ i=1 piµi; σ2 = k ∑ i=1 piµ2 i + k ∑ i=1 piσ2 i − µ2 ; Eθ,p(X2 ) = k ∑ i=1 piµ2 i + k ∑ i=1 piσ2 i which implies that (µ1,...,µk,σ1,...,σk) belongs to a specific ellipse. New reparametrisation: Modifying the parameterization of the location-scale mixture in terms of the global mean and variance of the mixture distribution. Writing f(x θ,p) = k ∑ i=1 pif(x µ + σγi/ √ pi,σηi/ √ pi), (2) leads a parameter space such that (p1,...,pk,γ1,...,γk,η1,...,ηk) is constrained by pi,ηi ≥ 0 (1 ≤ i ≤ k) k ∑ i=1 pi = 1 k ∑ i=1 √ piγi = 0 k ∑ i=1 {η2 i + γ2 i } = 1. which implies ∀i 0 ≤ pi ≤ 1, 0 ≤ γi ≤ 1, 0 ≤ ηi ≤ 1. The constraints lead that (γ1,...,η) belongs to an hypersphere of R2k centered at the origin with the radius of r = 1 intersected with an hyperplane of this space passing the origin that results in a circle centered at the origin with radius 1. Spherical coordinate representation of γ’s: Suppose that ∑ k i=1 γ2 i = ϕ2. The vector γ belongs both to the hypersphere of radius ϕ and to the hyperplane orthogonal to √ pi;i = 1,...,k. s-th orthogonal base Λs: ̃Λ1,j = ⎧⎪⎪ ⎨ ⎪⎪⎩ − √ p2, j = 1 √ p1, j = 2 0, j > 2 s-th vector is given by ̃Λs,j = ⎧⎪⎪⎪⎪⎪⎪ ⎨ ⎪⎪⎪⎪⎪⎪⎩ −(pjps+1)1/2/(∑ s l=1 pl) 1/2 , s > 1, j ≤ s (∑ s l=1 pl) 1/2 , s > 1, j = s + 1 0, s > 1, j > s + 1 and s-th orthonormal base is Fs = Λs/Λs . Figure: Image from Robert Osserman. (γ1,...,γk) can be written as (γ1,...,γk) = ϕcos( 1)F1 + ϕsin( 1)cos( 2)F2 + ... + ϕsin( 1)⋯sin( k−2)Fk−1 with the angles 1,..., k−3 in [0,π] and k−2 in [0,2π]. Foundational consequences: The restriction is compact and helpful in selecting improper and non-informative priors over mixtures. Prior modeling: Global mean and variance: The posterior distribution associated with the prior π(µ,σ) = 1/σ is proper when (a) proper distributions are used on the other parameters and (b) there are at least two observations in the sample. Component weights: (p1,...,pk) ∼ Dir(α0,...,α0), Angles ’s: 1,..., k−3 ∼ U[0,π] and k ∼ U[0,2π], Raduis ϕ and η1,...,ηk: If k is small, (ϕ2,η2 1,...,η2 k) ∼ Dir(α,...,α) while for k more than 3, (η1,...,ηk) is written through spherical coordinates ηi = ⎧⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪ ⎨ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎩ √ 1 − ϕ2 cos(ξi), i = 1 √ 1 − ϕ2 i−1 ∏ j=1 sin(ξj)cos(ξi), 1 < i < k √ 1 − ϕ2 i−1 ∏ j=1 sin(ξj), i = k Unlike , the support for all angles ξ1,⋯,ξk−1 is limited to [0,π/2], due to the positivity requirement on the ηi’s. (ξ1,⋯,ξk−1) ∼ U([0,π/2]k−1 ). MCMC algorithm Metropolis-within-Gibbs algorithm for reparameterised mixture model: 1 Generate initial values (µ(0),σ(0),p(0),ϕ(0),ξ (0) 1 ,...,ξ (0) k−1, (0) 1 ,..., (0) k−2). 2 For t = 1,...,T, the update of (µ(t),σ(t),p(t),ϕ(t),ξ (t) 1 ,...,ξ (t) k−1, (t) 1 ,..., (t) k−2) follows; 2.1 Generate a proposal µ′ ∼ N(µ(t−1),εµ) and update µ(t) against π(⋅ x,σ(t−1) ,p(t−1) ,ϕ(t−1) ,ξ(t−1) , (t−1) ). 2.2 Generate a proposal log(σ)′ ∼ N(log(σ(t−1)),εσ) and update σ(t) against π(⋅ x,µ(t) ,p(t−1) ,ϕ(t−1) ,ξ(t−1) , (t−1) ). 2.3 Generate a proposal (ϕ2)′ ∼ Beta((ϕ2)(t)εϕ + 1,(1 − (ϕ2)(t))εϕ + 1) and update ϕ(t) against π(⋅ x,µ(t) ,σ(t) ,p(t−1) ,ξ(t) , (t) ). 2.4 Generate a proposal p′ ∼ Dir(p (t−1) 1 εp + 1,...,p (t−1) k εp + 1), and update p(t) against π(⋅ x,µ(t) ,σ(t) ,ϕ(t) ,ξ(t) , (t) ). 2.5 Generate proposals ξ′ i ∼ U[ξ (t) i − εξ,ξ (t) i + εξ], i = 1,⋯,k − 1, and update (ξ (t) 1 ,...,ξ (t) k−1) against π(⋅ x,µ(t) ,σ(t) ,p(t) ,ϕ(t) , (t) ). 2.6 Generate proposals ′ i ∼ U[ (t) i − ε , (t) i + ε ], i = 1,⋯,k − 2, and update ( (t) 1 ,..., (t) k−2) against π(⋅ x,µ(t) ,σ(t) ,p(t) ,ϕ(t) ,ξ(t) ). where p(t) = (p (t) 1 ,...,p (t) k ), x = (x1,...,xn), ξ(t) = (ξ (t) 1 ,...,ξ (t) k−1) and (t) = ( (t) 1 ,..., (t) k−2). Ultimixt package ▸ Implementation of the Metropolis-within-Gibbs algorithm for reparametrized mixture distribution; ▸ Calibrate the scales of the various proposals by aiming an average acceptance rate of either 0.44 or 0.234 depending on the dimension of the simulated parameter; ▸ Accurately estimate the component parameters; Point estimator of the component parameters in the case of label switching: ▸ K-means clustering algorithm; ▸ Reordering labels towards producing the shortest distance between the current posterior sample and the (or a) maximum posterior probability (MAP) estimate; [1]. Mixture of two normal distributions A sample of size 50 simulated from .65N(−8,2) + .35N(−.5,1), Figure: Empirical densities of 10 sequences of running Metropolis-within-Gibbs algorithm in parallel with 2e + 05 iterations. ▸ Outcomes of 10 parallel chains started randomly from different starting values, are indistinguishable; ▸ Chains are well-mixed; ▸ Sampler output covers the entire sample space; ▸ Estimated densities converge to a neighborhood of the true values; ▸ Estimated mixture density is remarkably smooth; Mixture of three normal distributions A sample of size 50 is simulated from model .27N(−4.5,1) + .4N(10,1) + .33N(3,1) Figure: Sequences of µi,σi and pi and estimated mixture density; mixture density estimate based on 104 MCMC iterations Overfitting case Extreme valued posterior samples for an overfitted model. Galaxy dataset: Point estimator of the parameters of a mixture of (Left) 6 components; (Right) 4 components. References [1] S. Früwirth. Schnatter. (2001). Markov chain Monte Carlo estimation of classical and dynamic switching and mixture models. J. American Statist. Assoc., 96 194–209. [2] K. Mengersen and C. Robert. (1996) Testing for mixtures: A Bayesian entropic approach (with discussion). In Bayesian Statistics 5 (J. Berger, J. Bernardo, A. Dawid, D. Lindley and A. Smith, eds). Oxford University Press, Oxford, 255–276. [3] J. Diebolt and C. Robert. (1994) Estimation of finite mixture distributions by Bayesian sampling. J. Royal Statist. Society Series B, 56 363–375. kamary@ceremade.dauphine.fr