Homepage Solution manuals Kevin P. Murphy Machine Learning: a Probabilistic Perspective Exercise 11.11 - Visible mixtures of Gaussians are in exponential family

Exercise 11.11 - Visible mixtures of Gaussians are in exponential family

Answers

Encode latent variable as binary code:

z k = 𝕀 ( x  is generated from the  k -th base distribution ) ,

then

p ( 𝐳 | 𝜃 ) = k = 1 K π k z k , p ( x | 𝐳 , 𝜃 ) = k = 1 K ( 1 2 π σ k 2 exp { 1 2 σ k 2 ( x μ k ) 2 } ) z k ,

where 𝜃 = ( π , μ , σ 2 ) .

The logarithm for the joint distribution is:

log p ( x , 𝐳 | 𝜃 ) = log k = 1 K ( π k 2 π σ k 2 exp { 1 2 σ k 2 ( x μ k ) 2 } ) z k = k = 1 K z k ( log π c 1 2 log 2 π σ k 2 1 2 σ k 2 ( x μ k ) 2 ) .

To reduce p ( x , 𝐳 | 𝜃 ) into the exponential family, note that log p ( x , 𝐳 | 𝜃 ) is linearly dependent on 𝐳 and x , hence we can rewrite:

ϕ ( x , 𝐳 ) = ( 𝐳 T , x 𝐳 T , x 2 𝐳 T ) T ,

the parameters for this form are:

( log π 1 2 log 2 π σ 2 , μ σ 2 , 1 2 σ 2 ) T

in vector form.

For the mixture of MVN, since:

log p ( 𝐱 , 𝐳 | 𝜃 ) = k = 1 K z k ( log π c D 2 log 2 π 1 2 log | Σ k | 1 2 ( 𝐱 μ k ) T Σ k 1 ( 𝐱 μ k ) ) 2 ,

so the distribution is still an exponential family member, with sufficient statistics:

ϕ ( 𝐱 , 𝐳 ) = ( 𝐳 , 𝐱 𝐳 , 𝐱 𝐱 𝐳 ) ,

rearranged as a vector. ( denotes the tensor/outer product.)

User profile picture
2021-03-24 13:42
Comments