Exercise 4.12 - BIC for Gaussians

Answers

For question (a), recall that the maximum likelihood estimation for a MVN model is:

μ MLE = 1 N n = 1 N 𝐱 n ,

Σ MLE = 1 N n = 1 N ( 𝐱 n μ MLE ) ( 𝐱 n μ MLE ) T .

So the likelihood reads:

p ( 𝒟 | μ MLE , Σ MLE ) = n = 1 N p ( 𝐱 n | μ MLE , Σ MLE ) = n = 1 N ( 2 π ) D 2 | Σ MLE | 1 2 exp ( 1 2 ( 𝐱 n μ MLE ) T Σ MLE 1 ( 𝐱 n μ MLE ) ) = ( 2 π ) 𝑁𝐷 2 | Σ MLE | N 2 exp ( 1 2 n = 1 N ( 𝐱 n μ MLE ) T Σ MLE 1 ( 𝐱 n μ MLE ) ) .

Denote:

𝐘 = ( 𝐱 1 μ MLE 𝐱 N μ MLE ) ,

then Σ MLE = 1 N 𝐘 𝐘 T , while the term in the exponential of the likelihood is:

1 2 tr ( 𝐘 T Σ MLE 1 𝐘 ) = 1 2 tr ( Σ MLE 1 𝐘 𝐘 T ) = 𝑁𝐷 2 .

Thus the BIC is:

𝑁𝐷 2 log ( 2 π e ) N 2 log | Σ MLE | D + D ( D + 1 ) 2 2 log N .

For question (b), the fitting of a diagonal MVN model is tantamount to fitting D independent 1d Gaussian models simultaneously, thus the d -th diagnoal component of Σ MLE d is:

1 N n = 1 N 𝐱 n , d 2 ,

where we have assumed 𝐱 = 0 w.l.o.g. Thus the term inside the exponential of the likelihood remains 𝑁𝐷 2 . So the BIC in this case is:

𝑁𝐷 2 log ( 2 π e ) N 2 log | Σ MLE d | D log N .

We observe that if all D components are mutually independent, i.e., Σ MLE is diagonal then the BIC for diagonal MVN model is strictly larger than that for general MVN, hence the diagonal version is always preferred. In cases there exists dependence among components, the BIC for general MVN is still not necessarily larger than that of disgonal MVN. This is a reflection of the trade-off between complexity and generalization.

The Bayesian information criterion is an approximation of a model’s evidence, p ( 𝒟 ) . Let us start from:

p ( 𝒟 ) = p ( 𝒟 | 𝜃 ) p ( 𝜃 ) d 𝜃 ,

where 𝜃 is the collection of all parameters within the current model. The trick here is to expand log p ( 𝐱 | 𝜃 ) as a function of 𝜃 and taking approximation to the second order at 𝜃 MAP so the first order gradient vanishes:

log p ( 𝐱 | 𝜃 ) log p ( 𝐱 | 𝜃 0 ) 1 2 ( 𝜃 𝜃 0 ) T 𝐇 ( 𝜃 𝜃 0 ) ,

where 𝐇 is the Hessian matrix at log p ( 𝐱 | 𝜃 0 ) . Thus we have:

p ( 𝐱 | 𝜃 ) p ( 𝐱 | 𝜃 0 ) exp ( 1 2 ( 𝜃 𝜃 0 ) T 𝐇 ( 𝜃 𝜃 0 ) ) .

We are now ready to perform the integral, with:

p ( 𝒟 | 𝜃 ) = p ( 𝐱 | 𝜃 0 ) N exp ( N 2 ( 𝜃 𝜃 0 ) T 𝐇 ( 𝜃 𝜃 0 ) ) ,

conducting the integral on the neighbour of 𝜃 0 = 𝜃 MAP :

p ( 𝒟 | 𝜃 ) p ( 𝜃 ) d 𝜃 p ( 𝒟 | 𝜃 MAP ) p ( 𝜃 MAP ) exp ( N 2 ( 𝜃 𝜃 MAP ) T 𝐇 ( 𝜃 𝜃 MAP ) ) d 𝜃 = p ( 𝒟 | 𝜃 MAP ) p ( 𝜃 MAP ) ( 2 π ) d 2 | N 1 𝐇 1 | 1 2 = p ( 𝒟 | 𝜃 MAP ) p ( 𝜃 MAP ) ( 2 π ) d 2 N d 2 | 𝐇 1 | 1 2 ,

where d is the number of components in 𝜃 . Taking the logarithm of both side of the evidence yields the BIC. One can see how many compromises and assumptions have been applied in deriving an analytic form of the evidence, which is arguably the most complex variable for Bayesian analysis.

See also PRML, Section 4.4.

User profile picture
2021-03-24 13:42
Comments