Homepage › Solution manuals › Kevin P. Murphy › Machine Learning: a Probabilistic Perspective › Exercise 3.12 - MAP estimation for the Bernoulli with non-conjugate priors
Exercise 3.12 - MAP estimation for the Bernoulli with non-conjugate priors
Answers
For question (a), we adopt the different prior:
The posterior distribution now reads:
whose support is , so the MAP is:
For question (b), it is intuitive that the non-conjugate has better performance when is small. But the conjugate Bayesian method prevails with grows. For a solid verification, consider the case where is large, the probability that deviates from can be bounded by the Chernoff bounding. Let be a collection of i.i.d. Bernoulli random variables with distribution:
Denote as the random variable marks their summation. Then:
Where can be an arbitrary positive number. The probability that exceeds , denoted by is bounded by the lower bound of the last line in the deduction above. With , , the probability is numerically bounded by . The other side of error , whose probability is , can be derived in a similar way. The probability that the Bayesian way is dominated by the non-uniform prior is no higher than . Taking and , this bound remains negligible.