Homepage › Solution manuals › Yaser Abu-Mostafa › Learning from Data › Exercise 4.7
Exercise 4.7
Answers
- (a)
- Note that the expectation w.r.t.
is equivalent to
because the
are assumed to be generated by a true .
- (b)
- In classification problem, .
We have
So the variance is:
- (c)
- In the end
- (d)
- The squared error is unbounded. The variance of it is also unbounded. So there’s no uniform upper bound for .
- (e)
- For regression with squared error, if we train using fewer points (smaller ) to get , then the resulting will be worse, the expectation of the squared error becomes larger. For continuous, non-negative random variables, higher mean often implies higher variance, so will be higher.
- (f)
- When we increasing the size of validation set
, the error
between
and
is .
It can drop in the case of classification. But for regression, it depends on which of
or
increases faster,
so the as an
estimate of
can become worse or better.
Does it mean for classification the estimate will always become better when we increase the K?
But note, the is only for the hypothesis , which can be pretty bad when is large. So for classification problem, even the error between and goes to zero, but the can be quite large.