Exercise 6.5

Answers

(a)
Since we are selecting hypothesis from a fixed set of Htrain, there are N K hypotheses and the validation data set is the ’input data set’, which has a size of K. We apply the generalization bound equation (2.1), and for any gk,k = 1,2,,N K, we have

Eout(gk) Eval(gk) + 1 2K ln 2(NK) δ

If we assume K log (NK) , then we have Eout(gk) Eval(gk)

Since g is the hypothesis with minimum validation error Eval(g), so we have

Eout(g) Eval(g) Eval(g) Eout(g)

On the other hand, g minimizes Eout, so we always have Eout(g) Eout(g)

Compare the two inequalities, we conclude Eout(g) Eout(g).

(b)
If N K , according to Theorem 6.2, we can find a k(N K), such that k(N K) and k(NK) NK 0, then we know that Ein(gk) Eout(gk) and Eout(gk) Eout.

Since Eout is the optimal out-of-sample error we can ever achieve, so we know that Eout(gk) Eout(g), by problem (a), we thus conclude Eout(g) Eout

(c)
If we used the k NN rule on the full data set D, we would see performance improvement because the learning curve tells us we should use more data to achieve better performance.
User profile picture
2021-12-08 09:39
Comments