Discussion about this post

User's avatar
Maxim Raginsky's avatar

I was an early partisan of the fundamental importance of the test set! In this 2005 NIPS paper with Lana on estimating the intrinsic dimension (our first paper together!), we had to use test-set estimates in order to correct for the negative bias of ERM, which led to deflated estimates in earlier work: https://proceedings.neurips.cc/paper_files/paper/2005/hash/892c3b1c6dccd52936e27cbd0ff683d6-Abstract.html.

Also, estimating the intrinsic dimension! Does anybody care about manifold learning anymore? Back then it was a thing (if only a minor thing), but here's also a bittersweet tidbit: Partha Niyogi was chairing the session in which I gave the talk and Sam Roweis asked the first question.

Expand full comment
Lior Fox's avatar

Isn't one factor in the explanation the fact that "classic" learning theory was mostly heavily biased to analyze "worst-case" scenarios? A randomized hold-out set is somewhat opposed to this approach (as it will implicitly be asking what happens if you made a misrepresentative / "bad" train-test split?)

On the other hand, it will probably also be hard to find explicit discussion of the hold-out evaluation practice in the works that tried to develop alternative theories of generalization, to overcome that limitation [I'm thinking of the Seung, Sompolinsky, Tishby "Statistical mechanics of learning from examples" paper/s, and related ideas].

Expand full comment
16 more comments...

No posts