Discussion about this post

User's avatar
Prateek Garg's avatar

Though imagining such stochastic processes help you build useful simulators of data.

Expand full comment
Rob Nowak's avatar

A common viewpoint is that there is a set of all possible data out there (e.g., all images on the web). We collect a subset of these data and train our model. The main (and likely unreasonable) assumption is that the training subset is an iid sample (or uniformly sampled w/o replacement) from the set of all possible data. This assumption is the central ingredient in theoretical generalization bounds. While it's probably not perfectly reasonable, it does give us a framework for comparing models and algorithms, and aligns with common practices like hold-out validation. I think this is a useful model of the (training) data-generating distribution.

Expand full comment
6 more comments...

No posts

Ready for more?