Discussion about this post

User's avatar
Sarah Dean's avatar

It's definitely interesting to consider why or when pure prediction (in the ML sense) is useful for science. One example that I think of often is the negative one: prediction of life outcomes (https://www.pnas.org/doi/10.1073/pnas.1915006117). What does it mean when our fanciest ML models and most extensive data to date *can't* make more accurate predictions that embarassingly simple models? There are definitely ethical/validity implications of this observation (https://predictive-optimization.cs.princeton.edu/), but maybe there are also scientific ones? It's also interesting to consider the role of mass collaboration/prediction competitions in justifying such a negative result.

Expand full comment
Bill Taylor's avatar

"Some criticize black-box predictive models as punting on scientific understanding, but we learn things when the predictions fail." Yes, but. When we are using large ML models and then trying to learn from their failures, we're studying something truly different. The underlying cause of such failures are by nature not human-understandable; and the improvements we make with more data or larger models are often nothing more than overtraining of the model to drive toward the single truth we are seeking... and not global truths which others will demand of the same model.

Take the example of a ML-based vision processor trying to detect human movers in a given space.

In our example, we may observe that a given model doesn't detect bicyclists very well. More training data with more labeled bicyclists may certainly improve the model on this axis, and we would rejoice... if we're in the bicyclists-detection business. But what has that updated training done to the other predictions?... of pedestrian walkers and resting/sitting/standing persons for example?

In simpler software (or big software performing clearly defined functions) we can run regression tests to ensure we've not stepped backwards. In modern meta-models, I'm not sure we are very good at regression testing. I am sure we're overfitting a lot of models in the name of improving them.

Expand full comment
5 more comments...

No posts