Discussion about this post

User's avatar
Yaroslav Bulatov's avatar

As an engineer, I'm constantly disappointed by the amount of work wasted due to misleading claims. We have a feedback loop -- 1) too many papers to review carefully, 2) forcing reviewers to focus on a few hackable signals 3) leading authors to optimize for those signals instead of usefulness.

IE, you can get a paper published by claiming a new optimizer that beats Adam uniformly. Reviewers don't have time to try it themselves, so they let it through on face value. If they tested it, they probably would have objected to claims, independent reproduction effort has found 0 success (https://arxiv.org/pdf/2007.01547 ) .

A recent personal example, there's a specific work that several groups spent significant effort trying to reproduce. I ran into an author at a party and he told me I should stay away, because it doesn't really work (!). I couldn't tell that by reading the paper, evals looked impressive.

Expand full comment
Maxim Raginsky's avatar

As Lana put it in a recent talk she gave (with the absolutely glorious title “Non Serviam”), the problem is that Kids These Days have been conditioned to chase and optimize dense rewards instead of sparse rewards like in the olden days:

https://drive.google.com/file/d/1yjBpvvyxwHJvd99NdLk-d7io7dHtp1ZU/view?usp=drivesdk

Also, in the context of overproduction of CS papers, a couple of recent studies by Vladlen Koltun and collaborators:

1) https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0253397&trk=public_post_comment-text

2) https://arxiv.org/abs/2105.08089

Expand full comment
59 more comments...

No posts