Discussion about this post

User's avatar
Yuval Rabani's avatar

"Everyone in computer science has bought into the pact of evaporating oceans for better coding assistants."

That's because nobody in computer science likes coding, or teaching programming courses. Evaporating an ocean after teaching a class using Java is a mild reaction.

But some of us still think that better scaling means something other than exponential growth rate of energy bills or Nvidia's valuation.

Expand full comment
Taorui Wang's avatar

I don't mean to be aggressive. But I wonder if you test the most advanced model in openai by yourself with your own problem ? While Sutton's argument unfairly attaches his weird assumption to many researchers, my personal experience with the LLM is that they are improving greatly from GPT 3.5. I am doing computational math with application in various fields. I test the GPT 5 thinking and pro with long proof statements and concepts in stochastic control. It can consistently generalize those math deduction pattern from control problem to HJB and even consider the math constraints and potential weakness of its own arguments. There are some minor mistakes and error. But O1 and O3 cannot do it. You can argue that all those patterns occur in the training set and can be simulate with enough computation. But the "emergence" of computation capability from scaling those "stupid" algorithms is very strange. Maybe, we still don't know how far they can get with scaling.

Expand full comment
10 more comments...

No posts