Discussion about this post

User's avatar
Carl Boettiger's avatar

Thanks Ben, great piece as always!

I think an overlooked corollary to this is that you don't actually need a super LLM sophisticated LLM for tool calling to work. Even the largest LLMs still can't add reliably, but they've all learned not try, they just use calculators. But this is a game small/local/open LLMs can do to. Have you poked at the open LLM ecosystem for tool calls (err, "agents")?

Sure, if the tool is just "here's a bash shell" then yeah, an LLM needs to still be pretty clever. But you can give an LLM tools with more narrowly scoped and clearly explained uses, and voila, even a tiny model can suddenly be very powerful. The beautiful thing about this is, as you point out so nicely here, building a tool doesn't involve any GPUs or transformers, here in good conventional-software development land of JSON schemas and function calls. We've had great success building simple MCP tools that a model like gpt-oss or nvidia nemotron-3 can easily outperform what Opus can do with only the generic tools claude-code gives it...

Dragor's avatar
3hEdited

The subject matter of this essay was pretty interesting and seemed to explain something neat about the nature of LLMs (albeit one that my rudimentary and abandoned coding ability did not enable my sleep deprived brain to grant me understanding off. But. That subtitle! Uggh! I can’t say whether I believe or disbelieve what happened next, but I certainly believe that phrase is used in contexts where it is either untrue or the concept of dis/belief does not apply in a proportion of contexts that approximates to always. I felt serious desire to unsubscribe, but for seeing in that desire a certain petulance I would have. Uggh!

2 more comments...

No posts

Ready for more?