Discussion about this post

User's avatar
Luke Schoen's avatar

Counter,

prediction really is modeling and having a model really is one step from intelligence

its true that LLMs are not humans and don't work like humans but they can model our culture

math breaks LLMs because it's dense, you can't 'hide infinity competence' in a little box like you can with some other things (like summarization, knowledge extraction, etc).

That just means we have yet another layer of resource management to handle :P perhaps we always will.

uploading consciousness really is just predicting culture and to the extend that anyone needs AI it's here.

enjoy

Expand full comment
Jake's avatar

This seems like a matter of training. Early LLMs didn't know how to answer questions either - they just completed text where you would start writing something and it would finish it. But then we built corpora of "Instruction Tuning" to show example transcripts of chats and bias the system to generate content that was more useful in general context per the benchmarks that have been defined. We have the start of something similar with tool use or function calling where the models identify when they need to use a tool, such as a calculator. Early examples of this are generally pretty good. Likewise reasoning models are increasingly getting guided transcripts to train on how to break down problems to get more accurate results ... leveraging the same underlying training data (except for the reasoning training data) as more vanilla llms. The internet at large and books and such, don't often call out step-by-step instructions with, and "now plug this into a calculator" and to the extent they do, models didn't know how to literally do that with the function call, other than output the text, until recently. LLM based approaches might reach some limit where they can't address certain tasks like maths ... but right now that limit seems to be data. If we can can guide them to know how to approach the problem from a combination of reasoning, using external tools, introspection or feedback, and iteratively addressing the problem we can approach something similar to humans on a wide variety of problem domains. This is why there is a bunch of hype about "Agentic Systems"

Expand full comment
7 more comments...

No posts