I sometime ask people why later, more powerful models will be agentic[1]. I think the most common cluster of reasons hangs out around “Meta-learning requires metacognition. Metacognition is, requires, or scales to agency.”
(Sometimes it could be generalisation rather than meta-learning or just high performance. And it might be other kinds of reasoning than metacognition)
I sometime ask people why later, more powerful models will be agentic[1]. I think the most common cluster of reasons hangs out around “Meta-learning requires metacognition. Metacognition is, requires, or scales to agency.”
(Sometimes it could be generalisation rather than meta-learning or just high performance. And it might be other kinds of reasoning than metacognition)
People vary in if they think it’s possible for scaled-up transformers to be powerful in this way