I feel like there’s so much agentic overhang. They seem to be way too over-trained for knowledge than for execution.
On one hand, this is “good”—they’re more like the Oracles/CAIs. But the gap is so glaring and there’s a sizeable demand for agentic AIs, so this imbalance basically forces the AI labs to go and adjust.
We would not like a “dumber” model in general and chopping the knowledge training is detrimental for grounding. This is demonstrated by the oX-mini series of models—they IMO suck to the point of being unusable and uninteresting.
So, I expect the upcoming models to not lose more smarts (like we’ve seen with 4.5->5 which got OpenAI into crosshairs) while gaining more in agency. GPT-5 is a visible step-up for me in this aspect. Even coming from o3 it is clearly more in control and more prepared for unagumented tasks.
I feel like there’s so much agentic overhang. They seem to be way too over-trained for knowledge than for execution.
On one hand, this is “good”—they’re more like the Oracles/CAIs. But the gap is so glaring and there’s a sizeable demand for agentic AIs, so this imbalance basically forces the AI labs to go and adjust.
We would not like a “dumber” model in general and chopping the knowledge training is detrimental for grounding. This is demonstrated by the
oX-mini
series of models—they IMO suck to the point of being unusable and uninteresting.So, I expect the upcoming models to not lose more smarts (like we’ve seen with 4.5->5 which got OpenAI into crosshairs) while gaining more in agency. GPT-5 is a visible step-up for me in this aspect. Even coming from o3 it is clearly more in control and more prepared for unagumented tasks.