Thus, I think it’s reasonable to think of post-training as “privileging some pretrained behavioral patterns over other pretrained behavioral patterns”, rather than “developing new behavioral patterns from scratch”. Ditto for prompting, constitutional AI, and other such interventions.
If I thought this was true, then I wouldn’t think that scaling the reasoning models would lead to superintelligence.
If I thought this was true, then I wouldn’t think that scaling the reasoning models would lead to superintelligence.
Relevant paper: Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model?
YouTube explanatory video