Claude Opus 4.6 and other frontier models have gotten really, impressively good without continual learning, so it is possible that isn’t strictly necessary.
If continual learning is required for AGI, then there’s a lot of understudied (potentially unstudyable?) risk there.
IMO, the threat/thing to measure is the system, so it doesn’t much matter what a badly run model can do. I’m with you here.