Whilst interesting, this post feels very assertive.
You claim that biological systems work by maintaining alignment as they scale. In what sense is this true?
You say that current methods lack a vision of a current whole. In what sense? There’s something extremely elegant about pre-training to learn a world model, doing supervised learning to select a sub-distribution and using RL to develop past the human level. In what sense does this “lack a vision”?
I’m open to the possibility that we need to align a model as we make it more intelligent to prevent the agent sabotaging the process. But it’s unclear from this article if this is why you want alignment first or for some other reason.
Whilst interesting, this post feels very assertive.
You claim that biological systems work by maintaining alignment as they scale. In what sense is this true?
You say that current methods lack a vision of a current whole. In what sense? There’s something extremely elegant about pre-training to learn a world model, doing supervised learning to select a sub-distribution and using RL to develop past the human level. In what sense does this “lack a vision”?
I’m open to the possibility that we need to align a model as we make it more intelligent to prevent the agent sabotaging the process. But it’s unclear from this article if this is why you want alignment first or for some other reason.