I don’t recall Eliezer (or anyone worth taking seriously with similar views) expressing confidence that whatever came after the previous generation of GPT would be anything in particular (except “probably not AGI”). Curious if you have a source for something like this.
Not claiming that, just, from my impressions, I thought their model was more like AlphaZero, with a clear utility function-style agency? Idk. Eliezer also talked about trusting an inscrutable matrix of floating point numbers, but the merging patterns, like grokking and interpretability seem unexpectedly close to how humans think.
Not claiming that, just, from my impressions, I thought their model was more like AlphaZero, with a clear utility function-style agency? Idk. Eliezer also talked about trusting an inscrutable matrix of floating point numbers, but the merging patterns, like grokking and interpretability seem unexpectedly close to how humans think.