I think you’ve really hit the nail on the head on what’s wrong (and right) with the MIRI approach. The Cartesian Frames stuff seems to be the best stuff they’ve done in this direction.
I’ve also felt that our lack of understanding of abstraction is one of the key bottlenecks. How concerned are you about insights on this question also applying to unaligned AGI development?
How concerned are you about insights on this question also applying to unaligned AGI development?
Enough that I have considered keeping it secret, but I think keeping it public is a strong net positive relative to our current state (i.e. giant inscrutable vectors of floating-points). If there were, say, another AI winter, then I could easily imagine changing my mind about that.
I think you’ve really hit the nail on the head on what’s wrong (and right) with the MIRI approach. The Cartesian Frames stuff seems to be the best stuff they’ve done in this direction.
I’ve also felt that our lack of understanding of abstraction is one of the key bottlenecks. How concerned are you about insights on this question also applying to unaligned AGI development?
Enough that I have considered keeping it secret, but I think keeping it public is a strong net positive relative to our current state (i.e. giant inscrutable vectors of floating-points). If there were, say, another AI winter, then I could easily imagine changing my mind about that.