In his earlier thinking (~2023) he was also quite focused on non-standard approaches to AI existential safety, and it was clear that he was expecting to collaborate with advanced AI systems on that.
That’s an indirect evidence, but it does look like he is continuing in the same mindset.
It would be nice if his org finds ways to publish those aspects of their activity which might contribute to AI existential safety[[1]].
Since almost everyone is using “alignment” for “thing 2″ these days, I am trying to avoid the word; I doubt solving “thing 2” would contribute much to existential safety, and I can easily see how that might turn counterproductive instead.
In his earlier thinking (~2023) he was also quite focused on non-standard approaches to AI existential safety, and it was clear that he was expecting to collaborate with advanced AI systems on that.
That’s an indirect evidence, but it does look like he is continuing in the same mindset.
It would be nice if his org finds ways to publish those aspects of their activity which might contribute to AI existential safety [[1]] .
Since almost everyone is using “alignment” for “thing 2″ these days, I am trying to avoid the word; I doubt solving “thing 2” would contribute much to existential safety, and I can easily see how that might turn counterproductive instead.