I buy this as an argument that “we should put more effort into making philosophy work to make the outcome of AI better, because we risk losing large amounts of value” rather than “our efforts to get a pretty good future are doomed unless we make tons of progress on this” or something like that.
What about the other post I linked, Two Neglected Problems in Human-AI Safety? A lot more philosophical progress would be one way to solve those problems, and I don’t see many other options.
What about the other post I linked, Two Neglected Problems in Human-AI Safety? A lot more philosophical progress would be one way to solve those problems, and I don’t see many other options.