Stuart Russell’s view seems to be similar to the one described by 180 in another comment: humans have preferences about how to do moral deliberation, and an IRL agent ought to learn to deliberate in a way that humans endorse, and then actually execute that deliberation, rather than directly learning arbitrarily complex values about e.g. population ethics.
(At least, I discussed this issue with him once and this was the impression I got, but I may have misunderstood.)
This view looks very reasonable to me. You and I have gone back and forth on this point a little bit but I don’t understand your position as well as I would like.
Stuart Russell’s view seems to be similar to the one described by 180 in another comment: humans have preferences about how to do moral deliberation, and an IRL agent ought to learn to deliberate in a way that humans endorse, and then actually execute that deliberation, rather than directly learning arbitrarily complex values about e.g. population ethics.
(At least, I discussed this issue with him once and this was the impression I got, but I may have misunderstood.)
This view looks very reasonable to me. You and I have gone back and forth on this point a little bit but I don’t understand your position as well as I would like.