This is not the same as CEV. CEV involves the AI extrapolating a user’s idealized future values and acting to implement them, even overriding current preferences if needed, whereas my model forbids that. In my framework, the AI never drives or predicts value change; it simply provides accurate world models and optimal plans based on the user’s current values, which only the user can update.
CEV also assumes convergence; my model protects normative autonomy and allows value diversity to persist.
CEV extrapolates the volition of humanity, that’s one reason it has to be “coherent”.
In your proposal, people have autonomy, but this principle can be violated in “extremely dangerous” situations. People are free to do what they want (“volition”)… but their AI advisors look ahead (“extrapolated”)… and people are not allowed to exercise their freedom so as to jeopardize the freedom of others (“coherent”).
I think this ends up being the same thing as CEV…
This is not the same as CEV. CEV involves the AI extrapolating a user’s idealized future values and acting to implement them, even overriding current preferences if needed, whereas my model forbids that. In my framework, the AI never drives or predicts value change; it simply provides accurate world models and optimal plans based on the user’s current values, which only the user can update.
CEV also assumes convergence; my model protects normative autonomy and allows value diversity to persist.
CEV extrapolates the volition of humanity, that’s one reason it has to be “coherent”.
In your proposal, people have autonomy, but this principle can be violated in “extremely dangerous” situations. People are free to do what they want (“volition”)… but their AI advisors look ahead (“extrapolated”)… and people are not allowed to exercise their freedom so as to jeopardize the freedom of others (“coherent”).