It’s often called “indirect normativity”: a strategy in which instead of directly encoding the goal for an AI (or moral agent), we specify a certain way of “learning what to value/inferring human values” so that the AI can then deduce human values (and then implement it).
Ah, so it means the same thing as “value learning?” For some reason when I read “value loading” I thought of, like, overloading a function :D “I want cake, and that desire is also a carnal lust for BEES!”
What helped me was thinking of it in terms of: “Oh, like ‘reading’ human preferences as if they were an XML config file that the program loads at runtime.”
It’s more an FHI term than a SI/LessWrong term.
It’s often called “indirect normativity”: a strategy in which instead of directly encoding the goal for an AI (or moral agent), we specify a certain way of “learning what to value/inferring human values” so that the AI can then deduce human values (and then implement it).
Ah, so it means the same thing as “value learning?” For some reason when I read “value loading” I thought of, like, overloading a function :D “I want cake, and that desire is also a carnal lust for BEES!”
What helped me was thinking of it in terms of: “Oh, like ‘reading’ human preferences as if they were an XML config file that the program loads at runtime.”