And even if somehow you could program an intelligence to optimize for those four competing utility functions at the same time, that would just cause it to optimize for conflict resolution, and then it would just tile the universe with tiny artificial conflicts between artificial agents for it to resolve as quickly and efficiently as possible without letting those agents do anything themselves.
I don’t believe an AI which simultaneously optimized multiple utility functions using a moral parliament approach would tile the universe with tiny artificial agents as described here.
“optimizing for competing utility functions” is not the same as optimizing for conflict resolution. There are various schemes for combining utility functions (some discussion on this podcast for instance). But let’s wave our hands a bit and say each of my utility functions outputs a binary approve/disapprove signal for any given action, and we choose randomly among those actions which are approved of by all of my utility functions. Then if even a single utility function doesn’t approve of the action “tile the universe with tiny artificial conflicts between artificial agents for it to resolve as quickly and efficiently as possible without letting those agents do anything themselves”, this action will not be done.
good point, I missed that, will fix later. more likely that effect would result from programming the AI with the overlap between those utility functions, but I’m not totally sure so I’ll have to think about it. I don’t think that point is actually necessary for the crux of my argument, though. Like I said, I’ll have to think about it. Right now it’s almost 4am and Im really sick now.
I don’t believe an AI which simultaneously optimized multiple utility functions using a moral parliament approach would tile the universe with tiny artificial agents as described here.
“optimizing for competing utility functions” is not the same as optimizing for conflict resolution. There are various schemes for combining utility functions (some discussion on this podcast for instance). But let’s wave our hands a bit and say each of my utility functions outputs a binary approve/disapprove signal for any given action, and we choose randomly among those actions which are approved of by all of my utility functions. Then if even a single utility function doesn’t approve of the action “tile the universe with tiny artificial conflicts between artificial agents for it to resolve as quickly and efficiently as possible without letting those agents do anything themselves”, this action will not be done.
good point, I missed that, will fix later. more likely that effect would result from programming the AI with the overlap between those utility functions, but I’m not totally sure so I’ll have to think about it. I don’t think that point is actually necessary for the crux of my argument, though. Like I said, I’ll have to think about it. Right now it’s almost 4am and Im really sick now.