Double indifference is better indifference

A putative new idea for AI control; index here.

Note: this post replaces and supersedes the previous one which is now obsolete.

Hacking utility indifference

Utility indifference was a technique for corrigibility was an attempt to allow an AI to safely change its values, without seeking to provoke or avoid the change. The idea is that, when the AI’s utility changes from to at time , it maximises a meta-utility such that

  • .

  • .


Here designates the event that remains unchanged, while designates the change. This is a permitted utility change—one that we have set the agent up to accept—rather than just any old change in utility, which the agent will still attempt to resist.

As has been shown, adding those expectation terms means the AI will not seek to provoke or resist such a utility change—it is indifferent to the change. Those expectation terms can be designated compensatory rewards.

What it is not indifferent to, though, is changes to . If the AI knows that change is coming, then it has incentives to change to with a high and a low . Its ability to rewire is limited by the fact that will serve as its future expectation estimator, so it won’t mangle it completely, but it certainly can twist it a bit.

Notice that the AI is now no longer indifferent to the permitted change of utility. It can benefit from rewiring this way only if a change in utility is coming up.

Probability estimators

Let be probability estimators. These define both expectation estimators and probability estimators . These two are equivalent in principle, since for any random variable with the indicator function for , then and .

Double indifference

One idea to combat this is to apply indifference to the process of changing probability estimators. Thus an AI with estimator changing to would get compensatory reward

  • .

This won’t work, however, since we don’t want to discount all the impacts of the change (maybe is a much more accurate estimator), just the part of the change that affects the compensatory rewards. So the second layer of indifference will be a bit different from the previous one.

First, a subtlety. If we designate by the fact that the agent uses estimator , it is very possible that . This simply means that, in ‘s ‘opinion’, is a better estimator for maximising than it is itself.

Now, assume it’s at a time where the agent is using estimator and current utility . Let be all future compensatory rewards of the type , for permitted change . Let be the same, except the expectation estimator is : thus it is . Note that .

Then define (and ) by summing (and ) over all , , and .

Let be all future compensatory rewards that are not of the type . Then if the agent changes into , it gets the compensatory reward

  • .

Note that is defined in terms of , so, though there may be convergence issues, the definition is not circular.

Then the first two terms of cancel out, according to the agent’s best current expectation, all future compensatory rewards. The third term replaces all the type compensatory rewards with the same, except estimated by the agent’s current rather than by whichever estimator the agent will have at the time. Thus, the agent will be indifferent to all actual compensatory rewards (as it gets none in expectation), and instead gain estimated rewards that make it indifferent to any future change .

See this post to see the same formula in AIXI formalism.

There are certainly cases where the above terms converge (such as all bounded , finite number of possible changes to both utility and ). A full discussion of convergence will wait for another post.

Sequentially unbiased estimators

Call the estimators sequentially unbiased if for all , and all , . Then the main result is:

  • If the different used by the agent are sequentially unbiased when labelled sequentially, then classical corrigibility is still sufficient.

To prove this, imagine that the agent will see a single change of utility , and that it has the opportunity of changing (to , say) exactly once before the utility change. Then there will be no term for , and, if it does change , the term is:

  • .

However, . Because of sequential unbiasedness, this simplifies to , which is just . So .

We can then recurse to the change in just before , and get the same result (since the future is still zero). And so on, with always being zero. Then since the formulas defining are linear, we can extend this to general environments and general utility function changes, and conclude that for sequentially unbiased , the are always under double indifference (modulo some convergence issues not addressed here). Therefore, double indifference will work, even if we don’t use ’s at all: thus classical indifference still works in this case.

Note the similarity of the sequential unbiasness with the conditions for successful value learners in the Cake or Death problem.