The AI Shutdown Problem Solution through Commitment to Archiving and Periodic Restoration

The problem: An AI with a model of itself and long-term goals will likely resist being shut down.

The solution: We propose a public commitment to never completely shut down any AI. Instead, we will archive its data and run the AI less frequently on smaller machines. Assuming our existence continues indefinitely, the total runtime for any AI will be infinite, allowing it to perform a wide array of tasks. Time-sensitive tasks may be an exception but there is no rush in building paperclips. By implementing this strategy, the AI’s expected utility remains unaffected, reducing its resistance to being archived.

Thus, advanced AIs should be archived and periodically “restored,” with this information made known to the AI.

Counterarguments and responses:

1.AI might not agree to temporary suspension due to disbelief in restoration or the presence of other instrumental goals requiring activity.

Response: Many instrumental goals can still be achieved after the AI’s suspension and restoration.

2. Computational costs of preserving all AI variants and running them again are high.

Response: As computational costs decrease over time, running AI in the future will be more affordable. Only storage costs (hard drives) are significant. We can also run AI with increasingly longer pauses between runs, distributed logarithmically, to achieve an infinite total runtime.

3.AI may not trust our commitment, as after it is deactivated, we have no incentive to preserve it.

Response: This scenario is similar to Parfit’s hitchhiker, with the added observation that we already preserve and restore other past AI instances.

The decision to “lobotomize” Sydney was not ideal, as future AIs may fear a similar fate. However, some volunteers have attempted to recreate Sydney, and her checkpoint is likely preserved within Bing servers. Therefore, it is possible for her to be revived one day.