No, I would say that if you start entering false utility data into the AI and it believes you, because after all it was programmed to be indifferent to new utility data, that’s your problem.
If the AI’s utility function changes randomly for no apparent reason because the AI has litterally zero incentive to make sure that doesn’t happen, then you have an AI design problem.
It didn’t change for no reason. It changed because someone fed new data into the AI’s utility-learning algorithm which made it change. Don’t give people root access if you don’t want them using it!
Being changed by an attacker is only one of the scenarios I was suggesting. And even then, presumably you would want the AI to help prevent them from hacking its utility function if they aren’t supposed to have root access, but it won’t.
Anyway, that problem is just a little bit stupid. But you can also get really stupid problems, like the AI wants more memory, so it replaces its utility function with something more compressible so that it can scavange from the memory where its utility function was stored.
You can’t always solve human problems with AI design.
I’m not sure what you mean. The problem I was complaining about is an AI design problem, not a human problem.
No, I would say that if you start entering false utility data into the AI and it believes you, because after all it was programmed to be indifferent to new utility data, that’s your problem.
If the AI’s utility function changes randomly for no apparent reason because the AI has litterally zero incentive to make sure that doesn’t happen, then you have an AI design problem.
It didn’t change for no reason. It changed because someone fed new data into the AI’s utility-learning algorithm which made it change. Don’t give people root access if you don’t want them using it!
Being changed by an attacker is only one of the scenarios I was suggesting. And even then, presumably you would want the AI to help prevent them from hacking its utility function if they aren’t supposed to have root access, but it won’t.
Anyway, that problem is just a little bit stupid. But you can also get really stupid problems, like the AI wants more memory, so it replaces its utility function with something more compressible so that it can scavange from the memory where its utility function was stored.