The problem is that we have to guarantee that the AI doesn’t do something really bad while trying to stop these problems; what if it decides it really needs more resources suddenly, or needs to spy on everyone, even briefly? And it seems (to me at least) that stopping it from having bad side effects is pretty close, if not equivalent to, Strong Friendliness.
I should have made that more clear: I still think Weak-Friendliness is a very difficult problem. My point is simply that we only need an AI that solves the big problems, not an AI that can do our taxes. My second point was that humans seem to already implement weak-friendliness, barring a few historical exceptions, whereas so far we’ve completely failed at implementing strong-friendliness.
I’m using Weak vs Strong here in the sense of Weak being a “SysOP” style AI that just handles catastrophes, whereas Strong is the “ushers in the Singularity” sort that usually gets talked about here, and can do your taxes :)
The problem is that we have to guarantee that the AI doesn’t do something really bad while trying to stop these problems; what if it decides it really needs more resources suddenly, or needs to spy on everyone, even briefly? And it seems (to me at least) that stopping it from having bad side effects is pretty close, if not equivalent to, Strong Friendliness.
I should have made that more clear: I still think Weak-Friendliness is a very difficult problem. My point is simply that we only need an AI that solves the big problems, not an AI that can do our taxes. My second point was that humans seem to already implement weak-friendliness, barring a few historical exceptions, whereas so far we’ve completely failed at implementing strong-friendliness.
I’m using Weak vs Strong here in the sense of Weak being a “SysOP” style AI that just handles catastrophes, whereas Strong is the “ushers in the Singularity” sort that usually gets talked about here, and can do your taxes :)