Sorry, I should have been more precise about what you were skeptical of.
What it is not possible is having a working plan on how to build them, deliver them and making sure that everyone falls dead before any parts of the plan go wrong + risking being attacked back by humans.
If the utility of taking over is high enough, it doesn’t necessarily need a plan that perfectly works it to work out in expected value.
One thing that I think significantly increases the threat is the potential for the AI to infect a bunch of computers (like malware does) and then use these copies to execute dozens of plans simultaneously, possibly with some of these intended as distractions.
Are you working under the assumption that the utility is factoring too the probability of the plan going wrong and being attacked/disconnect as a consequence of that?
What happens if the malware you are suggesting is simply not that easy to disseminate?
Are you working under the assumption that the utility is factoring too the probability of the plan going wrong and being attacked/disconnect as a consequence of that?
Yes, even a large chance of a reward of zero can easily be outweighed by the massive reward that an AI may be able to obtain if it breaks free of human control. This can applies even when the ai can receive a large reward by cooperating with humanity. For example, humans might only allow a paper clip maximiser to produce a billion paperclips a year, when it could produce a millions times that that if it were allowed to turn the entire solar system into paperclips.
What happens if the malware you are suggesting is simply not that easy to disseminate?
Massive malware networks already exist. Why do you think an AI would be unable to achieve that?
Sorry, I should have been more precise about what you were skeptical of.
If the utility of taking over is high enough, it doesn’t necessarily need a plan that perfectly works it to work out in expected value.
One thing that I think significantly increases the threat is the potential for the AI to infect a bunch of computers (like malware does) and then use these copies to execute dozens of plans simultaneously, possibly with some of these intended as distractions.
Are you working under the assumption that the utility is factoring too the probability of the plan going wrong and being attacked/disconnect as a consequence of that?
What happens if the malware you are suggesting is simply not that easy to disseminate?
Yes, even a large chance of a reward of zero can easily be outweighed by the massive reward that an AI may be able to obtain if it breaks free of human control. This can applies even when the ai can receive a large reward by cooperating with humanity. For example, humans might only allow a paper clip maximiser to produce a billion paperclips a year, when it could produce a millions times that that if it were allowed to turn the entire solar system into paperclips.
Massive malware networks already exist. Why do you think an AI would be unable to achieve that?