Reasoning from fictional evidence, I see.The point wasn’t that this failure mode was likely, it was that approximately every objection we’ve seen as to why AI won’t become unsafe fails.
I wouldn’t assume this particular failure mode is how things will go down in real life, just a potential counter-measure assuming the premises of the fiction
Reasoning from fictional evidence, I see.
The point wasn’t that this failure mode was likely, it was that approximately every objection we’ve seen as to why AI won’t become unsafe fails.
I wouldn’t assume this particular failure mode is how things will go down in real life, just a potential counter-measure assuming the premises of the fiction