While we try to ensure that it will pursue the goals that we gave it, we are NOT sure that it won’t develop its own goals
“Not sure that not X” is very different to “sure that X”.
I’m not arguing for 0% p(doom) , I’m arguing against 99%.
The point about commercially viable AI being aligned well enough
It’s a tautology. It’s aligned well enough to be usable, because it’s usable. If it were unaligned in a binary sense, it wouldn’t follow instructions.
as dumb as claiming that the AIs cannot fake alignment, gather power and take over the world.
Possible is very far from certain.
“Not sure that not X” is very different to “sure that X”.
I’m not arguing for 0% p(doom) , I’m arguing against 99%.
It’s a tautology. It’s aligned well enough to be usable, because it’s usable. If it were unaligned in a binary sense, it wouldn’t follow instructions.
Possible is very far from certain.