They also made an error in implicitly arguing that because they didn’t think unaligned behavior seems intelligent, then we have nothing to worry about from such AI—they wouldn’t be “intelligent”. I think leaving this out was a good choice.
They also made an error in implicitly arguing that because they didn’t think unaligned behavior seems intelligent, then we have nothing to worry about from such AI—they wouldn’t be “intelligent”. I think leaving this out was a good choice.