For my part, if I think about things people might do that might cause a powerful AI to feel threatened and thereby have significantly bad results, FAI theory and implementation not only doesn’t float to the top of the list, it’s hardly even visible in the hypothesis space (unless, as here, I privilege it inordinately by artificially priming it).
OK.
For my part, if I think about things people might do that might cause a powerful AI to feel threatened and thereby have significantly bad results, FAI theory and implementation not only doesn’t float to the top of the list, it’s hardly even visible in the hypothesis space (unless, as here, I privilege it inordinately by artificially priming it).