Like Self-fulfilling misalignment data might be poisoning our AI models, what are historical examples of self-fulfilling prophecies that have affected AI alignment and development?
Put a few potential examples below to seed discussion.
Like Self-fulfilling misalignment data might be poisoning our AI models, what are historical examples of self-fulfilling prophecies that have affected AI alignment and development?
Put a few potential examples below to seed discussion.
https://x.com/sama/status/1621621724507938816
Training on Documents About Reward Hacking Induces Reward Hacking
Situational Awareness and race dynamics? h/t Jan Kulveit @Jan_Kulveit
Situational Awareness probably caused Project Stargate to some extent. Getting the Republican party to take AI seriously enough to let them launch in the White House is no joke and less likely without the essay.
It also started the website-essay meta which is part of why AI 2027, The Compendium, and Gradual Disempowerment all launched the way they did, so there are knock-on effects too.
Superintelligence Strategy is pretty explicitly trying to be self-fulfilling, e.g. “This dynamic stabilizes the strategic landscape without lengthy treaty negotiations—all that is necessary is that states collectively recognize their strategic situation” (which this paper popularly argues exists in the first place)
https://x.com/saffronhuang/status/1907863453009867183
Grok prompts lately, kinda “Don’t think about elephants”
https://x.com/bitcloud/status/1942792945238983022