What’s your version of the story for how the “moderates” at OpenPhil ended up believing stuff even others can now see to be fucking nuts in retrospect and which “extremists” called out at the time, like “bio anchoring” in 2021 putting AGI in median fucking 2050, or Carlsmith’s Multiple Stage Fallacy risk estimate of 5% that involved only an 80% chance anyone would even try to build agentic AI?
Were they no true moderates? How could anyone tell the difference in advance?
From my perspective, the story is that “moderates” are selected to believe nice-sounding moderate things, and Reality is off doing something else because it doesn’t care about fitting in the same way. People who try to think like reality are then termed “extremist”, because they don’t fit into the nice consensus of people hanging out together and being agreeable about nonsense. Others may of course end up extremists for other reasons. It’s not that everyone extreme is reality-driven, but that everyone who is getting pushed around by reality (instead of pleasant hanging-out forces like “AGI in 2050, 5% risk” as sounded very moderate to moderates before the ChatGPT Moment) ends up departing from the socially driven forces of what entitles you to sound terribly reasonable to the old AIco-OpenPhil cluster and hang out at their social gatherings without anyone feeling uncomfortable.
Anyone who loves being an extremist will of course go instantly haywire a la Yampolskiy imagining that he has proven alignment impossible via Godelian fallacy so he can say 99.9999% doom. But yielding to the psychological comfort of being a “moderate” will not get you any further in science than that.
I accept your correction and Buck’s as to these simple facts (was posting from mobile).