Here’s one that I don’t think has a name—the belief that [some desirable thing] should just happen. For example, the belief that people should just have different emotional reactions than they do, or that a government policy should just have good effects.
Incidentally, this expression is very intuitive and has an amazingly low inferential distance. Multiple times IRL and online I would reply to someone “it’s too bad we don’t live in a should-universe” in response to a should-statement, and my reply is instantly understood, without having to explain much, except maybe saying “the should-universe is an imaginary one where what you think should happen actually does, every time”.
Related notion: in the humongous SSC post Reactionary Philosophy in an Enormous Planet-Sized Nutshell, the section “Reach for the Tsars” deals with proposals to solve problems which could only be implemented by dictatorial fiat, and describes it as a “czar’s-eye view” solution.
It may be a typical mind fallacy if the person actually has the emotional habits they’re demanding from other people. Now that I think about it, people sometimes demand that their own emotions should just be different.
However, a statement can include more than one fallacy, and I think fantasies of lack of process can also be in play.
The whole FAI project resulted from Eiiezer realizing that a process was needed for AIs to be benevolent rather than a disaster.
As may be obvious, I now think the bias could be named the lack of process bias, though the “it should just happen!” bias might be more intuitive.
I was going to ask, “Do we ever demand emotional habits we don’t have ourselves?”, but then I noticed it was yet another typical mind fallacy on my part.
Here’s one that I don’t think has a name—the belief that [some desirable thing] should just happen. For example, the belief that people should just have different emotional reactions than they do, or that a government policy should just have good effects.
Eliezer called this believing in the should-universe.
Incidentally, this expression is very intuitive and has an amazingly low inferential distance. Multiple times IRL and online I would reply to someone “it’s too bad we don’t live in a should-universe” in response to a should-statement, and my reply is instantly understood, without having to explain much, except maybe saying “the should-universe is an imaginary one where what you think should happen actually does, every time”.
https://en.wikipedia.org/wiki/Wishful_thinking or perhaps https://en.wikipedia.org/wiki/Is%E2%80%93ought_problem ?
Related notion: in the humongous SSC post Reactionary Philosophy in an Enormous Planet-Sized Nutshell, the section “Reach for the Tsars” deals with proposals to solve problems which could only be implemented by dictatorial fiat, and describes it as a “czar’s-eye view” solution.
Expecting different emotions than the ones actually observed looks to me like typical mind fallacy.
It may be a typical mind fallacy if the person actually has the emotional habits they’re demanding from other people. Now that I think about it, people sometimes demand that their own emotions should just be different.
However, a statement can include more than one fallacy, and I think fantasies of lack of process can also be in play.
The whole FAI project resulted from Eiiezer realizing that a process was needed for AIs to be benevolent rather than a disaster.
As may be obvious, I now think the bias could be named the lack of process bias, though the “it should just happen!” bias might be more intuitive.
I was going to ask, “Do we ever demand emotional habits we don’t have ourselves?”, but then I noticed it was yet another typical mind fallacy on my part.