Then you’re just dealing with beliefs about preferences, which are a kind of beliefs, so this reduces to PCR for beliefs.
You’re right there. And PCR for beliefs is trivial in principle, just use Bayes’ rule and the Universal Prior based on the programming language of your choice. Nobody seems to be good enough at actually evaluating that prior to care much about which programming language you use to represent the hypotheses yet.
So if someone introspects and says they will make choices as though they have unbounded utility, and the math makes it seem impossible for them to really do that, then I can reply “I don’t believe you” and move on, just as though they had professed believing in an invisible dragon in their garage.
That’s a really simple solution to get rid of a large pile of garbage, contingent on the math working out right. Thanks. I’ll pay more attention to the math.
ETA: I edited the OP to point to this comment. This was an excellent outcome from the conversation, by the way. LessWrong works.
You’re right there. And PCR for beliefs is trivial in principle, just use Bayes’ rule and the Universal Prior based on the programming language of your choice. Nobody seems to be good enough at actually evaluating that prior to care much about which programming language you use to represent the hypotheses yet.
So if someone introspects and says they will make choices as though they have unbounded utility, and the math makes it seem impossible for them to really do that, then I can reply “I don’t believe you” and move on, just as though they had professed believing in an invisible dragon in their garage.
That’s a really simple solution to get rid of a large pile of garbage, contingent on the math working out right. Thanks. I’ll pay more attention to the math.
ETA: I edited the OP to point to this comment. This was an excellent outcome from the conversation, by the way. LessWrong works.
(obligatory xkcd reference)
LessWrong: It works, bitches.