We humans compartmentalize by default, because brains don’t automatically enforce belief propagation.
Belief propagation is an exact computation that brains can’t be expected to perform (or even represent a problem statement for). Pointing to (absence of) it as an explanation for compartmentalization feels rather arbitrary (similarly with the reference to decision theory).
What I mean is that if brains enforced belief propagation (and thus, were configured to do so), there wouldn’t be compartmentalization. I guess I can clarify that by adding a period and a few words.
But in fact, you don’t have a tail because you’re not a kangaroo. And if we were all fairly familiar with kangaroos and thought they were fairly analogous to Vladimir_Nesovs, then we would make note of the distinction.
I don’t see how it helps. I think the idea is wrong, not the wording. This situation also seems somewhat analogous to that with your use of Aumann agreement term: drawing a loose analogy with a technical tool that isn’t really relevant.
(To alleviate the usual worry, I note that I upvoted the post itself, and this trivial isolated point has no bearing on overall impression.)
Belief propagation is an exact computation that brains can’t be expected to perform (or even represent a problem statement for). Pointing to (absence of) it as an explanation for compartmentalization feels rather arbitrary (similarly with the reference to decision theory).
What I mean is that if brains enforced belief propagation (and thus, were configured to do so), there wouldn’t be compartmentalization. I guess I can clarify that by adding a period and a few words.
This doesn’t mark it as a natural explanation. By the same pattern, I don’t have a tail because I’m not a kangaroo.
But in fact, you don’t have a tail because you’re not a kangaroo. And if we were all fairly familiar with kangaroos and thought they were fairly analogous to Vladimir_Nesovs, then we would make note of the distinction.
Is the new wording still confusing?
I don’t see how it helps. I think the idea is wrong, not the wording. This situation also seems somewhat analogous to that with your use of Aumann agreement term: drawing a loose analogy with a technical tool that isn’t really relevant.
(To alleviate the usual worry, I note that I upvoted the post itself, and this trivial isolated point has no bearing on overall impression.)