diffusion planning is entirely made from motivated reasoning and performs pretty well. this is imo a reasonable exemplar for a slightly broader belief I have, which is that the simplest hypothesis for motivated reasoning is that reasoning from percept-feature to outcome-feature (prediction), outcome feature to motor feature (control), outcome feature to percept feature (wishful thinking), are not trivially distinguishable when you have something confuseable for “big ol’ diffusion model over whatever”, and so avoiding motivated reasoning is hard in a messy system. There’s enough pressure to avoid a lot of it, but given that reasoning from outcome-feature to motor-feature is a common need, going through correlational features that mix representations is totally allowed-by-substrate and thus common anywhere there isn’t sufficient pressure against it.
of course, I’m being kind of sloppy in my claims here.
(This is a less-specific version of saying “it’s just active inference”, because the active inference math hasn’t clicked for me yet, so I can’t claim that it’s exactly active inference; but it does seem like in general, planning-by-inference ought to be the default, as hinted by the fact that you can get it just by jiggling stuff around diffusion style.)
diffusion planning is entirely made from motivated reasoning and performs pretty well. this is imo a reasonable exemplar for a slightly broader belief I have, which is that the simplest hypothesis for motivated reasoning is that reasoning from percept-feature to outcome-feature (prediction), outcome feature to motor feature (control), outcome feature to percept feature (wishful thinking), are not trivially distinguishable when you have something confuseable for “big ol’ diffusion model over whatever”, and so avoiding motivated reasoning is hard in a messy system. There’s enough pressure to avoid a lot of it, but given that reasoning from outcome-feature to motor-feature is a common need, going through correlational features that mix representations is totally allowed-by-substrate and thus common anywhere there isn’t sufficient pressure against it.
of course, I’m being kind of sloppy in my claims here.
(This is a less-specific version of saying “it’s just active inference”, because the active inference math hasn’t clicked for me yet, so I can’t claim that it’s exactly active inference; but it does seem like in general, planning-by-inference ought to be the default, as hinted by the fact that you can get it just by jiggling stuff around diffusion style.)