I’m not sure of my reading of this was correct but I will describe it below in the hopes of someone engaging with me and telling me if I’m wrong or not. (Low confidence in the following claim:)
This view seems like what you would have if you thought that P(y|x) would be here:
y is what the brain is doing as an agent and x is the prior which in this case is that the human body is a homeostatic control system trying to minimise the time it spends out of balance then it seems you might converge to a view like this?
There’s also a couple of ther assumptions around:
Fixed points within our evolutionary past not mattering as much as the functional of homeostatic control.
Emotions not being related to exploration somehow? (In that there’s no innate positive drives?)
That the brain is a relatively straightforwardly coupled control system?
It seems a bit like taking dynamic programming and saying that it is what Reinforcement Learning is? You’re missing part of the algorithm (the action part of the action-perception loop?).
(I also pre-emptively apologise for invoking Active Inference in a Steven Byrnes comment field)
I’m not sure of my reading of this was correct but I will describe it below in the hopes of someone engaging with me and telling me if I’m wrong or not. (Low confidence in the following claim:)
This view seems like what you would have if you thought that P(y|x) would be here:
y is what the brain is doing as an agent and x is the prior which in this case is that the human body is a homeostatic control system trying to minimise the time it spends out of balance then it seems you might converge to a view like this?
There’s also a couple of ther assumptions around:
Fixed points within our evolutionary past not mattering as much as the functional of homeostatic control.
Emotions not being related to exploration somehow? (In that there’s no innate positive drives?)
That the brain is a relatively straightforwardly coupled control system?
It seems a bit like taking dynamic programming and saying that it is what Reinforcement Learning is? You’re missing part of the algorithm (the action part of the action-perception loop?).
(I also pre-emptively apologise for invoking Active Inference in a Steven Byrnes comment field)