Probably the most sensible response to the interview I’ve seen so far.
Also I’ll probably start referencing this when people argue whether thinking in active inference frame has any advantage over thinking in the RL frame. Clearly it does: it’s way easier to see what happens if you drop the “reward” term Sutton is imagining as necessary and keep just the prediction error minimization terms. You still get intelligent systems, they still learn powerful abstractions (because they need to compress data), they still learn a generative world model. (It’s probably good the active inference frame is antimemetic in the orthodox RL crowd)
Probably the most sensible response to the interview I’ve seen so far.
Also I’ll probably start referencing this when people argue whether thinking in active inference frame has any advantage over thinking in the RL frame. Clearly it does: it’s way easier to see what happens if you drop the “reward” term Sutton is imagining as necessary and keep just the prediction error minimization terms. You still get intelligent systems, they still learn powerful abstractions (because they need to compress data), they still learn a generative world model. (It’s probably good the active inference frame is antimemetic in the orthodox RL crowd)