Agreed. Humans are constantly optimizing a reward function, but it sort of ‘changes’ from moment to moment in a near-focal way, so it often looks irrational or self-defeating, but once you know what the reward function is, the goal-directedness is easy to see too.
Doesn’t this become tautological? If the reward function changes from moment to moment, then the reward function can just be whatever explains the behaviour.
Since everything can fit into the “agent with utility function” model given a sufficiently crumpled utility function, I guess I’d define “is an agent” as “goal-directed planning is useful for explaining a large enough part of its behavior.” This includes humans while discluding bacteria. (Hmm unless, like me, one knows so little about bacteria that it’s better to just model them as weak agents. Puzzling.)
Doesn’t this become tautological? If the reward function changes from moment to moment, then the reward function can just be whatever explains the behaviour.
Since everything can fit into the “agent with utility function” model given a sufficiently crumpled utility function, I guess I’d define “is an agent” as “goal-directed planning is useful for explaining a large enough part of its behavior.” This includes humans while discluding bacteria. (Hmm unless, like me, one knows so little about bacteria that it’s better to just model them as weak agents. Puzzling.)