The inference goes beyond the definition and claims, additionally, that in a case like this diagram, the agent will only take actions that “follow the arrows,” changing the world state from less-preferred to more-preferred at each step.
Hmm. I’m not sure about that inference but might be due to absence of specific argument you’re making that inference from.
The preference arrow to me is simply pointing in a direction. The path between the less preferred state to more preferred state is pretty much a black box at this point. So I think the inference is more like: If agent 86 stays in state A, there is a lack of resources or a higher transition cost to get from A to C than a difference in utility/value between the two states.
There is probably something of a granularity here as well. Does taking a second job for a while, or doing some type of gig-work one doesn’t enjoy at all, count as a state or is that wrapped up in some type of non-preference type of state?
Hmm. I’m not sure about that inference but might be due to absence of specific argument you’re making that inference from.
The preference arrow to me is simply pointing in a direction. The path between the less preferred state to more preferred state is pretty much a black box at this point. So I think the inference is more like: If agent 86 stays in state A, there is a lack of resources or a higher transition cost to get from A to C than a difference in utility/value between the two states.
There is probably something of a granularity here as well. Does taking a second job for a while, or doing some type of gig-work one doesn’t enjoy at all, count as a state or is that wrapped up in some type of non-preference type of state?