Hi thanks for share this interesting perspective on RL as a training process! Although it seems to only be a matter of seeking vs obeying and reward vs cost, the effect on the reader’s mind seem to be huge!
One thing that seems to be happening here and I have not fully digested is the “intrinsicness” of rewards. In frameworks parallel to mainstream RL, such as active inference and the free energy principle, policy is a part of the agent’s model such that the agent “self-organizes” to a characteristic state of the world. The policy can be constructed either through reward or not. However, in the active inference literature, how policies are constructed in real agents are currently unanswered (discussions exist but don’t close the case).
How this intrinsic perspective is related to the post and safety and alignment? I am still thinking about it. If you have any thoughts please share!
Hi thanks for share this interesting perspective on RL as a training process! Although it seems to only be a matter of seeking vs obeying and reward vs cost, the effect on the reader’s mind seem to be huge!
One thing that seems to be happening here and I have not fully digested is the “intrinsicness” of rewards. In frameworks parallel to mainstream RL, such as active inference and the free energy principle, policy is a part of the agent’s model such that the agent “self-organizes” to a characteristic state of the world. The policy can be constructed either through reward or not. However, in the active inference literature, how policies are constructed in real agents are currently unanswered (discussions exist but don’t close the case).
How this intrinsic perspective is related to the post and safety and alignment? I am still thinking about it. If you have any thoughts please share!