I was excited by the first half, seeing you relate classic Agent Foundations thinking to current NN training regimes, and try to relate the optimist/pessimist viewpoints.
Then we hit free energy and entropy. These seem like needlessly complex metaphors, providing no strong insight on the strength of the factors pushing toward and pulling away from alignment.
Analyzing those “forces” or tendencies seems like it’s crucially important, but needs to go deeper than a metaphor or use a much more fitting metaphor to get traction.
Nonetheless, upvoted for working on the important stuff even when it’s hard!
I probably shouldnt have used the free energy terminology.
Does complexity accuracy tradeoff work better ?
To be clear, I very much dont mean these things as a metaphor. I am thinking there may be an actual numerical complexity—accuracy that is some elaboration of Watanabe s “free energy” formula that actually describes these tendencies.
I was excited by the first half, seeing you relate classic Agent Foundations thinking to current NN training regimes, and try to relate the optimist/pessimist viewpoints.
Then we hit free energy and entropy. These seem like needlessly complex metaphors, providing no strong insight on the strength of the factors pushing toward and pulling away from alignment.
Analyzing those “forces” or tendencies seems like it’s crucially important, but needs to go deeper than a metaphor or use a much more fitting metaphor to get traction.
Nonetheless, upvoted for working on the important stuff even when it’s hard!
I probably shouldnt have used the free energy terminology. Does complexity accuracy tradeoff work better ?
To be clear, I very much dont mean these things as a metaphor. I am thinking there may be an actual numerical complexity—accuracy that is some elaboration of Watanabe s “free energy” formula that actually describes these tendencies.