In both cases, I understand them as saying that the loss function used for training is an entirely different sort of thing from the goals an intelligent system pursues after training.
I think Turntrout would object to that charecterization as it is privileging the hypothesis that you get systems which pursue goals after training. I’m assuming you mean the agent does some sort of EV maximization by “goals an intelligent systems pursues”. Though I have a faint suspicion Turntrout would disagree even with a more general interpretation of “pursues goals”.
I think Turntrout would object to that charecterization as it is privileging the hypothesis that you get systems which pursue goals after training. I’m assuming you mean the agent does some sort of EV maximization by “goals an intelligent systems pursues”. Though I have a faint suspicion Turntrout would disagree even with a more general interpretation of “pursues goals”.