I guess you could rephrase it as “suppose a UR optimizer had a button which randomly caused an agent to be a UR optimizer” or something along those lines and have similair results.
Yeah, instead of asking it a question, we can just see what happens when we put it in a world where it can influence another robot going left or right. Set it up the right way, and Stuart’s arguement should go through.
I guess you could rephrase it as “suppose a UR optimizer had a button which randomly caused an agent to be a UR optimizer” or something along those lines and have similair results.
Do you mean that as a way to understand what Stuart is talking about when he says that a UR-optimiser would answer questions in a certain way?
Yeah, instead of asking it a question, we can just see what happens when we put it in a world where it can influence another robot going left or right. Set it up the right way, and Stuart’s arguement should go through.