It’s an essential aspect of decision making for an agent to figure out where it might be. Thought experiments try to declare the current situation, but they don’t necessarily need to be able to convincingly succeed. Algorithmic induction, such as updating from Solomonoff prior, is the basic way an agent figures out which situations it should care about, and declaring that we are working with a particular thought experiment doesn’t affect the prior. In line with updatelessness, an agent should be ready for observations in general (according to which of them it cares about more), rather than particular “fair” observations, so distinguishing observations that describe “fair” thought experiments doesn’t seem right either.
It’s an essential aspect of decision making for an agent to figure out where it might be. Thought experiments try to declare the current situation, but they don’t necessarily need to be able to convincingly succeed. Algorithmic induction, such as updating from Solomonoff prior, is the basic way an agent figures out which situations it should care about, and declaring that we are working with a particular thought experiment doesn’t affect the prior. In line with updatelessness, an agent should be ready for observations in general (according to which of them it cares about more), rather than particular “fair” observations, so distinguishing observations that describe “fair” thought experiments doesn’t seem right either.