I haven’t fully thought it out, but there might be some counterargument in the style of anti-Pascals-mugging counterargument, where if your priors say that you might be modeled by a hostile entity, there is an incentive to confuse it, and it’s all going to balance out (somehow) and you just need to use your decision theory as if you are real always.
I haven’t fully thought it out, but there might be some counterargument in the style of anti-Pascals-mugging counterargument, where if your priors say that you might be modeled by a hostile entity, there is an incentive to confuse it, and it’s all going to balance out (somehow) and you just need to use your decision theory as if you are real always.