[Question] Is acausal extortion possible?

Disclaimer: this question discusses potential infohazards, which may cause harm or psychological distress if true. If you decide to read this post, please be warned and do so at your own discretion.


Hi, I am someone new to LessWrong so please excuse any mistakes I may make in my discussions (you’re more than welcome to point them out). This question is intended to ask whether or not acausal extortion is possible and to pose a scenario which I am unsure about, hence please feel free to critique my points and arguments.

Acausal trade is essentially when two agents simulate each other to predict each other’s actions, and acausal extortion (or blackmail) is when one agent applies a negative incentive to influence the behavior of another agent when the latter agent models the former agent. From what I’ve gathered, acausal trade (and by extension acausal extortion) between humans and artificial superintelligences (ASIs) is impossible as it requires lots of computing power on the human’s end in order for the ASI to not have an incentive to defect instead cooperate. However, this still leaves open the possibility of a non-superintelligence (such as an alien) who has control over an ASI using it to acausally extort us. To see what I mean, consider the scenario below:

On some branch of the multiverse, there exists an alien with control over an aligned ASI. The alien then asks the ASI to create many simulations of the other branches of the multiverse and you are contained in one of these simulations. The alien observes you and realizes that you realize the possibility of such a scenario taking place (since you are reading this paragraph). The alien then thinks: If you do not take action X, it will ask the ASI to create an arbitrarily large number of simulations of you and torture them for an arbitrarily long period of time. And since these simulations have the exact same subjective experience as you, you are compelled to carry out action X as you are uncertain about whether or not you’re in a simulation (indexical uncertainty). And since it is much easier to model non-superintelligences than ASIs, this scenario avoids the pitfall of scenarios like Roko’s Basilisk where humans are limited by our computing power.

The scenario above is partly inspired by the famous Roko’s Basilisk and Stuart Armstrong’s AI in a Box thought experiment.


Now I’d like to address some of the common refutations to acausal extortion scenarios and why I don’t think any of them work very well against this particular scenario. (Though it’s very likely I have made mistakes or engaged strawman versions of these arguments, and if so, please do not hesitate to point them out).


Yeah, sure, this scenario could happen, but it’s very very very unlikely right?

The above scenario, however unlikely, is guaranteed to take place with probability 1 if the Many-Worlds Interpretation of quantum mechanics is correct, as every possible outcome occurs in such a multiverse. And even if the measure of this particular branch of the multiverse is extremely low, if the alien creates an extremely large number of simulations to create indexical uncertainty, it can still have a significant effect on your most probable environment and possible future observer-moments.

Ok, sure, but I can imagine any action X and there will be a quantum branch where the alien wants you to take that particular action, this is just a Pascal’s Mugging where the many-gods refutation still applies.

The action X can be indeed be any arbitrary action, but such aliens are likely to have convergent instrumental goals such that the probability distribution of X will not be random. And I don’t find the many-gods refutation to be satisfactory since, well, it just concludes that you will be tortured no matter what you do (not a good ending).

Hmm, well then if I commit to ignore any such attempts of blackmail, then the blackmailer will have no incentive to employ such a tactic.

Sure, it works in principle, but humans cannot commit to an action with 100% certainty. How confident are you that none of your bazillions of copies will give in? And also, in the MWI every possible outcome occurs, including outcomes where you do give in to the blackmailer’s demand, so it’s likely that doing so will yield positive expected utility (however small) for the blackmailer.


In the above section I addressed some of the refutations that I hear commonly but do not find all that convincing. I am not sure how much I believe in this whole thing yet but since I am not an expert in decision theory and this sort of thing in general I wanted to hear some criticism from more knowledgeable people on the scenario I’ve posed as well as on how to avoid feeling anxious over acausal extortion. Thanks in advance.