Counterfactuals are an Answer, Not a Question

I’m going to subtly contradict my last post on logical counterfactuals. I now think that raw or actually-consistent counterfactuals are just an especially useful model of what counterfactuals are as opposed to the be all and end all.

An Old Classic

You may have heard this one before: The Ship of Theseus is a very old ship, so old that all the original parts had been replaced. Is it still the Ship of Theseus? What if we gathered all of the original parts and used them to rebuild the ship? Would this new construct be the Ship of Theseus now?

If you’ve internalised the idea of the map and the territory, the answer should be clear. Terms like “The Ship of Theseus” are human constructions with no real existence. They can be defined in whatever manner is most useful.

Counterfactuals Don’t Exist in the Universe

So let’s start with regular counterfactuals. What are they? Do they exist in the universe itself? Unless you’re a modal realist (ie. David Lewis and his clones) the answer is no. Given the exact state of universe and an agent, the agent can only make one decision*. In other words, counterfactuals are something we construct.

This means that it is a mistake to search for an objectively true definition. Just as in the Ship of Theseus, our expectation should be that multiple definitions could have value. Of course, some definitions may be more useful or natural than others and indeed, as I’ve argued, Raw Counterfactuals are particularly natural. However, I wouldn’t go as far as I went in that post where I argued that insofar any other kind of counterfactual had meaning, it was derived from Raw Counterfactuals.

Logical counterfactuals are no different. We may not know everything about logic, but this doesn’t mean that logic could have been different. We can construct a model where we imagine that logic being different than it is, but there isn’t a fact of the matter about how logic would be if 1+1=3 instead of 2 that exists as part of standard logic without any extensions, any more than counterfactuals exist in-universe without any extensions.

An Answer, Not a Question

Since counterfactuals don’t have an objectively true definition, asking, “What are counterfactuals?” is a confused question. It’d be better to ask, “What definitions of counterfactuals are useful?”, but then this is still kind of vague. A better answer is to figure out the kinds of questions that we tend to want to answer when counterfactuals arise. This will vary, but in general counterfactuals relate to problems that can be solved via iteration.

An Initial Question

Raw counterfactuals are especially important because they answer a particularly common kind of question related to partial knowledge. One common simplification is to assume that we can enumerate all the states of the universe that would match our given knowledge. Since states are only valid if they are consistent, the states we enumerate will be raw counterfactuals. Indeed, as I previously argued, most uses of CDT-style counterfactuals are justified in terms of how they approximate raw counterfactuals. Pretending an agent’s magically decides to turn left instead of right at the last moment, isn’t that different from assuming that the universe was slightly different so that agent was always going to turn left, but that this change wasn’t ever going to affect anything else.

When the Story Breaks

So why doesn’t the story end here? Why might we desire other notions of counterfactual? I won’t claim that all of the following situations would lead to an alternative notion of counterfactual, but I wouldn’t want to rule it out.

Firstly, the assumption in the previous question is that we already have abstracted out a world model; that is, for any observation, we have an idea of what worlds are possible and other relevant facts like probabilities. While this describes ideal Bayesian agents, it doesn’t really describe more realistic agents who choose their model based at least somewhat upon their observations.

Secondly, we might sometimes want to consider the possibility that our logical assumptions or deductions might be incorrect. This is possible, but tricky, since as if our logic is incorrect, we’ll be using our flawed logic to determine how to handle this issues.

Thirdly, it’s not always immediately obvious whether a particular state is consistent or not. For example, if you know that you’re a utility maximiser, then it would be inconsistent for you to pick any option that provides suboptimal utility, but if you knew the utility was suboptimal you wouldn’t need to examine that option. It would often be useful to be able to assign a utility value to inconsistent situations so we could use the common technique of iterating over a list of possibly-inconsistent situations and pick whichever was assigned the highest utility.

Ambient Decision Theory (much of which has been wrapped into FDT) takes advantage of this. You consider various counterfactuals (X chooses action1, action2, ect.), only one of which should be consistent given the fact that the program is a utility maximiser. The hope is that the only consistent result will represent the best move and it will in the right circumstance.

Inconsistency

This last strategy is neat, but it faces significant challenges. For one, what does it mean to say that when we choose the only consistent result we choose the best outcome? “Best” requires a comparison, which requires a value. In other words, we have assumes a way of assigning utility to inconsistent situations. What does this mean?

From a logically viewpoint, if we ever have P and Not P, then we have a contradiction and can prove anything. So we shouldn’t strictly view these inconsistent situations as logical entities, at least in the classical sense. Instead, it may be more useful to view them as a datastructure and the utility function as something that extracts a particular element or elements, then runs a regular utility function over these.

For example, suppose we have:

{mouse=alive, cat=alive, cat=dead}

Viewed as logical propositions, we can prove that the mouse is also dead from the contradiction, but viewed as a datastructure, the mouse is only alive. This doesn’t perfectly map to the situation where you’re given an inconsistent list of logical propositions and told to look for contradictions, but it is close enough to be analogous as in either case, the inconsistent object is more a pseudo-logical object than a logical object. Similarly, we could imagine assigned a utility based upon the consistent part of such a datastructure.

So given these datastructures, we can define a notion of “best” for inconsistent situations, but this begs the question, “Why do we care about the best such (inconsistent) datastructure?” Do we only care about these inconsistent objects insofar as they stand in for an actually consistent situation? I suspect the answer could be no and I don’t currently have a very good answer for this, but I’m hoping to have one soon. In any case, my previous stance completely dismissing these possibilities without solid reasons was too absolutionist for me to still maintain it.

*Quantum mechanics allows this to be a probability distribution, but then it’s just probabilistically deterministic instead, so it only complicates the issue without really changing anything