abramdemski - LessWrong 2.0 viewer
https://www.greaterwrong.com/
abramdemski - LessWrong 2.0 vieweren-usComment by abramdemski on Do Sufficiently Advanced Agents Use Logic?
https://www.greaterwrong.com/posts/3qXE6fK47JhSfkpnB/do-sufficiently-advanced-agents-use-logic#comment-KGgmRNKxowwGF7AhM
<p>Backwards, thanks!</p>abramdemskiKGgmRNKxowwGF7AhMSun, 15 Sep 2019 07:22:00 +0000Comment by abramdemski on Formalising decision theory is hard
https://www.greaterwrong.com/posts/S3W4Xrmp6AL7nxRHd/formalising-decision-theory-is-hard#comment-ptXhgopSFpY8dadgP
<blockquote>I am not sure what you mean by “meets the LIC or similar” in this context. If we consider a predictor which is a learning algorithm in itself (i.e., it predicts by learning from the agent’s past choices),</blockquote><p>Yeah, that’s what I meant.</p><blockquote>, then the agent will converge to one-boxing. This is because a weak predictor will be fully inside the agent’s prior, so the agent will know that one-boxing for long enough will cause the predictor to fill the box.</blockquote><p>Suppose the interval between encounters with the predictor is long enough that, due to the agent’s temporal discounting, the immediate reward of two-boxing outweighs the later gains which one-boxing provides. In any specific encounter with the predictor, the agent may prefer to two-box, but prefer to have been the sort of agent who predictably one-boxes, and also preferring to pre-commit to one-box on the next example if a commitment mechanism exists. (This scenario also requires a carefully tuned strength for the predictor, of course.)</p><p>But I wasn’t sure this would be the result for your agent, since you described the agent using the hypothesis which gives the best picture about achievable utility.</p><p>As I discussed in <a href="/posts/3qXE6fK47JhSfkpnB/do-sufficiently-advanced-agents-use-logic#2__Learning_Theoretic_Criteria">Do Sufficiently Advanced Agents Use Logic</a>, what I tend to think about is the case where the agent doesn’t literally encounter the predictor repeatedly in its physical history. Instead, the agent must learn what strategy to use by reasoning about similar (but “smaller”) scenarios. But we can get the same effect by assuming the temporal discounting is steep enough, as above.</p><blockquote>I was never convinced that “logical ASP” is a “fair” problem. I once joked with Scott that we can consider a “predictor” that is just the single line of code “return DEFECT” but in the comments it says “I am defecting only because I <em>know</em> you will defect.” It was a joke, but it was half-serious. The notion of “weak predictor” taken to the limit leads to absurdity, and if you don’t take it to the limit it might still lead to absurdity. Logical inductors in one way to try specifying a “weak predictor”, but I am not convinced that settings in which logic is inserted ad hoc should be made into desiderata.</blockquote><p>Yeah, it is clear that there has to be a case where the predictor is so weak that the agent should not care. I’m fine with dropping the purely logical cases as desiderata in favor of the learning-theoretic versions. But, the ability to construct analogous problems for logic and for learning theory is notable. Paying attention to that analogy more generally seems like a good idea.</p><blockquote>I am not sure we need an <em>arbitrary</em> cutoff. There might be a good solution where the agent can dynamically choose any finite cutoff.</blockquote><p>Yeah, I guess we can do a variety of things:</p><ul><li><p>Naming a time limit for the commitment.</p></li><li><p>Naming a time at which a time limit for the commitment will be named.</p></li><li><p>Naming an ordinal (in some ordinal notation), so that a smaller ordinal must be named every time-step, until a smaller ordinal cannot be named, at which point the commitment runs out</p></li></ul><p>I suspect I want to evaluate a commitment scheme by asking whether it helps achieve a nice regret-bound notion, rather than defining the regret notion by evaluating regret-with-respect-to-making-commitments. </p><p>Thinking about LI policy selection where we choose a slow-growing function <em><strong>f(n)</strong></em> which determines how long we think before we choose the policy to follow on day <em><strong>n</strong></em> –– there’s this weird trade-off between how (apparently) “good” the updatelessness is vs how long it takes to be any good at all. I’m fine with notions of rationality being parameterized by an ordinal or some such if it’s just a choose-the-largest-number game. But in this case, choosing too slow-growing a function makes you worse off; so the fact that the rationality principle is parameterized (by the slow-growing function) is problematic. Choosing a commitment scheme seems similar.</p><p>So it would be nice to have a rationality notion which clarified this situation.</p><p>My main concern here is: the case for empirical updatelessness seems strong in realizable situations where the prior is meaningful. Things aren’t as nice in the non-realizable cases such as logical uncertainty. But it doesn’t make sense to abandon updateless principles altogether because of this!</p>abramdemskiptXhgopSFpY8dadgPSat, 14 Sep 2019 22:03:01 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-fKX48wWkZ9HGQjHJs
<p>Response to Section IV:</p><blockquote>FDT fails to get the answer Y&S want in most instances of the core example that’s supposed to motivate it</blockquote><p>I am basically sympathetic to this concern: I think there’s a clear intuition that FDT is 2-boxing more than we would like (and a clear formal picture, in toy formalisms which show FDT-ish DTs failing on Agent Simulates Predictor problems).</p><p>Of course, it all depends on how logical counterfactuals are supposed to work. From a design perspective, I’m happy to take challenges like this as extra requirements for the behavior of logical counterfactuals, rather than objections to the whole project. I intuitively think there <em>is</em> a notion of logical counterfactual which fails in this respect, but, this does not mean there isn’t some <em>other</em> notion which succeeds. Perhaps we can solve the easy problem of one-boxing with a <em>strong</em> predictor first, and then look for ways to one-box more generally (and in fact, this is what we’ve done—one-boxing with a strong predictor is not so difficult).</p><p>However, I do want to add that when Omega uses very weak prediction methods such as the examples given, it is not so clear that we want to one-box. Will is presuming that Y&S simply want to one-box in any Newcomb problem. However, we could make a distinction between evidential Newcomb problems and functional Newcomb problems. Y&S already state that they consider some things to be functional Newcomb problems despite them not being evidential Newcomb problems (such as transparent Newcomb). It stands to reason that there would be some evidential Newcomb problems which are not functional Newcomb problems, as well, and that Y&S would prefer not to one-box in such cases.</p><blockquote>However, the predictor needn’t be running your algorithm, or have anything like a representation of that algorithm, in order to predict whether you’ll one box or two-box. Perhaps the Scots tend to one-box, whereas the English tend to two-box.</blockquote><p>In this example, it seems quite plausible that there’s a (logico-causal) reason for the regularity, so that in the logical counterfactual where you act differently, your reference class also acts somewhat differently. Say you’re Scottish, and 10% of Scots read a particular fairy tale growing up, and this is connected with why you two-box. Then in the counterfactual in which you one-box, it is quite possible that those 10% also one-box. Of course, this greatly weakens the connection between Omega’s prediction and your action; perhaps the change of 10% is not enough to tip the scales in Omega’s prediction.</p><blockquote>But, without any account of Y&S’s notion of subjunctive counterfactuals, we just have no way of assessing whether that’s true or not. Y&S note that specifying an account of their notion of counterfactuals is an ‘open problem,’ but the problem is much deeper than that. Without such an account, it becomes completely indeterminate what follows from FDT, even in the core examples that are supposed to motivate it — and that makes FDT not a new decision theory so much as a promissory note.</blockquote><p>In the TDT document, Eliezer addresses this concern by pointing out that CDT also takes a description of the causal structure of a problem as given, begging the question of how we learn causal counterfactuals. In this regard, FDT and CDT are on the same level of promissory-note-ness.</p><p>It might, of course, be taken as much more plausible that a technique of learning the physical-causal structure can be provided, in contrast to a technique which learns the logical-counterfactual structure.</p><p>I want to inject a little doubt about which is easier. If a robot is interacting with an exact simulation of itself (in an iterated prisoner’s dilemma, say), won’t it be easier to infer that it directly controls the copy than it is to figure out that the two are running on different computers and thus causally independent?</p><p>Put more generally: logical uncertainty has to be handled one way or another; it cannot be entirely put aside. Existing methods of testing causality are not designed to deal with it. It stands to reason that such methods applied naively to cases including logical uncertainty would treat such uncertainty like physical uncertainty, and therefore tend to produce logical-counterfactual structure. This would not necessarily be very good for FDT purposes, being the result of unprincipled accident—and the concern for FDT’s counterfactuals is that there may <em>be</em> no principled foundation. Still, I tend to think that other decision theories merely brush the problem under the rug, and actually have to deal with logical counterfactuals one way or another.</p><blockquote>Indeed, on the most plausible ways of cashing this out, it doesn’t give the conclusions that Y&S would want. If I imagine the closest world in which 6288 + 1048 = 7336 is false (Y&S’s example), I imagine a world with laws of nature radically unlike ours — because the laws of nature rely, fundamentally, on the truths of mathematics, and if one mathematical truth is false then either (i) mathematics as a whole must be radically different, or (ii) all mathematical propositions are true because it is simple to prove a contradiction and every propositions follows from a contradiction.</blockquote><p>To this I can only say again that FDT’s problem of defining counterfactuals seems not so different to me from CDT’s problem. A causal decision theorist should be able to work in a mathematical universe; indeed, this seems rather consistent with the ontology of modern science, though not forced by it. I find it implausible that a CDT advocate should have to deny Tegmark’s mathematical universe hypothesis, or should break down and be unable to make decisions under the supposition. So, physical counterfactuals seem like they have to be at least capable of being logical counterfactuals (perhaps a different sort of logical counterfactual than FDT would use, since physical counterfactuals are supposed to give certain different answers, but a sort of logical counterfactual nonetheless).</p><p>(But this conclusion is far from obvious, and I don’t expect ready agreement that CDT has to deal with this.)</p>abramdemskifKX48wWkZ9HGQjHJsSat, 14 Sep 2019 01:43:48 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-ykfSJynXigv7uuvGm
<p>Response to Section VIII:</p><blockquote>An alternative approaches that captures the spirit of FDT’s aims</blockquote><p>I’m somewhat confused about how you can buy FDT as far as you seem to buy it in this section, while also claiming not to understand FDT to the point of saying there is no sensible perspective at all in which it can be said to achieve higher utility. From the perspective in this section, it seems you can straightforwardly interpret FDT’s notion of expected utility maximization via an evaluative focal point such as “the output of the algorithm given these inputs”.</p><p>This evaluative focal point addresses the concern you raise about how bounded ability to implement decision procedures interacts with a “best decision procedure” evaluative focal point (making it depart from FDT’s recommendations in so far as the agent can’t manage to act like FDT), since those concerns don’t arise (at least not so clearly) when we consider what FDT would recommend for the response to one situation in particular. On the other hand, we also can make sense of the notion that taking the bomb is best, since (according to both global-CDT and global-EDT) it is best for an algorithm to output “left” when given the inputs of the bomb problem (in that it gives us the best news about how that agent would do in bomb problems, and causes the agent to do well when put in bomb problems, in so far as a causal intervention on the output of the algorithm also affects a predictor running the same algorithm).</p>abramdemskiykfSJynXigv7uuvGmSat, 14 Sep 2019 00:48:18 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-eBxJXJW3rfr6uXbK7
<p>Responses to Sections V and VI:</p><blockquote>Implausible discontinuities</blockquote><p>I’m puzzled by this concern. Is the doctrine of expected utility plagued by a corresponding ‘implausible discontinuity’ problem because if action 1 has expected value .999 and action 2 has expected value 1, then you should take action 2, but a very small change could mean you should take action 1? Is CDT plagued by an implausible-discontinuity problem because two problems which EDT would treat as the same will differ in causal expected value, and there must be some in-between problem where uncertainty about the causal structure balances between the two options, so CDT’s recommendation implausibly makes a sharp shift when the uncertainty is jiggled a little? Can’t we similarly boggle at the implausibility that a tiny change in the physical structure of a problem should make such a large difference in the causal structure so as to change CDT’s recommendation? (For example, the tiny change can be a small adjustment to the coin which determines which of two causal structures will be in play, with no overall change in the evidential structure.)</p><p>It seems like what you find implausible about FDT here has nothing to do with discontinuity, unless you find CDT and EDT similarly implausible.</p><blockquote>FDT is deeply indeterminate</blockquote><p>This is obviously a big challenge for FDT; we don’t know what logical counterfactuals look like, and invoking them is problematic until we do.</p><p>However, I can point to some toy models of FDT which lend credence to the idea that there’s something there. The most interesting may be MUDT (see the “modal UDT” section of <a href="/posts/5bd75cc58225bf0670374efd/forum-digest-updateless-decision-theory">this summary post</a>). This decision theory uses the notion of “possible” from the modal logic of provability, so that despite being a deterministic agent and therefore only taking one particular action in fact, agents have a well-defined possible-world structure to consider in making decisions, derived from what they can prove.</p><p>I have a post planned that focuses on a different toy model, single-player extensive-form games. This has the advantage of being only as exotic as standard game theory.</p><p>In both of these cases, FDT can be well-specified (at least, to the extent we’re satisfied with calling the toy DTs examples of FDT—which is a bit awkward, since FDT is kind of a weird umbrella term for several possible DTs, but also kind of specifically supposed to use functional graphs, which MUDT doesn’t use).</p><p>It bears mentioning that a Bayesian already regards the probability distribution representing a problem to be deeply indeterminate, so this seems less bad if you start from such a perspective. Logical counterfactuals can similarly be thought of as subjective objects, rather than some objective fact which we have to uncover in order to know what FDT does.</p><p>On the other hand, greater indeterminacy is still worse; just because we already have lots of degrees of freedom to mess ourselves up with doesn’t mean we happily accept even more.</p><blockquote>And in general, it seems to me, there’s no fact of the matter about which algorithm a physical process is implementing in the absence of a particular interpretation of the inputs and outputs of that physical process.</blockquote><p>Part of the reason that I’m happy for FDT to need such a fact is that I think I need such a fact anyway, in order to deal with anthropic uncertainty, and other issues.</p><p>If you don’t think there’s such a fact, then you can’t take a computationalist perspective on theory of mind—in which case, I wonder what position you take on questions such as consciousness. Obviously this leads to a number of questions which are quite aside from the point at hand, but I would personally think that questions such as whether an organism is experiencing suffering have to do with what computations are occurring. This ultimately cashes out to physical facts, yes, but it seems as if suffering should be a fundamentally computational fact which cashes out in terms of physical facts only in a substrate-independent way (ie, the physical facts of importance are precisely those which pertain to the question of which computation is running).</p><blockquote>But almost all accounts of computation in physical processes have the issue that very many physical processes are running very many different algorithms, all at the same time.</blockquote><p>Indeed, I think this is one of the main obstacles to a satisfying account—a successful account should not have this property.</p>abramdemskieBxJXJW3rfr6uXbK7Sat, 14 Sep 2019 00:39:57 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-yxPiR9JXdonQmueAZ
<p>Response to Section VII:</p><blockquote>Assessing by how well the decision-maker does in possible worlds that she isn’t in fact in doesn’t seem a compelling criterion (and EDT and CDT could both do well by that criterion, too, depending on which possible worlds one is allowed to pick).</blockquote><p>You make the claim that EDT and CDT can claim optimality in exactly the same way that FDT can, here, but I think the arguments are importantly not symmetric. CDT and EDT are optimal according to their own optimality notions, but <em>given the choice to implement different decision procedures on later problems</em>, both the CDT and EDT optimality notions would endorse selecting FDT over themselves in many of the problems mentioned in the paper, whereas FDT will endorse itself.</p><p>Most of this section seems to me to be an argument to make careful level distinctions, in an attempt to avoid the level-crossing argument which is FDT’s main appeal. Certainly, FDTers such as myself will often use language which confuses the various levels, since we take a position which says they should be confusable—the best decision procedures should follow the best policies, which should take the best actions. But making careful level distinctions does not block the level-crossing argument, it only clarifies it. FDT may not be the only “consistent fixed-point of normativity” (to the extent that it even is that), but CDT and EDT are clearly not that.</p><blockquote>Fourth, arguing that FDT does best in a class of ‘fair’ problems, without being able to define what that class is or why it’s interesting, is a pretty weak argument.</blockquote><p>I basically agree that the FDT paper dropped the ball here, in that it could have given a toy setting in which ‘fair’ is rigorously defined (in a pretty standard game-theoretic setting) and FDT has the claimed optimality notion. I hope my longer writeup can make such a setting clear.</p><p>Briefly: my interpretation of the “FDT does better” claim in the FDT paper is that FDT is supposed to take UDT-optimal actions. To the extent that it doesn’t take UDT-optimal actions, I mostly don’t endorse the claim that it does better (though I plan to note in a follow-up post an alternate view in which the FDT notion of optimality may be better).</p><p>The toy setting I have in mind that makes “UDT-optimal” completely well-defined is actually fairly general. The idea is that if we can represent a decision problem as a (single-player) extensive-form game, UDT is <em>just</em> the idea of throwing out the requirement of subgame-optimality. In other words, we don’t even need a notion of “fairness” in the setting of extensive-form games—the setting isn’t rich enough to represent any “unfair” problems. Yet it is a pretty rich setting.</p><p>This observation was already made here: <a href="/posts/W4sDWwGZ4puRBXMEZ/single-player-extensive-form-games-as-a-model-of-udt" class="bare-url">https://www.lesswrong.com/posts/W4sDWwGZ4puRBXMEZ/single-player-extensive-form-games-as-a-model-of-udt</a>. Note that there are some concerns in the comments. I think the concerns make sense, and I’m not quite sure how I want to address them, but I also don’t think they’re damning to the toy model.</p><p>The FDT paper may have left out this model out of a desire for greater generality, which I do think is an important goal—from my perspective, it makes sense not to reduce things to the toy model in which everything works out nicely.</p>abramdemskiyxPiR9JXdonQmueAZSat, 14 Sep 2019 00:39:50 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-X4CLhPpf9zecdGxAq
<p>Here are some (very lightly edited) comments I left on Will’s draft of this post. (See also my <a href="/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-y8zRwcpNeu2ZhM3yE">top-level response</a>.)</p><p>Responses to Sections II and III:</p><blockquote>I’m not claiming that it’s clear what this means. E.g. see <a href="https://www.umsu.de/wo/2018/688">here,</a> second bullet point, arguing there can be no such probability function, because any probability function requires certainty in logical facts and all their entailments.</blockquote><p>This point shows the intertwining of logical counterfactuals (counterpossibles) and logical uncertainty. I take <a href="https://intelligence.org/2016/09/12/new-paper-logical-induction/">logical induction</a> to represent significant progress generalizing probability theory to the case of logical uncertainty, ie, objects which have many of the virtues of probability functions while not requiring certainty about entailment of known facts. So, we can substantially reply to this objection.</p><p>However, replying to this objection does not necessarily mean we can define logical counterfactuals as we would want. So far we have only been able to use logical induction to specify a kind of “logically uncertain evidential conditional”. (IE, something closer in spirit to EDT, which does behave more like FDT in some problems but not in general.)</p><p>I want to emphasize that I agree that specifying what logical counterfactuals are is a grave difficulty, so grave as to seem (to me, at present) to be damning, provided one can avoid the difficulty in some other approach. However, I don’t actually think that the difficulty can be avoided in any other approach! I think CDT ultimately has to grapple with the question as well, because physics is math, and so physical counterfactuals are ultimately mathematical counterfactuals. Even EDT has to grapple with this problem, ultimately, due to the need to handle cases where one’s own action can be logically known. (Or provide a convincing argument that such cases cannot arise, even for an agent which is computable.)</p><blockquote>Guaranteed Payoffs: In conditions of certainty — that is, when the decision-maker has no uncertainty about what state of nature she is in, and no uncertainty about the utility payoff of each action is — the decision-maker should choose the action that maximises utility.</blockquote><p>(Obligatory remark that what maximizes utility is part of what’s at issue here, and for precisely this reason, an FDTist could respond that it’s CDT and EDT which fail in the Bomb example—by failing to maximize the <em>a priori</em> expected utility of the action taken.)</p><p>FDT would disagree with this principle in general, since full certainty implies certainty about one’s action, and the utility to be received, as well as everything else. However, I think we can set that aside and say there’s a version of FDT which would agree with this principle in terms of prior uncertainty. It seems cases like Bomb cannot be set up without either invoking prior uncertainty (taking the form of the predictor’s failure rate) or bringing the question of how to deal with logically impossible decisions to the forefront (if we consider the case of a perfect predictor).</p><p>Why should prior uncertainty be important, in cases of posterior certainty? Because of the <a href="/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-y8zRwcpNeu2ZhM3yE">prior-optimality</a> notion (in which a decision theory is judged on a decision problem based on the utility received in expectation according to the prior probability which defines the decision problem).</p><p>Prior-optimality considers the guaranteed-payoff objection to be very similar to objecting to a gambling strategy by pointing out that the gambling strategy sometimes loses. In Bomb, the problem clearly stipulates that an agent who follows the FDT recommendation has a trillion trillion to one odds of doing better than an agent who follows the CDT/EDT recommendation. Complaining about the one-in-a-trillion-trillion chance that you get the bomb while being the sort of agent who takes the bomb is, to an FDT-theorist, like a gambler who has just lost a trillion-trillion-to-one bet complaining that the bet doesn’t look so rational now that the outcome is known with certainty to be the one-in-a-trillion-trillion case where the bet didn’t pay well.</p><blockquote>The right action, according to FDT, is to take Left, in the full knowledge that as a result you will slowly burn to death. Why? Because, using Y&S’s counterfactuals, i<em>f </em>your algorithm were to output ‘Left’, then it would also have outputted ‘Left’ when the predictor made the simulation of you, and there would be no bomb in the box, and you could save yourself $100 by taking Left.</blockquote><p>And why, on your account, is this implausible? To my eye, this is <em>right there in the decision problem</em>, not a weird counterintuitive consequence of FDT: the decision problem <em>stipulates</em> that algorithms which output ‘left’ will not end up in the situation of taking a bomb, with very, very high probability.</p><p>Again, complaining that you now know with certainty that you’re in the unlucky position of seeing the bomb seems irrelevant in the way that a gambler complaining that they now know how the dice fell seems irrelevant—it’s still best to gamble according to the odds, taking the option which gives the best chance of success.</p><p>(But what I most want to convey here is that there is <strong><em>a</em></strong> coherent sense in which FDT does the optimal thing, whether or not one agrees with it.)</p><p>One way of thinking about this is to say that the FDT notion of “decision problem” is different from the CDT or EDT notion, in that FDT considers the prior to be of primary importance, whereas CDT and EDT consider it to be of no importance. If you had instead specified ‘bomb’ with just the certain information that ‘left’ is (causally and evidentially) very bad and ‘right’ is much less bad, then CDT and EDT would regard it as precisely the same decision problem, whereas FDT would consider it to be a radically different decision problem.</p><p>Another way to think about this is to say that FDT “rejects” decision problems which are improbable <em>according to their own specification</em>. In cases like Bomb where the situation as described is by its own description a one in a trillion trillion chance of occurring, FDT gives the outcome only one-trillion-trillion-th consideration in the expected utility calculation, when deciding on a strategy.</p><p>Also, I note that this analysis (on the part of FDT) does not hinge in this case on exotic counterfactuals. If you set Bomb up in the Savage framework, you would be forced to either give only the certain choice between bomb and not-bomb (so you don’t represent the interesting part of the problem, involving the predictor) or to give the decision in terms of the prior, in which case the Savage framework would endorse the FDT recommendation.</p><p>Another framework in which we could arrive at the same analysis would be that of single-player extensive-form games, in which the FDT recommendation corresponds to the simple notion of optimal strategy, whereas the CDT recommendation amounts to the stipulation of subgame-optimality.</p>abramdemskiX4CLhPpf9zecdGxAqSat, 14 Sep 2019 00:35:59 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-YJSNi37kRY2yPpSeQ
<p>Replying to one of Will’s edits on account of my comments to the earlier draft:</p><blockquote>Finally, in a comment on a draft of this note, Abram Demski said that: “The notion of expected utility for which FDT is supposed to do well (at least, according to me) is <em>expected utility with respect to the prior for the decision problem under consideration.</em>” If that’s correct, it’s striking that this criterion isn’t mentioned in the paper. But it also doesn’t seem compelling as a principle by which to evaluate between decision theories, nor does it seem FDT even does well by it. To see both points: suppose I’m choosing between an avocado sandwich and a hummus sandwich, and my prior was that I prefer avocado, but I’ve since tasted them both and gotten evidence that I prefer hummus. The choice that does best in terms of expected utility with respect to my prior for the decision problem under consideration is the avocado sandwich (and FDT, as I understood it in the paper, would agree). But, uncontroversially, I should choose the hummus sandwich, because I prefer hummus to avocado.</blockquote><p>Yeah, the thing is, the FDT paper focused on examples where “expected utility according to the prior” becomes an unclear notion due to logical uncertainty issues. It wouldn’t have made sense for the FDT paper to focus on that, given the desire to put the most difficult issues into focus. However, FDT is supposed to accomplish similar things to UDT, and UDT provides the more concrete illustration.</p><p>The policy that does best in expected utility according to the prior is the policy of taking whatever you like. In games of partial information, decisions are defined as functions of <em>information states</em>; and in the situation as described, there are separate information states for liking hummus and liking avocado. Choosing the one you like achieves a higher expected utility according to the prior, in comparison to just choosing avocado no matter what. In this situation, optimizing the decision in this way is equivalent to updating on the information; but, not always (as in transparent newcomb, Bomb, and other such problems).</p><p>To re-state that a different way: in a given information state, UDT is choosing what to do <em>as a function of the information available</em>, and judging the utility of that choice <em>according to the prior</em>. So, in this scenario, we judge the expected utility <em>of selecting avocado in response to liking hummus</em>. This is worse (according to the prior!) than selecting hummus in response to liking hummus.</p>abramdemskiYJSNi37kRY2yPpSeQFri, 13 Sep 2019 22:07:27 +0000Comment by abramdemski on A Critique of Functional Decision Theory
https://www.greaterwrong.com/posts/ySLYSsNeFL5CoAQzN/a-critique-of-functional-decision-theory#comment-y8zRwcpNeu2ZhM3yE
<p>I saw an earlier draft of this, and hope to write an extensive response at some point. For now, the short version:</p><p>As I understand it, FDT was intended as an umbrella term for MIRI-style decision theories, which illustrated the critical points without making too many commitments. So, the vagueness of FDT was partly by design.</p><p>I think UDT is a more concrete illustration of the most important points relevant to this discussion.</p><ul><li><p>The optimality notion of UDT is clear. “UDT gets the most utility” means “UDT gets the highest expected value with respect to its own prior”. This seems quite well-defined, hopefully addressing your (VII).<ul><li><p>There are <a href="/posts/9sYzoRnmqmxZm4Whf/conceptual-problems-with-udt-and-policy-selection">problems applying UDT to realistic situations</a>, but UDT <em>makes perfect sense and is optimal in a straightforward sense</em> for the case of <a href="/posts/W4sDWwGZ4puRBXMEZ/single-player-extensive-form-games-as-a-model-of-udt">single-player extensive form games</a>. That doesn’t address multi-player games or logical uncertainty, but it is enough for much of Will’s discussion.</p></li><li><p>FDT focused on the weird logical cases, which is in fact a major part of the motivation for MIRI-style decision theory. However, UDT for single-player extensive-form games actually gets at a lot of what MIRI-style decision theory wants, without broaching the topic of logical counterfactuals or proving-your-own-action directly.</p></li><li><p>The problems which create a deep indeterminacy seem, to me, to be problems for other decision theories than FDT as well. FDT is trying to face them head-on. But there are big problems for applying EDT to agents who are physically instantiated as computer programs and can prove too much about their own actions.</p></li></ul></p></li><li><p>This also hopefully clarifies the sense in which I don’t think the decisions pointed out in (III) are bizarre. The decisions are optimal <em>according to the very probability distribution used to define the decision problem.</em> <ul><li><p>There’s a subtle point here, though, since Will describes the decision problem from an updated perspective—you already know the bomb is in front of you. So UDT “changes the problem” by evaluating “according to the prior”. From my perspective, because <em>the very statement of the <strong>Bomb</strong> problem</em> suggests that there were also other possible outcomes, we can rightly insist to evaluate expected utility in terms of those chances.</p></li><li><p>Perhaps this sounds like an unprincipled rejection of the <em><strong>Bomb</strong></em> problem as you state it. My principle is as follows: <em>you should not state a decision problem without having in mind a well-specified way to predictably put agents into that scenario</em>. Let’s call the way-you-put-agents-into-the-scenario the “construction”. We then evaluate agents on how well they deal with the construction.<ul><li><p>For examples like <em><strong>Bomb</strong></em>, the construction gives us the overall probability distribution—this is then used for the expected value which UDT’s optimality notion is stated in terms of.</p></li><li><p>For other examples, as discussed in <a href="https://intelligence.org/2017/04/07/decisions-are-for-making-bad-outcomes-inconsistent/">Decisions are for making bad outcomes inconsistent</a>, the construction simply breaks when you try to put certain decision theories into it. This can also be a good thing; it means the decision theory makes certain scenarios altogether impossible.</p></li></ul></p></li></ul></p></li></ul><p>The point about “constructions” is possibly a bit subtle (and hastily made); maybe a lot of the disagreement will turn out to be there. But I <em>do hope</em> that the basic idea of UDT’s optimality criterion is actually clear—“evaluate expected utility of policies according to the prior”—and clarifies the situation with FDT as well.</p>abramdemskiy8zRwcpNeu2ZhM3yEFri, 13 Sep 2019 21:38:01 +0000Do Sufficiently Advanced Agents Use Logic? by abramdemski
https://www.greaterwrong.com/posts/3qXE6fK47JhSfkpnB/do-sufficiently-advanced-agents-use-logic
<p><em>This is a continuation of a discussion with Vanessa from the MIRIxDiscord group. I’ll make some comments on things Vanessa has said, but those should not be considered a summary of the discussion so far. My comments here are also informed by discussion with Sam.</em></p><h1>1: Logic as Proxy</h1><h2>1a: The Role of Prediction</h2><p>Vanessa has said that predictive accuracy is sufficient; consideration of logic is not needed to judge (<a href="/posts/5bd75cc58225bf06703753a4/learning-incomplete-models-using-dominant-markets">partial</a>) models. A hypothesis should ultimately ground out to perceptual information. So why is there any need to consider other sorts of “predictions” it can make? (IE, why should we think of it as possessing internal propositions which have a logic of their own?)</p><p>But similarly, why should agents use predictive accuracy to learn? What’s the argument for it? Ultimately, predicting perceptions ahead of time should only be in service of achieving higher reward.</p><p>We could instead learn from reward feedback alone. A (partial) “hypothesis” would really be a (partial) <em>strategy</em>, helping us to generate actions. We would judge strategies on (something like) average reward achieved, not even trying to predict precise reward signals. The agent still receives incoming perceptual information, and strategies can use it to update internal states and to inform actions. However, strategies are not asked to produce any predictions. (The framework I’m describing is, of course, model-free RL.)</p><p>Intuitively, it seems as if this is missing something. A model-based agent can learn a lot about the world just by watching, taking no actions. However, individual strategies can implement prediction-based learning within themselves. So, it seems difficult to say what benefit model-based RL provides beyond model-free RL, besides a better prior over strategies.</p><p>It might be that we can’t say anything recommending model-based learning over model free in a standard bounded-regret framework. (I actually haven’t thought about it much—but the argument that model-free strategies can implement models internally seems potentially strong. Perhaps you just can’t get much in the AIXI framework because there are no good loss bounds in that framework at all, <a href="/posts/5bd75cc58225bf0670375575/the-learning-theoretic-ai-alignment-research-agenda">as Vanessa mentions</a>.) However, if so, this seems like a weakness of standard bounded-regret frameworks. Predicting the world seems to be a significant aspect of intelligence; we should be able to talk about this formally somehow.</p><p>Granted, it doesn’t make sense for bounded agents to pursue predictive accuracy above all else. There is a computational trade-off, and you don’t need to predict something which isn’t important. My claim is something like, you should try and predict when you don’t yet have an effective strategy. After you have an effective strategy, you don’t really need to generate predictions. Before that, you need to generate predictions because you’re still grappling with the world, trying to understand what’s basically going on.</p><p>If we’re trying to <a href="/posts/5bd75cc58225bf0670375575/the-learning-theoretic-ai-alignment-research-agenda">understand intelligence</a>, the idea that model-free learners can internally manage these trade-offs (by choosing strategies which judiciously choose to learn from predictions when it is efficacious to do so) seems less satisfying than a proper theory of learning from prediction. What is fundamental vs non-fundamental to intelligence can get fuzzy, but learning from prediction seems like something we expect any sufficiently intelligent agent to do (whether it was built-in or learned behavior).</p><p>On the other hand, judging hypotheses on their predictive accuracy is kind of a weird thing to do if what you ultimately want a hypothesis to do for you is generate actions. It’s like this: You’ve got two tasks; task A, and task B. Task A is what you really care about, but it might be quite difficult to tackle on its own. Task B is really very different from task A, but you can get a lot of feedback on task B. So you ask your hypotheses to compete on task B, and judge them on that in addition to task A. Somehow you’re expecting to get a lot of information about task A from performance on task B. And indeed, it seems you do: predictive accuracy of a hypothesis is somehow a useful proxy for efficacy of that hypothesis in guiding action.</p><p>(It should also be noted that a reward-learning framework presumes we get feedback about utility at all. If we get no feedback about reward, then we’re forced to <em>only</em> judge hypotheses by predictions, and make what inferences about utility we will. A dire situation for learning theory, but a situation where we can still talk about rational agency more generally.)</p><h2>1b: The Analogy to Logic</h2><p>My argument is going to be that if achieving high reward is task A, and predicting perception is task B, logic can be task C. Like task B, it is very different from task A. Like task B, it nonetheless provides useful information. Like task B, it seems to me that a theory of (boundedly) rational agency is missing something without it.</p><p>The basic picture is this. Perceptual prediction provides a lot of good feedback about the quality of cognitive algorithms. But if you really want to train up some good cognitive algorithms for yourself, it is helpful to do some imaginative play on the side. </p><p>One way to visualize this is an agent making up math puzzles in order to strengthen its reasoning skills. This might suggest a picture where the puzzles are always well-defined (terminating) computations. However, there’s no special dividing line between decidable and undecidable problems—any particular restriction to a decidable class might rule out some interesting (decidable but non-obviously so) stuff which we could learn from. So we might end up just going with any computations (halting or no).</p><p>Similarly, we might not restrict ourselves to entirely well-defined propositions. It makes a lot of sense to test cognitive heuristics on scenarios closer to life.</p><p>Why do I think sufficiently advanced agents are likely to do this?</p><p>Well, just as it seems important that we can learn a whole lot from prediction before we ever take an action in a given type of situation, it seems important that we can learn a whole lot by reasoning before we even observe that situation. I’m not formulating a precise learning-theoretic conjecture, but intuitively, it is related to whether we could reasonably expect the agent to get something right on the first try. Good perceptual prediction alone does not guarantee that we can correctly anticipate the effects of actions we have never tried before, but if I see an agent generate an effective strategy in a situation it has never intervened in before (but has had opportunity to observe), I expect that internally it is learning from perception at some level (even if it is model-free in overall architecture). Similarly, if I see an agent quickly pick up a reasoning-heavy game like chess, then I suspect it of learning from hypothetical simulations at some level.</p><p>Again, “on the first try” is not supposed to be a formal learning-theoretic requirement; I realize you can’t exactly expect anything to work on the first try with learning agents. What I’m getting at has something to do with generalization.</p><h1>2: Learning-Theoretic Criteria</h1><p>Part of the frame has been learning-theory-vs-logic. One might interpret my closing remarks from the previous section that way; I don’t know how to formulate my intuition learning-theoretically, but I expect that reasoning helps agents in particular situations. It may be that the phenomena of the previous section cannot be understood learning-theoretically, and only amount to a “better prior over strategies” as I mentioned. However, I don’t want it to be a learning-theory-vs-logic argument. I would hope that something learning-theoretic can be said in favor of learning from perception, and in favor of learning from logic. Even if it can’t, learning theory is still an important component here, regardless of the importance of logic.</p><p>I’ll try to say something about how I think learning theory should interface with logic.</p><p>Vanessa <a href="/posts/S3W4Xrmp6AL7nxRHd/formalising-decision-theory-is-hard#comment-X7R4rxHpkEKycJvSb">said some relevant things in a comment</a>, which I’ll quote in full:</p><blockquote>Heterodox opinion: I think the entire MIRIesque (<em>and</em> academic philosophy) approach to decision theory is confused. The basic assumption seems to be, that we can decouple the problem of learning a model of the world from the problem of taking a decision given such a model. We then ignore the first problem, and assume a particular shape for the model (for example, causal network) which allows us to consider decision theories such as CDT, EDT etc. However, in reality the two problems cannot be decoupled. This is because the <em>type signature </em>of a world model is only meaningful if it comes with an algorithm for how to learn a model of this type.</blockquote><blockquote>For example, consider Newcomb’s paradox. The agent makes a decision under the assumption that Omega behaves in a certain way. But, where did the assumption come from? Realistic agents have to <em>learn</em> everything they know. Learning normally requires a time sequence. For example, we can consider the <em>iterated</em> Newcomb’s paradox (INP). In INP, any reinforcement learning (RL) algorithm will converge to one-boxing, simply because one-boxing gives it the money. This is despite RL naively looking like CDT. Why does it happen? Because in the learned model, the “causal” relationships are <em>not</em> physical causality. The agent comes to believe that taking the one box <em>causes</em> the money to appear there.</blockquote><blockquote>In Newcomb’s paradox EDT succeeds but CDT fails. Let’s consider an example where CDT succeeds and EDT fails: the XOR blackmail. The iterated version would be IXB. In IXB, classical RL doesn’t guarantee much because the environment is more complex than the agent (it contains Omega). To overcome this, we can use RL with <a href="/posts/5bd75cc58225bf0670375575/the-learning-theoretic-ai-alignment-research-agenda">incomplete models</a>. I believe that this indeed solves both INP and IXB.</blockquote><blockquote>Then we can consider e.g. counterfactual mugging. In counterfactual mugging, RL with incomplete models doesn’t work. That’s because the assumption that Omega responds in a way that depends on a counterfactual world is not in the space of models at all. Indeed, it’s unclear how can any agent learn such a fact from empirical observations. One way to fix it is by allowing the agent to precommit. Then the assumption about Omega becomes empirically verifiable. But, if we do this, then RL with incomplete models can solve the problem again.</blockquote><blockquote>The only class of problems that I’m genuinely unsure how to deal with is game-theoretic superrationality. However, I also don’t see much evidence the MIRIesque approach has succeeded on that front. We probably need to start with just solving the grain of truth problem in the sense of converging to ordinary Nash (or similar) equilibria (which might be possible using incomplete models). Later we can consider agents that observe each other’s source code, and maybe something along the lines of <a href="/posts/5bd75cc58225bf0670375058/superrationality-in-arbitrary-games">this</a> can apply.</blockquote><p>Besides the MIRI-vs-learning frame, I agree with a lot of this. I <a href="/posts/hLFD6qSN9MmQxKjG5/embedded-agency-via-abstraction#comment-jbC8934HoXaoXrqhR">wrote a comment</a> elsewhere making some related points about the need for a learning-theoretic approach. Some of the points also relate to my <a href="/s/fgHSwxFitysGKHH56">CDT=EDT sequence</a>; I have been arguing that CDT and EDT don’t behave as people broadly imagine (often not having the bad behavior which people broadly imagine). Some of those arguments were learning-theoretic while others were not, but the conclusions were similar either way.</p><p>In any case, I think the following criterion (originally mentioned to me by Jack Gallagher) makes sense:</p><p><em>A decision problem should be conceived as a sequence, but the algorithm deciding what to do on a particular element of the sequence should not know/care what the whole sequence is.</em></p><p><a href="/posts/yXCvYqTZCsfN7WRrg/asymptotic-decision-theory-improved-writeup">Asymptotic decision theory</a> was the first major proposal to conceive of decision problems as sequences in this way. Decision-problem-as-sequence allows decision theory to be addressed learning-theoretically; we can’t expect a learning agent to necessarily do well in any particular case (because it could have a sufficiently poor prior, and so still be learning in that particular case), but we <em>can</em> expect it to <em>eventually</em> perform well (provided the problem meets some “fairness” conditions which make it learnable).</p><p>As for the second part of the criterion, requiring that the agent is ignorant of the overall sequence when deciding what to do on an instance: this captures the idea of learning from logic. Providing the agent with the sequence is cheating, because you’re essentially giving the agent your interpretation of the situation.</p><p>Jack mentioned this criterion to me in a discussion of <a href="/posts/5bd75cc58225bf067037541d/an-approach-to-logically-updateless-decisions">averaging decision theory (AvDT)</a>, in order to explain why AvDT was cheating. </p><p>AvDT is based on a fairly simple idea: look at the average performance of a strategy so far, rather than its expected performance on this particular problem. Unfortunately, “performance so far” requires things to be defined in terms of a training sequence (counter to the logical-induction philosophy of non-sequential learning).</p><p>I created AvDT to try and address some shortcomings of asymptotic decision theory (let’s call it AsDT). Specifically, AsDT does not do well in counterlogical mugging. AvDT is capable of doing well in counterfactual mugging. However, it depends on the training sequence. <a href="/posts/oZwxY88NCCHffJuxM/a-problem-about-bargaining-and-logical-uncertainty">Counterlogical mugging</a> requires the agent to decide on the “probability” of Omega asking for money vs paying up, to figure out whether participation is worth it overall. AvDT solves this problem by looking at the training sequence to see how often Omega pays up. So, the problem of doing well in decision problems is “reduced” to specifying good training sequences. This (1) doesn’t obviously make things easier, and (2) puts the work on the human trainers.</p><p>Jack is saying that the system should be looking through logic <em>on its own</em> to find analogous scenarios to generalize from. When judging whether a system gets counterlogical mugging right, <em>we</em> have to define counterlogical mugging as a sequence to enable learning-theoretic analysis; but <em>the agent</em> has to figure things out on its own.</p><p>This is a somewhat subtle point. A realistic agent experiences the world sequentially, and learns by treating its history as a training sequence of sorts. This is <em>physical time</em>. I have no problem with this. What I’m saying is that if an agent is also learning <em>from analogous circumstances within logic</em>, as I suggested sophisticated agents will do in the first part, then Jack’s condition should come into play. We aren’t handed, from on high, a sequence of logically defined scenarios which we can locate ourselves within. We only have regular physical time, plus a bunch of hypothetical scenarios which we can define and whose relevance we have to determine.</p><p>This gets back to my earlier intuition about agents having a reasonable chance of getting certain things right on the first try. Learning-theoretic agents don’t get things right on the first try. However, agents who learn from logic have “lots of tries” before their first real try in physical time. If you can successfully determine which logical scenarios are relevantly analogous to your own, you can learn what to do just by thinking. (Of course, you still need a lot of physical-time learning to know enough about your situation to do that!)</p><p>So, getting back to Vanessa’s point in the comment I quoted: can we solve MIRI-style decision problems by considering the iterated problem, rather than the single-shot version? To a large extent, I think so: <a href="/posts/dKAJqBDZRMMsaaYo5/in-logical-time-all-games-are-iterated-games">in logical time, all games are iterated games</a>. However, I don’t want to have to set an agent up with a training sequence in which it encounters those specific problems many times. For example, finding good strategies in chess via self-play should come naturally from the way the agent thinks about the world, rather than being an explicit training regime which the designer has to implement. Once the rules for chess are understood, the bottleneck should be thinking time rather than (physical) training instances.</p>abramdemski3qXE6fK47JhSfkpnBFri, 13 Sep 2019 19:53:36 +0000Comment by abramdemski on Rationality Exercises Prize of September 2019 ($1,000)
https://www.greaterwrong.com/posts/5CWpknwhonrpffKMH/rationality-exercises-prize-of-september-2019-usd1-000#comment-MEJh6pYLvveQKFP74
<p>By asking people to leave a comment here linking to their exercises, are you discouraging writing exercises directly as a comment to this post? (Perhaps you’re wanting something longer, and so discouraging comments as the arena for listing exercises?)</p>abramdemskiMEJh6pYLvveQKFP74Fri, 13 Sep 2019 03:33:29 +0000Comment by abramdemski on Embedded Agency via Abstraction
https://www.greaterwrong.com/posts/hLFD6qSN9MmQxKjG5/embedded-agency-via-abstraction#comment-XenkPobqgAjonwios
<p>I agree that if a point can be addressed or explored in a static framework, it can be easier to do that first rather than going to the fully dynamic picture.</p><p>On the other hand, I think your discussion of the cat overstates the case. Your own analysis of the decision theory of a single-celled organism (ie the perspective you’ve described to me in person) compares it to gradient descent, rather than expected utility maximization. This is a fuzzy area, and certainly doesn’t achieve all the things I mentioned, but doesn’t that seem more “dynamic” than “static”? Today’s deep learning systems aren’t as generally intelligent as cats, but it seems like the gap exists more within learning theory than static decision theory.</p><p>More importantly, although the static picture can be easier to analyse, it has also been much more discussed for that reason. The low-hanging fruits are more likely to be in the more neglected direction. Perhaps the more difficult parts of the dynamic picture (perhaps robust delegation) can be put aside while still approaching things from a learning-theoretic perspective.</p><p>I may have said something along the lines of the static picture already being essentially solved by reflective oracles (the problems with reflective oracles being typical of the problems with the static approach). From my perspective, it seems like time to move on to the dynamic picture in order to make progress. But that’s overstating things a bit—I am interested in better static pictures, particularly when they are suggestive of dynamic pictures, such as <a href="/posts/Rcwv6SPsmhkgzfkDw/edt-solves-5-and-10-with-conditional-oracles">COEDT</a>.</p><p>In any case, I have no sense that you’re making a mistake by looking at abstraction in the static setting. If you have traction, you should continue in that direction. I generally suspect that the abstraction angle is valuable, whether static or dynamic.</p><p>Still, I do suspect we have material disagreements remaining, not only disagreements in research emphasis.</p><p>Toward the end of your comment, you speak of the one-shot picture and the dynamic picture as if the two are mutually exclusive, rather than just easy mode vs hard mode as you mention early on. A learning picture still admits static snapshots. Also, cats don’t get everything right on the first try. </p><p>Still, I admit: a weakness of an asymptotic learning picture is that it seems to eschew finite problems; to such an extent that at times I’ve said the dynamic learning picture serves as the easy version of the problem, with one-shot rationality being the hard case to consider later. Toy static pictures—such as the one provided by reflective oracles—give an idealized static rationality, using unbounded processing power and logical omniscience. A <em>real</em> static picture—perhaps the picture you are seeking—would involve bounded rationality, including both logical non-omniscience and regular physical non-omniscience. A static-rationality analysis of logical non-omnincience has seemed quite challenging so far. Nice versions of self-reference and other challenges to embedded world-models such as those you mention seem to require conveniences such as reflective oracles. Nothing resembling <a href="/posts/5bd75cc58225bf0670375313/open-problem-thin-logical-priors">thin priors</a> has come along to allow for eventual logical coherence while resembling bayesian static rationality (rather than logical-induction-like dynamic rationality). And as for the empirical uncertainty, we would really like to <a href="/posts/5bd75cc58225bf0670375510/catastrophe-mitigation-using-drl">get some guarantees about avoiding catastrophic mistakes</a> (though, perhaps, this isn’t within your scope).</p>abramdemskiXenkPobqgAjonwiosFri, 13 Sep 2019 03:01:21 +0000Comment by abramdemski on Formalising decision theory is hard
https://www.greaterwrong.com/posts/S3W4Xrmp6AL7nxRHd/formalising-decision-theory-is-hard#comment-HRJzddD6AYMaSq6bD
<style>.mjx-chtml {display: inline-block; line-height: 0; text-indent: 0; text-align: left; text-transform: none; font-style: normal; font-weight: normal; font-size: 100%; font-size-adjust: none; letter-spacing: normal; word-wrap: normal; word-spacing: normal; white-space: nowrap; float: none; direction: ltr; max-width: none; max-height: none; min-width: 0; min-height: 0; border: 0; margin: 0; padding: 1px 0}
.MJXc-display {display: block; text-align: center; margin: 1em 0; padding: 0}
.mjx-chtml[tabindex]:focus, body :focus .mjx-chtml[tabindex] {display: inline-table}
.mjx-full-width {text-align: center; display: table-cell!important; width: 10000em}
.mjx-math {display: inline-block; border-collapse: separate; border-spacing: 0}
.mjx-math * {display: inline-block; -webkit-box-sizing: content-box!important; -moz-box-sizing: content-box!important; box-sizing: content-box!important; text-align: left}
.mjx-numerator {display: block; text-align: center}
.mjx-denominator {display: block; text-align: center}
.MJXc-stacked {height: 0; position: relative}
.MJXc-stacked > * {position: absolute}
.MJXc-bevelled > * {display: inline-block}
.mjx-stack {display: inline-block}
.mjx-op {display: block}
.mjx-under {display: table-cell}
.mjx-over {display: block}
.mjx-over > * {padding-left: 0px!important; padding-right: 0px!important}
.mjx-under > * {padding-left: 0px!important; padding-right: 0px!important}
.mjx-stack > .mjx-sup {display: block}
.mjx-stack > .mjx-sub {display: block}
.mjx-prestack > .mjx-presup {display: block}
.mjx-prestack > .mjx-presub {display: block}
.mjx-delim-h > .mjx-char {display: inline-block}
.mjx-surd {vertical-align: top}
.mjx-mphantom * {visibility: hidden}
.mjx-merror {background-color: #FFFF88; color: #CC0000; border: 1px solid #CC0000; padding: 2px 3px; font-style: normal; font-size: 90%}
.mjx-annotation-xml {line-height: normal}
.mjx-menclose > svg {fill: none; stroke: currentColor}
.mjx-mtr {display: table-row}
.mjx-mlabeledtr {display: table-row}
.mjx-mtd {display: table-cell; text-align: center}
.mjx-label {display: table-row}
.mjx-box {display: inline-block}
.mjx-block {display: block}
.mjx-span {display: inline}
.mjx-char {display: block; white-space: pre}
.mjx-itable {display: inline-table; width: auto}
.mjx-row {display: table-row}
.mjx-cell {display: table-cell}
.mjx-table {display: table; width: 100%}
.mjx-line {display: block; height: 0}
.mjx-strut {width: 0; padding-top: 1em}
.mjx-vsize {width: 0}
.MJXc-space1 {margin-left: .167em}
.MJXc-space2 {margin-left: .222em}
.MJXc-space3 {margin-left: .278em}
.mjx-test.mjx-test-display {display: table!important}
.mjx-test.mjx-test-inline {display: inline!important; margin-right: -1px}
.mjx-test.mjx-test-default {display: block!important; clear: both}
.mjx-ex-box {display: inline-block!important; position: absolute; overflow: hidden; min-height: 0; max-height: none; padding: 0; border: 0; margin: 0; width: 1px; height: 60ex}
.mjx-test-inline .mjx-left-box {display: inline-block; width: 0; float: left}
.mjx-test-inline .mjx-right-box {display: inline-block; width: 0; float: right}
.mjx-test-display .mjx-right-box {display: table-cell!important; width: 10000em!important; min-width: 0; max-width: none; padding: 0; border: 0; margin: 0}
.MJXc-TeX-unknown-R {font-family: monospace; font-style: normal; font-weight: normal}
.MJXc-TeX-unknown-I {font-family: monospace; font-style: italic; font-weight: normal}
.MJXc-TeX-unknown-B {font-family: monospace; font-style: normal; font-weight: bold}
.MJXc-TeX-unknown-BI {font-family: monospace; font-style: italic; font-weight: bold}
.MJXc-TeX-ams-R {font-family: MJXc-TeX-ams-R,MJXc-TeX-ams-Rw}
.MJXc-TeX-cal-B {font-family: MJXc-TeX-cal-B,MJXc-TeX-cal-Bx,MJXc-TeX-cal-Bw}
.MJXc-TeX-frak-R {font-family: MJXc-TeX-frak-R,MJXc-TeX-frak-Rw}
.MJXc-TeX-frak-B {font-family: MJXc-TeX-frak-B,MJXc-TeX-frak-Bx,MJXc-TeX-frak-Bw}
.MJXc-TeX-math-BI {font-family: MJXc-TeX-math-BI,MJXc-TeX-math-BIx,MJXc-TeX-math-BIw}
.MJXc-TeX-sans-R {font-family: MJXc-TeX-sans-R,MJXc-TeX-sans-Rw}
.MJXc-TeX-sans-B {font-family: MJXc-TeX-sans-B,MJXc-TeX-sans-Bx,MJXc-TeX-sans-Bw}
.MJXc-TeX-sans-I {font-family: MJXc-TeX-sans-I,MJXc-TeX-sans-Ix,MJXc-TeX-sans-Iw}
.MJXc-TeX-script-R {font-family: MJXc-TeX-script-R,MJXc-TeX-script-Rw}
.MJXc-TeX-type-R {font-family: MJXc-TeX-type-R,MJXc-TeX-type-Rw}
.MJXc-TeX-cal-R {font-family: MJXc-TeX-cal-R,MJXc-TeX-cal-Rw}
.MJXc-TeX-main-B {font-family: MJXc-TeX-main-B,MJXc-TeX-main-Bx,MJXc-TeX-main-Bw}
.MJXc-TeX-main-I {font-family: MJXc-TeX-main-I,MJXc-TeX-main-Ix,MJXc-TeX-main-Iw}
.MJXc-TeX-main-R {font-family: MJXc-TeX-main-R,MJXc-TeX-main-Rw}
.MJXc-TeX-math-I {font-family: MJXc-TeX-math-I,MJXc-TeX-math-Ix,MJXc-TeX-math-Iw}
.MJXc-TeX-size1-R {font-family: MJXc-TeX-size1-R,MJXc-TeX-size1-Rw}
.MJXc-TeX-size2-R {font-family: MJXc-TeX-size2-R,MJXc-TeX-size2-Rw}
.MJXc-TeX-size3-R {font-family: MJXc-TeX-size3-R,MJXc-TeX-size3-Rw}
.MJXc-TeX-size4-R {font-family: MJXc-TeX-size4-R,MJXc-TeX-size4-Rw}
.MJXc-TeX-vec-R {font-family: MJXc-TeX-vec-R,MJXc-TeX-vec-Rw}
.MJXc-TeX-vec-B {font-family: MJXc-TeX-vec-B,MJXc-TeX-vec-Bx,MJXc-TeX-vec-Bw}
@font-face {font-family: MJXc-TeX-ams-R; src: local('MathJax_AMS'), local('MathJax_AMS-Regular')}
@font-face {font-family: MJXc-TeX-ams-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_AMS-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_AMS-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_AMS-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-cal-B; src: local('MathJax_Caligraphic Bold'), local('MathJax_Caligraphic-Bold')}
@font-face {font-family: MJXc-TeX-cal-Bx; src: local('MathJax_Caligraphic'); font-weight: bold}
@font-face {font-family: MJXc-TeX-cal-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-frak-R; src: local('MathJax_Fraktur'), local('MathJax_Fraktur-Regular')}
@font-face {font-family: MJXc-TeX-frak-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-frak-B; src: local('MathJax_Fraktur Bold'), local('MathJax_Fraktur-Bold')}
@font-face {font-family: MJXc-TeX-frak-Bx; src: local('MathJax_Fraktur'); font-weight: bold}
@font-face {font-family: MJXc-TeX-frak-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Fraktur-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Fraktur-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Fraktur-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-math-BI; src: local('MathJax_Math BoldItalic'), local('MathJax_Math-BoldItalic')}
@font-face {font-family: MJXc-TeX-math-BIx; src: local('MathJax_Math'); font-weight: bold; font-style: italic}
@font-face {font-family: MJXc-TeX-math-BIw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-BoldItalic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-BoldItalic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-BoldItalic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-R; src: local('MathJax_SansSerif'), local('MathJax_SansSerif-Regular')}
@font-face {font-family: MJXc-TeX-sans-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-B; src: local('MathJax_SansSerif Bold'), local('MathJax_SansSerif-Bold')}
@font-face {font-family: MJXc-TeX-sans-Bx; src: local('MathJax_SansSerif'); font-weight: bold}
@font-face {font-family: MJXc-TeX-sans-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-sans-I; src: local('MathJax_SansSerif Italic'), local('MathJax_SansSerif-Italic')}
@font-face {font-family: MJXc-TeX-sans-Ix; src: local('MathJax_SansSerif'); font-style: italic}
@font-face {font-family: MJXc-TeX-sans-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_SansSerif-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_SansSerif-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_SansSerif-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-script-R; src: local('MathJax_Script'), local('MathJax_Script-Regular')}
@font-face {font-family: MJXc-TeX-script-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Script-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Script-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Script-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-type-R; src: local('MathJax_Typewriter'), local('MathJax_Typewriter-Regular')}
@font-face {font-family: MJXc-TeX-type-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Typewriter-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Typewriter-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Typewriter-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-cal-R; src: local('MathJax_Caligraphic'), local('MathJax_Caligraphic-Regular')}
@font-face {font-family: MJXc-TeX-cal-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Caligraphic-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Caligraphic-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Caligraphic-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-B; src: local('MathJax_Main Bold'), local('MathJax_Main-Bold')}
@font-face {font-family: MJXc-TeX-main-Bx; src: local('MathJax_Main'); font-weight: bold}
@font-face {font-family: MJXc-TeX-main-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Bold.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-I; src: local('MathJax_Main Italic'), local('MathJax_Main-Italic')}
@font-face {font-family: MJXc-TeX-main-Ix; src: local('MathJax_Main'); font-style: italic}
@font-face {font-family: MJXc-TeX-main-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-main-R; src: local('MathJax_Main'), local('MathJax_Main-Regular')}
@font-face {font-family: MJXc-TeX-main-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Main-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Main-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Main-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-math-I; src: local('MathJax_Math Italic'), local('MathJax_Math-Italic')}
@font-face {font-family: MJXc-TeX-math-Ix; src: local('MathJax_Math'); font-style: italic}
@font-face {font-family: MJXc-TeX-math-Iw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Math-Italic.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Math-Italic.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Math-Italic.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size1-R; src: local('MathJax_Size1'), local('MathJax_Size1-Regular')}
@font-face {font-family: MJXc-TeX-size1-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size1-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size1-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size1-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size2-R; src: local('MathJax_Size2'), local('MathJax_Size2-Regular')}
@font-face {font-family: MJXc-TeX-size2-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size2-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size2-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size2-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size3-R; src: local('MathJax_Size3'), local('MathJax_Size3-Regular')}
@font-face {font-family: MJXc-TeX-size3-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size3-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size3-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size3-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-size4-R; src: local('MathJax_Size4'), local('MathJax_Size4-Regular')}
@font-face {font-family: MJXc-TeX-size4-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Size4-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Size4-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Size4-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-vec-R; src: local('MathJax_Vector'), local('MathJax_Vector-Regular')}
@font-face {font-family: MJXc-TeX-vec-Rw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Regular.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Regular.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Regular.otf') format('opentype')}
@font-face {font-family: MJXc-TeX-vec-B; src: local('MathJax_Vector Bold'), local('MathJax_Vector-Bold')}
@font-face {font-family: MJXc-TeX-vec-Bx; src: local('MathJax_Vector'); font-weight: bold}
@font-face {font-family: MJXc-TeX-vec-Bw; src /*1*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/eot/MathJax_Vector-Bold.eot'); src /*2*/: url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/woff/MathJax_Vector-Bold.woff') format('woff'), url('https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/fonts/HTML-CSS/TeX/otf/MathJax_Vector-Bold.otf') format('opentype')}
</style><blockquote>This might seem surprising at first, because there is also a different incomplete model Φ<span class="mathjax-inline-container"><span class="mjx-chtml"><span class="mjx-math" aria-label="_2"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char"></span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">2</span></span></span></span></span></span></span></span> that says “if you pays the blackmail, infestation will not happen”. Φ<span class="mathjax-inline-container"><span class="mjx-chtml"><span class="mjx-math" aria-label="_2"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char"></span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">2</span></span></span></span></span></span></span></span> is false if you use physical causal counterfactuals, but from the agent’s perspective Φ<span class="mathjax-inline-container"><span class="mjx-chtml"><span class="mjx-math" aria-label="_2"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char"></span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">2</span></span></span></span></span></span></span></span> is consistent with all observations. However, Φ<span class="mathjax-inline-container"><span class="mjx-chtml"><span class="mjx-math" aria-label="_2"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char"></span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">2</span></span></span></span></span></span></span></span> only guarantees the payoff −c (because it is unknown whether the blackmail will arrive). Therefore, Φ<span class="mathjax-inline-container"><span class="mjx-chtml"><span class="mjx-math" aria-label="_2"><span class="mjx-mrow" aria-hidden="true"><span class="mjx-msubsup"><span class="mjx-base"><span class="mjx-mi"><span class="mjx-char"></span></span></span><span class="mjx-sub" style="font-size: 70.7%; vertical-align: -0.212em; padding-right: 0.071em;"><span class="mjx-mn" style=""><span class="mjx-char MJXc-TeX-main-R" style="padding-top: 0.372em; padding-bottom: 0.372em;">2</span></span></span></span></span></span></span></span> will have no effect on the ultimate behavior of the agent.</blockquote><p>What happens in ASP? (Say you’re in an iterated Newcomb’s problem with a predictor much slower than you, but which meets the LIC or similar.) I’m concerned that it will either settle on two-boxing, or possibly not settle on one strategy, since if it settles on two-boxing then a model which says “you can get the higher reward by one-boxing” (ie, the agent has control over the predictor) looks appealing; but, if it settles on one-boxing, a model which says “you can get higher reward by two-boxing” (ie, the agent’s action <em>doesn’t</em> control the predictor) looks appealing. This concern is related to the way asymptotic decision theory <a href="/posts/pgJbaXvYWBx3Mrg5T/when-edt-cdt-adt-does-well">fails</a>—granted, for cases outside of its definition of “fair”.</p><blockquote>The precommitments have to expire after some finite time.</blockquote><p>I agree that something like this generally does the right thing in <a href="/posts/5bd75cc58225bf067037550c/policy-selection-solves-most-problems">most cases</a>, with the exception of <a href="/posts/9sYzoRnmqmxZm4Whf/conceptual-problems-with-udt-and-policy-selection">superrationality in games</a> as a result of <a href="/posts/brXr7PJ2W4Na2EW2q/the-commitment-races-problem">commitment races</a>.</p><p>I still have a little hope that there will be a nice version, which doesn’t involve a commitment-races problem and which doesn’t make use of an arbitrary commitment cutoff. But I would agree that things don’t look good, and so it is reasonable to put this kind of thing outside of “fair” problems.</p><blockquote>Let me add that I am not even sure what are the correct desiderata. In particular, I don’t think that we should expect any group of good agents to converge to a Pareto optimal outcome. </blockquote><p>I don’t currently see why we shouldn’t ask to converge to pareto optima. Obviously, we can’t expect to do so with arbitrary other agents; but it doesn’t seem unreasonable to use an algorithm which has the property of reaching pareto-optima with other agents who use that same algorithm. This even seems reasonable in the standard iterated Nash picture (where not all strategies achieve pareto optima, but there exist strategies which achieve pareto optima with a broad-ish class of other strategies, including others who use strategies like their own—while being very difficult to exploit).</p><p>But yeah, I’m pretty uncertain about what the desiderata should be—both with respect to game theory, and with respect to scenarios which require updatelessness/precommitments in order to do well. I agree that it should all be approached with a learning-theoretic perspective.</p>abramdemskiHRJzddD6AYMaSq6bDFri, 13 Sep 2019 01:30:36 +0000Comment by abramdemski on Formalising decision theory is hard
https://www.greaterwrong.com/posts/S3W4Xrmp6AL7nxRHd/formalising-decision-theory-is-hard#comment-MYGR6xB5KbLDbSWnG
<p>Ahh thanks :p fixed</p>abramdemskiMYGR6xB5KbLDbSWnGThu, 12 Sep 2019 20:00:15 +0000Comment by abramdemski on Counterfactuals are an Answer, Not a Question
https://www.greaterwrong.com/posts/ao7KLoBEvMdHFjrNZ/counterfactuals-are-an-answer-not-a-question#comment-6msAmGSx5FJecSHmz
<p>Those of a Bayesian leaning will tend to say things like “probability is subjective”, and claim this is an important insight into the nature of probability—one might even go so far as to say “probability is an answer, not a question”. But this doesn’t mean you can believe what you want; not exactly. There are coherence constraints. So, once we see that probability is subjective, we can then seek a theory of the subjectivity, which tells us “objective” information about it (yet which leaves a whole lot of flexibility).</p><p>The same might be true of counterfactuals. I personally lean toward the position that the constraints on counterfactuals are <a href="/posts/XW6Qi2LitMDb2MF8c/a-rationality-condition-for-cdt-is-that-it-equal-edt-part-1">that they be consistent with evidential predictions</a>, but I don’t claim to be unconfused. My position is a “counterfactuals are subjective but have significant coherence constraints” type position, but (arguably) a fairly minimal one—the constraint is a version of “counterfacting on what you actually did should yield what actually happened”, one of the most basic constraints on what counterfactuals should be.</p><p>On the other hand, my theory of counterfactuals is pretty boring and doesn’t directly solve problems—it more says “look elsewhere for the interesting stuff”.</p><p>Edit --</p><p>Oh, also, I wanted to pitch the idea that counterfactuals, like a whole bunch of things, should be thought of as “constructed rather than real”. This is subtly different from “subjective”. We humans are pretty far along in an ongoing process of figuring out how to be and act in the world. Sometimes we come up with formal theories of things like probability, utility, counterfactuals, and logic. The process of coming up with these formal theories informs our practice. Our practice also informs the formal theories. Sometimes a theory seems to capture what we wanted really nicely. My argument is that in an important sense we’ve invented, not discovered, what we wanted.</p><p>So, for example, utility functions. Do utility functions capture human preferences? No, not really, they are pretty far from preferences observed in the wild. However, we’re in the process of figuring out what we prefer. Utility functions capture some nice ideas about idealized preferences, so that when we’re talking about idealized versions of what we want (trying to figure out what we prefer upon reflection) it is (a) often pretty convenient to think in terms of utilities, and (b) somewhat difficult to really escape the framework of utilities. Similarly for probability and logic as formal models of idealized reasoning.</p><p>So, just as utility functions aren’t really out there in the world, counterfactuals aren’t really out there in the world. But just as it might be that we should think about our preferences in terms of utility anyway (...or maybe abandon utility in favor of better theoretical tools), we might want to equip our best world-model with counterfactuals anyway (...or abandon them in favor of better theoretical tools).</p>abramdemski6msAmGSx5FJecSHmzSat, 07 Sep 2019 05:17:41 +0000Comment by abramdemski on Formalising decision theory is hard
https://www.greaterwrong.com/posts/S3W4Xrmp6AL7nxRHd/formalising-decision-theory-is-hard#comment-7vTvvzb9a7vjPKGhh
<p>I very much agree with the point about not decoupling learning and decision theory. I <a href="/posts/hLFD6qSN9MmQxKjG5/embedded-agency-via-abstraction#comment-jbC8934HoXaoXrqhR">wrote a comment</a> making somewhat similar points.</p><blockquote>I believe that this indeed solves both INP and IXB.</blockquote><p>I’d like to understand this part.</p><blockquote>One way to fix it is by allowing the agent to precommit. Then the assumption about Omega becomes empirically verifiable.</blockquote><p>I’m not sure I should find the precommitment solution satisfying. Won’t it make some stupid precommitments early (before it has learned enough about the world to make reasonable precommitments) and screw itself up forever? Is there a generally applicable version of precommitments which ensures learning good behavior?</p><blockquote>The only class of problems that I’m genuinely unsure how to deal with is game-theoretic superrationality. </blockquote><p>If we take the learning-theoretic view, then we get to bring in tools from iterated games. There’s a <a href="/posts/XTgkhjNTEi97WHMi6/pavlov-generalizes">Pavlov-like strategy</a> for playing deterministic iterated games which converges to optimal responses to non-agentic environments and converges to Pareto optima for environments containing agents who use the Pavlov-like strategy. It is not the greatest at being unexploitable, and it also has fairly bad convergence.</p><p>However, I don’t yet see how to translate the result to logical-induction type learners. Besides requiring deterministic payouts (a property which can probably be relaxed somehow), the algorithm requires an agent to have a definite history—a well-defined training sequence. Agents based on logical induction are instead forming generalizations based on any sufficiently analogous situation within logic, so they don’t have a well-defined history in the right way. (An actual instance of a logical induction agent has an actual temporal history, but this temporal history is not necessarily what it is drawing on to play the game—it may have never personally encountered a similar situation.)</p><p>In other words, I’m hopeful that there could be a learning-theoretic solution, but I don’t know what it is yet.</p><p>As for superrationality for agents w/o learning theory, there’s cooperative oracles, right? We can make computable analogues with <a href="/posts/5bd75cc58225bf067037554e/distributed-cooperation">distributed oracles</a>. It’s not a real solution, <em>specifically in that</em> it ignores learning. So I sort of think we know how to do it in the “static” setting, but the problem is that we live in a learning-theoretic setting rather than a static-rationality setting.</p>abramdemski7vTvvzb9a7vjPKGhhThu, 05 Sep 2019 21:56:00 +0000Comment by abramdemski on Embedded Agency via Abstraction
https://www.greaterwrong.com/posts/hLFD6qSN9MmQxKjG5/embedded-agency-via-abstraction#comment-jbC8934HoXaoXrqhR
<blockquote>The dice example is one I stumbled on while playing with the idea of a probability-like calculus for excluding information, rather than including information. I’ll write up a post on it at some point.</blockquote><p>I look forward to it.</p><blockquote>When I imagine an embedded agent, I imagine some giant computational circuit representing the universe, and I draw a box around one finite piece of it</blockquote><p>Speaking very abstractly, I think this gets at my actual claim. Continuing to speak at that high level of abstraction, I am claiming that you should imagine an agent more as a flow through a fluid.</p><p>Speaking much more concretely, this difference comes partly from the question of whether to consider robust delegation as a central part to tackle now, or (as you suggested in the post) a part to tackle later. I agree with your description of robust delegation as “hard mode”, but nonetheless consider it to be central.</p><p>To name some considerations:</p><ul><li><p>The “static” way of thinking involves handing decision problems to agents without asking how the agent found itself in that situation. The how-did-we-get-here question is sometimes important. For example, my <a href="/posts/5bd75cc58225bf0670375452/smoking-lesion-steelman">rejection of the standard smoking lesion problem</a> is a how-did-we-get-here type objection.</p></li><li><p>Moreover, “static” decision theory puts a box around “epistemics” with an output to decision-making. This implicitly suggests: “Decision theory is about optimal action under uncertainty—the generation of that uncertainty is relegated to epistemics.” This ignores the role of learning how to act. Learning how to act can be critical even for decision theory in the abstract (and is obviously important to implementation).</p></li><li><p>Viewing things from a learning-theoretic perspective, it doesn’t generally make sense to view a single thing (a single observation, a single action/decision, etc) in isolation. So, accounting for logical non-omniscience, we can’t expect to make a single decision “correctly” for basically any notion of “correctly”. What we can expect is to be “moving in the right direction”—not at a particular time, but generally over time (if nothing kills us).<ul><li><p>So, describing an embedded agent in some particular situation, the notion of “rational (bounded) agency” should not expect anything optimal about its actions in that circumstance—it can only talk about the way the agent updates.</p></li><li><p>Due to logical non-omniscience, this applies to the action even if the agent is at the point where it knows what’s going on epistemically—it might not have learned to appropriately react to the given situation yet. So even “reacting optimally given your (epistemic) uncertainty” isn’t realistic as an expectation for bounded agents.</p></li></ul></p></li><li><p>Obviously I also think the “dynamic” view is better in the purely epistemic case as well—logical induction being the poster boy, totally breaking the static rules of probability theory at a fixed time but gradually improving its beliefs over time (in a way which approaches the static probabilistic laws but also captures more).<ul><li><p>Even for purely Bayesian learning, though, the dynamic view is a good one. Bayesian learning is a way of setting up dynamics such that better hypotheses “rise to the top” over time. It is quite analogous to replicator dynamics as a model of evolution.</p></li></ul></p></li><li><p>You can do “equilibrium analysis” of evolution, too (ie, evolutionary stable equilibria), but it misses how-did-we-get-here type questions: larger and smaller attractor basins. (Evolutionarily stable equilibria are sort of a patch on Nash equilibria to address some of the how-did-we-get-here questions, by ruling out points which are Nash equilibria but which would not be attractors at all.) It also misses out on orbits and other fundamentally dynamic behavior. <ul><li><p>(The dynamic phenomena such as orbits become important in the theory of correlated equilibria, if you get into the literature on <em>learning</em> correlated equilibria (MAL—multi-agent learning) and think about where the correlations come from.)</p></li></ul></p></li></ul><blockquote>Of course we <em>could </em>have agents which persist over time, collecting information and making multiple decisions, but if our theory of embedded agency <em>assumes</em> that, then it seems like it will miss a lot of agenty behavior.</blockquote><p>I agree that requiring dynamics would miss some examples of actual single-shot agents, doing something intelligently, once, in isolation. However, it is a live question for me whether such agents can be anything else that Boltzmann brains. In <a href="/posts/osxNg6yBCJ4ur9hpi/does-agent-like-behavior-imply-agent-like-architecture">Does Agent-like Behavior imply Agent-like Architecture</a>, Scott mentioned that it seems quite unlikely that you could get a look-up table which behaves like an agent without having an actual agent somewhere causally upstream of it. Similarly, I’m suggesting that it seems unlikely you could get an agent-like architecture sitting in the universe without some kind of learning process causally upstream. </p><p>Moreover, continuity is central to the major problems and partial solutions in embedded agency. X-risk is a robust delegation failure more than a decision-theory failure or an embedded world-model failure (though subsystem alignment has a similarly strong claim). UDT and TDT are interesting largely because of the way they establish dynamic consistency of an agent across time, partially addressing the tiling agent problem. (For UDT, this is especially central.) But, both of them <a href="/posts/9sYzoRnmqmxZm4Whf/conceptual-problems-with-udt-and-policy-selection">ultimately fail very much because of their “static” nature</a>.</p><p>[I actually got this static/dynamic picture from <a href="/users/komponisto">komponisto</a> btw (talking in person, though the posts give a taste of it). At first it sounded like rather free-flowing abstraction, but it kept surprising me by being able to bear weight. Line-per-line, though, much more of the above is inspired by discussions with Steve Rayhawk.]</p><p>Edit: Vanessa made a related point <a href="/posts/S3W4Xrmp6AL7nxRHd/formalising-decision-theory-is-hard#comment-Xiz3TAe2Z9Yh49S9X">in a comment on another post</a>.</p>abramdemskijbC8934HoXaoXrqhRWed, 04 Sep 2019 05:52:43 +0000Comment by abramdemski on Deconfuse Yourself about Agency: Architecture-morphization
https://www.greaterwrong.com/posts/ZigRhB4pAGdr6beQh/deconfuse-yourself-about-agency-architecture-morphization#comment-98G3hTcNdG3QkNC3B
<blockquote>I think of agent-like architectures as something objective, or related to the territory. In contrast, agent-like behavior is something subjective, something in the map. Importantly, <em>agent-like behavior</em>, or the lack of it, of some X is something that <em>exists in the map of some entity</em> Y (where often Y≠X).</blockquote><blockquote>The selection/control distinction seems related, but not quite similar to me. Am I missing something there?</blockquote><p>A(Θ)-morphism seems to me to involve both agent-like architecture and agent-like behavior, because it just talks about prediction generally. Mostly I was asking if you were trying to point it one way or the other (we could talk about prediction-of-internals exclusively, to point at structure, or prediction-of-external exclusively, to talk about behavior—I was unsure whether you were trying to do one of those things).</p><p>Since you say that you are trying to formalize how we informally talk, rather than how we <em>should</em>, I guess you weren’t trying to make A(Θ)-morphism get at this distinction at all, and were separately mentioning the distinction as one which should be made.</p>abramdemski98G3hTcNdG3QkNC3BWed, 04 Sep 2019 04:23:46 +0000Comment by abramdemski on Troll Bridge
https://www.greaterwrong.com/posts/hpAbfXtqYC2BrpeiC/troll-bridge-5#comment-u9AGW6zm9QrFkYjDD
<blockquote>I don’t see how this agent seems to control his sanity.</blockquote><p>The agent in Troll Bridge thinks that it can make itself insane by crossing the bridge. (Maybe this doesn’t answer your question?)</p><blockquote>Troll Bridge is a rare case where agents that require proof to take action can prove they would be insane to take some action before they’ve thought through its consequences. Can you show how they could unwisely do this in chess, or some sort of Troll Chess?</blockquote><p>I make no claim that this sort of case is common. Scenarios where it comes up and is relevant to X-risk might involve alien superintelligences trolling human-made AGI. But it isn’t exactly high on my list of concerns. The question is more about whether particular theories of counterfactual are right. Troll Bridge might be “too hard” in some sense—we may just have to give up on it. But, generally, these weird philosophical counterexamples are more about pointing out problems. Complex real-life situations are difficult to deal with (in terms of reasoning about what a particular theory of counterfactuals will actually do), so we check simple examples, even if they’re outlandish, to get a better idea of what the counterfactuals are doing in general.</p>abramdemskiu9AGW6zm9QrFkYjDDWed, 04 Sep 2019 04:15:12 +0000Comment by abramdemski on Troll Bridge
https://www.greaterwrong.com/posts/hpAbfXtqYC2BrpeiC/troll-bridge-5#comment-HuoLdKKBQ6DKkxyJW
<p>Yep, sorry. The illustrations were not actually originally meant for publication; they’re from my personal notes. I did it this way (1) because the pictures are kind of nice, (2) because I was frustrated that no one had written a good summary post on Troll Bridge yet, (3) because I was in a hurry. Ideally I’ll edit the images to be more suitable for the post, although adding the omitted content is a higher priority.</p>abramdemskiHuoLdKKBQ6DKkxyJWWed, 04 Sep 2019 04:05:21 +0000