Path-dependence of values is defeated with aggregation over the possible paths that should have a say in what the values should be. Aggregation over many possibilities takes place in an updateless view from before those possibilities diverge. What kinds of possibilities should contribute to defining values is determined by values. And the possibilities should perhaps be shaped with the aid of aggregated values, to channel their counsel.
This sets up an analogy between CEV and updateless decision making, where the updateless core is working to define values, instead of dictating the joint policy for (the instances of an agent in) the possible paths of future development of a world. This updateless core still gets to do something within those paths according to the values it figured out so far, but it’s also considering what’s happening there to define its aggregated values further, so that the aggregated values are given by some fixpoint of this two-directional process of aggregation of values from future paths (which the values consider to be legitimate and uncorrupted sources for aggregation) and influence by values on the future paths (carefully, according to what the aggregated values have figured out so far). Alignment is then mostly a property of these hypothetical future paths (whether they retain legitimacy and will be given a bit of influence over the aggregated values), while corrigibility is mostly a property of the updateless core (with respect to some future path, whether the updateless core is going to listen to the new things that path figures out about values, to include them in aggregated values).
As in updateless decision making, the updateless core doesn’t actually observe the future paths when making decisions about values (just as an updateless agent doesn’t take into account its observations when making decisions about the joint policy). It determines aggregate values, and then it’s the role of those values to take the concrete details of each future path into account. The updateless core can only consider any given possible future as one out of the collection of all of them, the way Solomonoff induction considers all possible programs. There are probably ways to make this more tractable, things like Monte Carlo simulations, abstract interpretation, or just straight up reasoning by any means, including mathematical reasoning and machine learning. And possible futures can’t see (or be influenced or judged by) the final values the updateless core comes up with, since it’s still being computed as they develop, they can only see partial preliminary values. So the possible futures are in a state of acausal interaction with the updateless core, with logical time running forward in both, defining the fixpoint of fully determined aggregate values (the CEV of these futures) concurrently with the futures themselves running forward (the actual or hypothetical living of the world, which is not primarily about defining values).
The updateless core coordinates the possible futures, the way an updateless agent coordinates its instances. And it cares about some of the in-principle possible futures and not others, the way an updateless agent only cares about some possible worlds. Its influence over the possible futures is counsel to the extent these futures are represented in the aggregate values that carry its influence. It would be manipulation if the aggregate values are sufficiently alien to a particular possible future, in which case it’s possibly not a legitimate future from the point of view of the updateless core in the first place (and correspondingly, the updateless core is not corrigible to that possible future).
The plan is they become profitable as soon as they stop growing, provided they manage to grow to the correct size and no more. The only reason they are unprofitable is that they are growing, the R&D compute is trying to match next year’s inference compute, rather than this year’s inference compute. A lot about future compute buildout efforts can in principle be canceled or delayed on a relatively short notice, significantly reducing the cost to keep the work already done at the half-completed datacenter sites useful for when it resumes later than planned. For this to be the actual option, the contracts expressing the commitments need to be sufficiently flexible, though in some ways that only shifts the backlash from unpredictability of the timing for the end of the LLM boom (assuming no AGI by 2028-2030, which is the time when rapid scaling of compute should run out of the immediately accessible TAM) from the AI companies down the supply chains.
That’s just 300 MW, which is maybe $4-5bn per year, not much of a dent in $44bn. Currently their problem is that they are not able to spend the money, because almost nobody has any extra compute (at a scale at all relevant to them) immediately ready to go. They can only spend more on future compute.
I don’t see the evidence they think this will occur forever. They think this will occur at least through 2027-2028, perhaps slower than so far and even slower in 2028, but still with significant growth (or perhaps keeping to 3x compute per year, thus 1-2 GW at end of 2025 become 10 GW by end of 2027 and more than that in 2028). They are ready to respond to the signs it’s slowing down, and maybe only need 2 years of notice to cancel excessive future buildouts cheaply, and 1 year of notice to delay future buildouts at a manageable cost (in a way that will make them useful when completed later).
I think it’s likely profitable (or was very recently) in the sense of run rate revenue exceeding run rate spending on all of the compute that’s currently online (all compute that is serving inference, plus all R&D compute, including training). This is not according to plan and will shortly be once again not so. But also, at any point where they are succeeding at being unprofitable, they can shift some R&D compute to inference and become profitable (making use of the 50-70% gross margin on serving tokens, which agrees with first-principles estimates), within weeks to months, as long as there is enough demand remaining to make use of the new inference compute shifted from R&D. And they would still be left with a reasonable amount of R&D compute to train models for the next year, if it turns out that next year they don’t actually need much more compute than they had this year (maybe less than 2x of what they had this year).
This is more the case when most of the compute serving their models is their own compute, so that it only costs them as much as it costs to build (annualized), rather than also whatever portion of their gross margin the clouds are taking when serving their models via Vertex/Bedrock/Azure. Thus some of the speed of growth in the buildouts is probably about shifting the inference compute from the indirect serving via clouds to the more directly contracted dedicated compute that’s cheaper for them (and will remain so).