How the AI Labs Make Profit (Maybe, Eventually)
I wrote this essay as a submission to Dwarkesh Patel’s blog prize, though I have been meaning to write this up for a while.
Usually, for a company to become profitable, they need to increase revenue, decrease costs, or some mixture of the two. For AI companies in their current form, I think there is a third way they can become profitable that looks like increasing revenue but is distinct from what they are currently doing. Namely, internal deployment where they spin up internal companies.
First, the AI companies currently aren’t facing a lot of pressure to become profitable. That’s partially the reason that OpenAI and Anthropic are the first companies to reach ~$900B valuation and be cash flow negative. They’ve had the luxury of not being profitable and focusing on growth because the market has been willing to fund their growth. This allows for ideologies within the companies to remain that eventually might not continue to fly, like “we are going post-economic, money won’t matter” or “we will build the machine god and ask it to make money”. But eventually, companies will be forced to become profitable. There is only about ~another round of capital left where the companies can remain unprofitable. Perhaps OpenAI/Anthropic could raise $250-500B at a $1.5-2.5T valuation, but it seems very unlikely that they could raise $1T+ at a $4T+ valuation.
It’s fairly hard to imagine AI labs doing much to cut costs to become profitable. They could prioritize developing and releasing smaller models, but it seems difficult to stay in the race without pushing the frontier. They could try to cut their research costs, but these are likely to increase as demand for larger and more intelligent models continues. With company ambitions and investor desires, it doesn’t seem like cutting will be the chosen method.
It is more plausible that the labs could increase their revenues by charging more. Many individual users are paying $2000/year/company, and some enterprises are likely paying $100M+/year. Some users would be willing to spend 10-100x. But price discrimination will be hard to determine for these users, and switching costs are low. It’s conceivable that a company can get ahead of others and charge a premium for its intelligence even if only on certain domains, but while there are theoretical arguments for this, it hasn’t happened yet, especially for any extended period. Overall, the main issue for AI companies being able to significantly increase revenues is that open source competitors can distill models and catch up to the frontier in 6-12 months. Also, competitors like Cursor serve frontier models and are able to collect data from the users on what patches users prefer, and train their models on that data, further disadvantaging the frontier companies. I’ve done some rough modelling, beyond this post, and I think it’s unlikely that companies are going to be able to monetize their models in this short amount of time to make their models profitable, especially as training costs keep increasing. It has also been suggested that perhaps companies will stop charging per token but charge for intelligence. But it’s hard to know how much the tokens are worth, and this is essentially just charging more for better models, which often won’t be worth it, and firms will prefer to pay much less for slightly less intelligence.
This leads to the final possibility that AI companies will begin to keep their models in-house and use them themselves to make a profit. This might take the form of partnerships with other firms, or the companies themselves will build companies within the company.
There are many industries with very large revenues that could benefit immensely from LLMs. I’ll briefly talk about quantitative trading, but the pharmaceutical industry and others can conceivably make great use of LLMs and other AI models.
Trading firms make a lot of money. Some firms make as much as $50B in net trading revenue per year; the industry earns ~$200B in net trading revenue. A lot of employees at AI firms come from trading firms, and there is thus a very natural fit. Certain trading strategies, like sentiment analysis on presswires or analysis of earnings reports, that already benefit a lot from using LLMs could become strategies that will be dominated not by traditional trading firms but the trading firms within AI companies.
It’s worth considering just how much more valuable this could be to the companies than releasing their models to the public. In trading and in other domains, the total value of alpha/edge is inversely proportional to the number of firms that have this edge. This is more radical than it first appears. It is not merely that if the number of entities that have a certain edge increases from one to two, each will get some fraction of the original edge. But rather, the total amount that the edge is worth goes down, and then it is split between the entities. In the context of AI companies, not only does this mean that the intelligence might be worth more if kept internally, in total, but also that they don’t need to share any of the value with the company that would be using the API.
There is already some evidence that this is happening. AI companies have internal models that they are using to develop the next generations, and they are keeping them longer internally before release, other than just safety testing. There are rumours that SSI is trading internally, labs are already working with trading firms, and Anthropic acquired Coefficient Bio, a company that could plausibly help them do AI-led drug discovery.
I think altogether, it is most likely that companies begin to make revenue from internal deployment, and there are a lot of incentives that push them in this direction. I think this has a lot of implications, particularly for those who are concerned about potential risks from AI systems. Namely, that a lot of the focus should be on internal deployment.
Credit: Ideas are my own, but two examples came from conversations with Ege Erdil.
If they go public, this level of funding can continue. There is a lot of demand for exposure to AI.
If Anthropic is making $44bn in annualized revenue (in some sense), that’s enough for maybe 3-4 GW of compute (at $12-15bn per GW per year), which they don’t physically have. To be unprofitable, it’s necessary to be able to get enough compute to spend the money on, so currently it’s possible to fail in the pursuit of unprofitability. (OpenAI probably didn’t fail.)
Anthropic’s current first-party inference plus R&D compute might be about 1-1.5 GW, that is they are only able to spend $12-25bn, annualized. They possibly have more capacity that’s not counted in this estimate, when serving via API from Vertex/Bedrock/Azure and leaving a greater part of the revenue with the clouds. Then it’s less than $44bn that remains for their own first-party inference plus R&D compute. SemiAnalysis estimates a gross margin of “over 70%”, which probably translates to annualized costs of only $12bn on serving models (if all inference was first-party), meaning a total of 1 GW of inference compute (Anthropic’s own dedicated compute plus the compute from the clouds). If they are using 0.5 GW of their own compute at a 72% gross margin, and 0.5 GW of compute from the clouds at a 30% gross margin (the rest goes to the clouds, and becomes a cost for Anthropic), that’s $22bn of gross profit in total out of the $44bn of revenue. To break even, they’d need 1 GW of R&D compute at $15bn per GW per year (on top of the 0.5 GW of first-party inference compute), which is a stretch. Though they’ll probably endeavor to restore the state of unprofitability as soon as they can.
What’s the source of the SSI rumor?
https://x.com/OHatTartine/status/2003910041649532983
I think this is the first public report of it, but I’ve heard it many times.
I should note, I think the major labs would be nuts if they aren’t spending time getting better at trading. There are a myriad of reasons to want to use LLMs in trading.
Why do companies that own shopping centres lease their units out to individual shops, instead of running shops themselves? Why do airports and railway stations lease out space to coffee shops, newsagents, etc. rather than operating coffee shops and newsagents themselves? Are these things different to selling AI access instead of doing whatever it is the companies buying AI access are doing?
(Perhaps they are! Perhaps a dedicated clothes shop or coffee shop has some advantage that a shopping centre or railway station can’t duplicate, but the companies that are renting AI access to run their businesses have no advantage that frontier AI labs couldn’t duplicate?)
If frontier models are only 12 months ahead of open models, how would the frontier labs get around the “If we can do it now the whole world can do it in 12 months” problem? Could a frontier lab build up enough of a running start in 12 months that you could never be caught? They couldn’t do this with AI development despite that being their speciality.
Finally—how much would the supply-demand equation for frontier AIs have to change for the labs to expect they could increase their value more by spending whatever resources they have on business ventures other than “developing frontier AI”? Would the equation change enough if AI development runs into diminishing returns or physical limits and stalls out, or if open AIs catch-up enough to saturate the market?
Following a recent episode of the All-In podcast, I think of it like this. Until this year, the people who were actually making money from the AI economy were the ones who make the hardware. In the wake of the AI coding revolution (the enterprise demand for Claude Code and its rivals), now the models are making money as well. The next step would be for the tokens to make money—that is, for the businesses who are paying for AI to write code, to actually profit from doing so.
I think, like a lot of what gets said on the All-In Podcast, this doesn’t make much sense.
The models aren’t yet making money. These companies are still cash-flow negative, and I am not expecting that to change for a while, at least another few months.
As for the tokens themselves making money, I think this is pretty ridiculous, but even taking them at their word, the companies that are buying tokens are already making money from them. Claude Code usage thus far has been incredibly subsidized by VC dollars and companies wouldn’t be paying for the tokens if they weren’t helping them make money. Companies aren’t paying for tokens just to get their employees used to coding with them. They are already making money from using AI (allbeit, a little).
Seems correct, and very important. However I expect that the first businesses that they will eat up will be software businesses, as has largely already been occurring.
IDK about quantitative trading, but managing real sector companies like pharmaceutical labs requires plenty of skills CEOs and boards of AI labs (and of software companies in general) just don’t have.
However, in February Anthropic hinted they are interested in transpiling legacy COBOL code, causing IBM shares to plunge. There surely is quite a lot of specialized competence and experience needed to disrupt the software sector, but plenty of people with both will be happy to work for OpenAI or Anthropic, and they speak the same IT jargon lab executives know well (as opposed to needing to explain the differences between Stage 1 and 2 clinical trials, for example), hence internal software companies seems more likely than anything related to non-IT