I think this analysis underestimates just how much compute OA and especially Anthropic currently have to spend on inference. Once they move to more efficient B200/300/Rubin systems, I expect a lot of compute to be freed up and for progress to accelerate.
I think this analysis underestimates just how much compute OA and especially Anthropic currently have to spend on inference. Once they move to more efficient B200/300/Rubin systems, I expect a lot of compute to be freed up and for progress to accelerate.
I think the compute they spend on inference will also just get scaled up over time.