Never Fight The Last War

Lieutenant Colonel J. L. Schley, of the Corps of Engineers wrote:

“There is a tendency in many armies to spend the peace time studying how to fight the last war”

After the Chinchilla paper, everyone got focused on the idea that large language models can be more effective.

The GPT3.5 we saw in ChatGPT was scary and got people to think about how we can prevent bigger models from being trained. There was a lot of capability gain when OpenAI released GPT4 but if we can trust Sam Altman, most of it was not because of an increase in model size but by doing a lot of different improvements.

From that perspective, many people think about whether or not GPT5 or GPT6 will be scary and dangerous by improving along the same axis as the GPT3.5 ->GPT4 improvement.

Right now, it seems that most of the danger is not in model improvement but in model utilization.

Palantir argues that models are commodities and released an AI product without providing any models that go beyond the existing open-source models. The value that Palantir does provide is that it provides interfaces of how a model can interact with the existing data a company has and how to create multi-step chains of data transformation. It provides a system to manage agents that use models to make decisions.

I read some people argue that Palantir’s product is worse than GPT4 and thus should not be seen as an advance in AI, but that’s flawed thinking. Autonomous agents are more dangerous than anything you can do in ChatGPT.