Sama Says the Age of Giant AI Models is Already Over

Link post

There wasn’t much content to this article, except for the below quotes.

Sam Altman says

further progress will not come from making models bigger. “I think we’re at the end of the era where it’s going to be these, like, giant, giant models,” he told an audience at an event held at MIT late last week. “We’ll make them better in other ways.” [...] Altman said there are also physical limits to how many data centers the company can build and how quickly it can build them. [...] At MIT last week, Altman confirmed that his company is not currently developing GPT-5. “An earlier version of the letter claimed OpenAI is training GPT-5 right now,” he said. “We are not, and won’t for some time.”

Nick Frosst, a cofounder at Cohere, says

“There are lots of ways of making transformers way, way better and more useful, and lots of them don’t involve adding parameters to the model,” he says. Frosst says that new AI model designs, or architectures, and further tuning based on human feedback are promising directions that many researchers are already exploring.

In the Lex Fridman interview with Sam Altman, Sam said that they had to do “hundreds of complicated things”. Does this, together with the above quote, suggest Sam thinks transformers are running out of oomph? Is he, perhaps, pausing progress whilst we await the next breakthrough in deep learning?

Edit: Added in a relevant question. Give your probability for the first option.

10%20%30%40%50%60%70%80%90%
Will the next LLM, which is to GPT-4 what GPT-4 is to GPT-3, be a transformer? Or will something else be needed to get there?

Or vote at this tweet, if you like:
https://​​twitter.com/​​Algon_33/​​status/​​1648047440065449993