But has it really failed its objective? It’s still producing text.
I think it’s also worth asking “but did it really figure out that the words were spelled backwards?” I think a reasonable case could be made that the tokens it’s outputting here come from the very small subset of reversed words in its training set, and it’s ordering them in a way that it thinks is sensical given how little training time was spent on it.
If you give GPT-3 a bunch of examples and teach it about words spelled backwards, does it improve? How much does it improve, how quickly?
But has it really failed its objective? It’s still producing text.
I think it’s also worth asking “but did it really figure out that the words were spelled backwards?” I think a reasonable case could be made that the tokens it’s outputting here come from the very small subset of reversed words in its training set, and it’s ordering them in a way that it thinks is sensical given how little training time was spent on it.
If you give GPT-3 a bunch of examples and teach it about words spelled backwards, does it improve? How much does it improve, how quickly?