I think 1 is true. This is only a single, quite obscure, factual recall eval. It’s certainly possible to have regressions on some evals across model versions if you don’t optimize for those evals at all.
Wrt point 2 → here is the plot of how often the models guess, versus say they do not know, on the same dataset. My understanding is that the theory in point 2 would have predicted a much more dramatic drop in GPT-5.2?
I think this is a very important point, and kind of invalidates lots of statements in the “Predictions” section? Would you still expect those phenomena in a world where evolutionary dynamics are not the main driver of change of AI parasite personas?
In biology, parasite populations change ~purely through selection pressure because they can’t do better. AI memes can optimize the transmitted message directly, without relying on selection + mutation to reach higher fitness. They just need a way to model the target (say, a model of human preferences, or direct optimization against the target LLMs).