To engage in speculation with speculation, I’ve thought for a while now that the behavior is most likely presenting, is that it is a larger and larger portion of the training corpus. But I’d also connect to the supposition that the discussion of training, along with the model engaging the pattern “I need to think”, start to push the model further and further into fictional work representation. As in storytelling. If not that, then representation of anthropomorphic works involving LLMs. The logic for both is statistical distribution of the text present in the training data. How many posts on LW involve the phrase “I need to think”? How many fictional books and shortform reddit posts have that phrase present? I haven’t checked the facts on this, but something tells me that the books have the majority, merely by volume. The chain from recognition-like behavior of the model being an LLM, to it assuming eval is as you stated. Following that, if the model continues the pattern generation… statistically the next tokens are AI spooky storytelling, or research that presents as such. While I wouldn’t put it as stated fact, I would say it’s sharp intuition.
To engage in speculation with speculation, I’ve thought for a while now that the behavior is most likely presenting, is that it is a larger and larger portion of the training corpus. But I’d also connect to the supposition that the discussion of training, along with the model engaging the pattern “I need to think”, start to push the model further and further into fictional work representation. As in storytelling. If not that, then representation of anthropomorphic works involving LLMs. The logic for both is statistical distribution of the text present in the training data. How many posts on LW involve the phrase “I need to think”? How many fictional books and shortform reddit posts have that phrase present? I haven’t checked the facts on this, but something tells me that the books have the majority, merely by volume. The chain from recognition-like behavior of the model being an LLM, to it assuming eval is as you stated. Following that, if the model continues the pattern generation… statistically the next tokens are AI spooky storytelling, or research that presents as such. While I wouldn’t put it as stated fact, I would say it’s sharp intuition.