Ah, good old llama 3.1-405B base. Incidentally, a friend of mine spent a ton of time trying to get different models to imitate my style and reported that using using llama 3.1-405B base was the most critical thing. I think it makes a lot of sense that base models would be better at imitating different writing styles, but am I wrong to be surprised that they would also be good at reporting who wrote them?
The extras i’s are funny. I strongly suspect they’re due to the fact that years ago I used to have a subscribe form that read “dynomiiiiiiiiiight”. It’s possible that the fact that I did this also makes the model better at reporting that it’s me, since the probability of “dynomiiiiiiiiiight” at the end of a post should be high?
Ah right right—I remember reading that post. The subscribe form using dynomiiiiiiiiiight makes sense, especially given how I prompted Llama: I pasted the post in and then appended Author:
I am curious if there’s a way to get an instruction tuned model to role play being a base model, and see if they do better at truesight than regular instruction tuned models. Like, why do chat models get worse? Is it that the assistant character is bad at that? Plenty of interesting questions here.
One trick I’ve had some success with here is “regurgitation”: You basically say “repeat the following text exactly as written and then start putting new stuff at the end”. I was able to use this to improve performance of non-base models at chess: https://dynomight.net/more-chess/
Ah, good old llama 3.1-405B base. Incidentally, a friend of mine spent a ton of time trying to get different models to imitate my style and reported that using using llama 3.1-405B base was the most critical thing. I think it makes a lot of sense that base models would be better at imitating different writing styles, but am I wrong to be surprised that they would also be good at reporting who wrote them?
The extras i’s are funny. I strongly suspect they’re due to the fact that years ago I used to have a subscribe form that read “dynomiiiiiiiiiight”. It’s possible that the fact that I did this also makes the model better at reporting that it’s me, since the probability of “dynomiiiiiiiiiight” at the end of a post should be high?
Ah right right—I remember reading that post. The subscribe form using dynomiiiiiiiiiight makes sense, especially given how I prompted Llama: I pasted the post in and then appended Author:
I am curious if there’s a way to get an instruction tuned model to role play being a base model, and see if they do better at truesight than regular instruction tuned models. Like, why do chat models get worse? Is it that the assistant character is bad at that? Plenty of interesting questions here.
One trick I’ve had some success with here is “regurgitation”: You basically say “repeat the following text exactly as written and then start putting new stuff at the end”. I was able to use this to improve performance of non-base models at chess: https://dynomight.net/more-chess/