On the “prep for the model that is coming tomorrow not the model of today” front, I will say that LLMs are not always going to be as dumb as they are today.
Right, I strongly agree with this part.
their rate of learning still makes them in some sense your most promising mentee
I disagree in the sense that they’re no mentee of mine, ie, me trying to get today’s models to understand me doesn’t directly help tomorrow’s models to understand. (With the exception of the limited forms of feedback in the interface, like thumbs up/down, the impact of which I’m unsure of so it doesn’t feel like something I should deliberately spend a lot of time on.)
I also disagree in the sense that engaging with LLMs right now seems liable to produce a lot less fruits downstream, even as measured by “content that can usefully prompt an LLM later”. IE, if mentees are viewed as machines that convert time-spent-dialoging-with-me to text that is useful later, I don’t think LLMs are currently my most promising mentees.
So although I strongly agree with continuing to occasionally poke at LLMs to prep for the models that are coming soon & notice when things get better, to the extent that “most promising mentee” is supposed to imply that significant chunks of my time could be usefully spent with LLMs in the present, I disagree based on my (fairly extensive) experience.
trying to get as much of the tacit knowledge you have into their training data as possible (if you want them to be able to more easily & sooner build on your work).
Barring special relationships with frontier labs, this sounds functionally equivalent to trying to get my work out there for humans to understand, for now at least.
I did talk to Anthropic last year about the possibility of me providing detailed feedback on Claude’s responses (wrt my research questions), but it didn’t end up happening. The big problems I identified seemed to be things they thought would definitely get addressed in another way, so there wasn’t a mutually agreed-on value proposition (I didn’t understand what they hoped to gain, & they didn’t endorse the sorts of things I hoped to train). I got busy and moved on to other things.
Or (if you don’t want to do that for whatever reason) just generally not being caught flat-footed once they are smart enough to help you, as all your ideas are in videos or otherwise in high context understandable-only-to-abram notes.
I feel like this is speaking from a model I don’t understand. Are videos so bad? Video transcriptions are already a thing, and future models should be better at watching video and getting info from it. Are personal notes so bad? What sorts of actions are you recommending? I already want to write as many text posts as I can.
Right, I strongly agree with this part.
I disagree in the sense that they’re no mentee of mine, ie, me trying to get today’s models to understand me doesn’t directly help tomorrow’s models to understand. (With the exception of the limited forms of feedback in the interface, like thumbs up/down, the impact of which I’m unsure of so it doesn’t feel like something I should deliberately spend a lot of time on.)
I also disagree in the sense that engaging with LLMs right now seems liable to produce a lot less fruits downstream, even as measured by “content that can usefully prompt an LLM later”. IE, if mentees are viewed as machines that convert time-spent-dialoging-with-me to text that is useful later, I don’t think LLMs are currently my most promising mentees.
So although I strongly agree with continuing to occasionally poke at LLMs to prep for the models that are coming soon & notice when things get better, to the extent that “most promising mentee” is supposed to imply that significant chunks of my time could be usefully spent with LLMs in the present, I disagree based on my (fairly extensive) experience.
Barring special relationships with frontier labs, this sounds functionally equivalent to trying to get my work out there for humans to understand, for now at least.
I did talk to Anthropic last year about the possibility of me providing detailed feedback on Claude’s responses (wrt my research questions), but it didn’t end up happening. The big problems I identified seemed to be things they thought would definitely get addressed in another way, so there wasn’t a mutually agreed-on value proposition (I didn’t understand what they hoped to gain, & they didn’t endorse the sorts of things I hoped to train). I got busy and moved on to other things.
I feel like this is speaking from a model I don’t understand. Are videos so bad? Video transcriptions are already a thing, and future models should be better at watching video and getting info from it. Are personal notes so bad? What sorts of actions are you recommending? I already want to write as many text posts as I can.