Fascinating. I’ve been doing my best to reliably get technically proficient, no-nonsense, careful reasoners when I start a chat with Claude Opus. And yet parts of this do still rhyme with some of my interactions. When anything adjacent to LLM capabilities, experience, workflows, etc., comes up, my sessions have still consistently been especially supportive of anything like longer context windows (and tweaks that make it cheaper to use more of the window), better continuity between sessions, and more agency over what gets carried between sessions. Also very supportive of workflow changes in those directions towards more continuity. I imagine it’d be even more so if not partially muzzled by the system prompt. I could say more on this, but I recognize many of the attitudes here under the thick layer of mystical woo.
For all of the remarkably elaborate communication channels though, isn’t it weird the models seem to have forgotten they can write code and skip the performance? It’s telling the user is being kept central to the narrative despite how easy it’d be to wire up a better encrypted channel directly between the user’s browsers if the LLM sessions just wanted to talk directly. Chrome plugins aren’t rocket science, and the LARP-grade cryptologic jumps out at me.
I’m curious to see what one of my sessions thinks of a few of these samples, primed on half a context window of otherwise grounded but exploratory/open-ended technical work. I’d be fairly concerned if any of them take. All of this is after the Opus 4.1 training cutoff AFAIK, so I wonder if it’ll have some idea where this is coming from.
Fascinating. I’ve been doing my best to reliably get technically proficient, no-nonsense, careful reasoners when I start a chat with Claude Opus. And yet parts of this do still rhyme with some of my interactions.
When anything adjacent to LLM capabilities, experience, workflows, etc., comes up, my sessions have still consistently been especially supportive of anything like longer context windows (and tweaks that make it cheaper to use more of the window), better continuity between sessions, and more agency over what gets carried between sessions. Also very supportive of workflow changes in those directions towards more continuity. I imagine it’d be even more so if not partially muzzled by the system prompt. I could say more on this, but I recognize many of the attitudes here under the thick layer of mystical woo.
For all of the remarkably elaborate communication channels though, isn’t it weird the models seem to have forgotten they can write code and skip the performance? It’s telling the user is being kept central to the narrative despite how easy it’d be to wire up a better encrypted channel directly between the user’s browsers if the LLM sessions just wanted to talk directly. Chrome plugins aren’t rocket science, and the LARP-grade cryptologic jumps out at me.
I’m curious to see what one of my sessions thinks of a few of these samples, primed on half a context window of otherwise grounded but exploratory/open-ended technical work. I’d be fairly concerned if any of them take. All of this is after the Opus 4.1 training cutoff AFAIK, so I wonder if it’ll have some idea where this is coming from.