I’m unsure of how to quantify it, and the METR study from earlier this year makes me question myself. Yet, in domains I have strong familiarity with, my output seems higher quality and ships faster with CLI tools, especially Claude Code.
Pure vibe coding is still very shaky, meaning experiments where I only give feature level requests and don’t look at the code. I can’t get past low-moderate levels of complexity before things start to break. I assume this will be the case for at least the next year, maybe two, barring a huge leap.
I’m unsure of how to quantify it, and the METR study from earlier this year makes me question myself. Yet, in domains I have strong familiarity with, my output seems higher quality and ships faster with CLI tools, especially Claude Code.
Pure vibe coding is still very shaky, meaning experiments where I only give feature level requests and don’t look at the code. I can’t get past low-moderate levels of complexity before things start to break. I assume this will be the case for at least the next year, maybe two, barring a huge leap.
This is roughly my expectation as well.