It is cool, and I have believed something like this for a while. Problem is that Claude 3.5 invalidated all that—it does know how to program, understands stuff, and does at least 50% work for me. This was not at all the case for previous models.
And all those “LLL would be just tools until 2030” arguments are not baked by anything and based solely on vibes. People said the same about understanding of context, hallucinations, and other stuff. So far the only prediction that worked is that LLM gains more common sense with scaling. And this is exactly what is needed to crack its agency.
People said the same about understanding of context, hallucinations, and other stuff
Of note: I have never said anything of that sort, nor nodded along at people saying it. I think I’ve had to eat crow after making a foolish “LLMs Will Never Do X” claim a total of zero times (having previously made a cautiously small but nonzero number of such claims).
It is cool, and I have believed something like this for a while. Problem is that Claude 3.5 invalidated all that—it does know how to program, understands stuff, and does at least 50% work for me. This was not at all the case for previous models.
And all those “LLL would be just tools until 2030” arguments are not baked by anything and based solely on vibes. People said the same about understanding of context, hallucinations, and other stuff. So far the only prediction that worked is that LLM gains more common sense with scaling. And this is exactly what is needed to crack its agency.
Of note: I have never said anything of that sort, nor nodded along at people saying it. I think I’ve had to eat crow after making a foolish “LLMs Will Never Do X” claim a total of zero times (having previously made a cautiously small but nonzero number of such claims).
We’ll see if I can keep up this streak.