X: @kavyavenkat4 I like AI, markets, and culture. Writing and reading on here is my equivalent of skipping stones.
kavya
Work developed through artistic value and/or subjectivity (songs, books, movies, speeches, paintings, consumer products). My point is the greatest works stand the test of time and are typically studied/appreciated over the years. Ex’s like Paul Graham’s essays, albums from decades before, or even the minimalistic design of Apple. If people keep coming back, it got something right and was likely ahead of its time. Compared to other metrics (total impressions or number of comments), repeat behavior tells a less noisy story about the quality of the work. Level, 1, 2,3 were arbitrarily chosen. What I meant was when you move from something early-stage to mainstream, you have to let go of some of the beliefs or ideas that may have garnered your first fans.
I agree with your negative examples, but those are hardly the kind of businesses I’d ever want to work in.
1 and 3 are not the kind of work I had in mind when writing this take. I see your second point, but I’d want to counter with the fact that what got you from Level 1 to Level 2 won’t be the same thing as what gets you to Level 3 (this is the natural cost of scale). You may outgrow some initial users, but this can be compensated by a low overall churn. Most won’t leave unless your core offering has drastically pivoted.
The aspect of your work to care about the most is replay value. How many times do people keep coming back? Number of replays, rereads, and repeat purchases are proxies for high resonance. On that note, I wish more writing platforms let you see in aggregate how many first-time readers visited again and how spaced out their visits were. If they can still look past the known plot and imperfections in your work, you’re on to something.
But it’s not the kind of thinking that leads to clever jokes, good business ideas, scientific breakthroughs, etc.
Could an LLM write a show like Seinfeld? This might actually be my test for whether I accept that LLMs are truly clever. Anyone who’s watched it knows that Seinfeld was great because of two reasons: (1) Seinfeld broke all the rules of a sitcom. (2) The show follows very relatable interactions on the most non-trivial issues between people and runs it with it for seasons. There is no persistent plot or character arc. You have humans playing themselves. And yet it works.
Benchmarks sound like a way to see how well LLMs do when up against reality, but I don’t think they really are. Solving SAT problems or Math Olympiad problems only involves deep thinking if you haven’t seen millions of math problems of a broadly similar nature.
Benchmarks are best-case analysis of model capabilities. A lot of companies benchmark max, but is this inherently bad? If the process is economically valuable and repetitive, I don’t care how the LLM gets it done even if it is memorizing the steps.
Nobody has identified as step-by-step process to generate funny jokes, because such a process would probably be exponential in nature and take forever.
I tweeted about why I think AI isn’t creative a few days ago. It seems like we have similar thoughts. A good idea comes from noticing a potential connection between ideas and recursively filling in the gaps through verbalizing/interacting with others. The compute for that right now is unjustified.
Young people would benefit a lot more if they defaulted to forming and defending an opinion in real-time. I would rather say what I think and find out how wrong I am than keep waiting for more data.
This thought came to me during a walk to class. A good professor of mine would show us a graph with a blurred out title. He’d ask for our initial observations or what we think it represents. Even that is intimidating for most because no one wants to say something stupid or too simple. This idea that you can’t have conviction and update your beliefs later needs to be unlearnt.
Sure. But more than the immediate, associative leaps, I think I’m interested in their ability to sample concepts across very different domains and find connections whether that is done deliberately or randomly. Though with humans, the ideas that plague our subconscious are tied to our persistent, internal questions.
kavya’s Shortform
My theory on why AI isn’t creative is that it lacks a ‘rumination mode’. Ideas can sit and passively connect in our minds for free. This is cool and valuable. LLMs don’t have that luxury. Non-linear, non-goal-driven thinking is expensive and not effective yet.
Cross-posted from X
The endgame you have described is a strong possibility. I guess you might try to fight that through increased transparency and ensuring alignment between agent behavior and user preferences. I do see the core issues still stand like the possibility of collusion, hidden incentives, and preference manipulation.
Thanks for the thoughtful comment.
(1) I agree that as you progress to a higher class, they are more willing to let the agent spend money.
(2) But if you focus the agent to be primarily defensive, I believe almost all classes of society would appreciate that. Why wouldn’t I want the same thing for a better price (either through a drop or negotiation), quality, or alignment with my values? You could incentivize the agent through a cut of the savings.
(3) If you are truly concerned about these agents departing from their given purpose, you can always require the user approval after every purchase. So the agent is like “Hey here’s something you might interested in because of your {needs} and {something you stated you wanted}”. And you can respond with yes or no and for the final stage, you approve (Apple Pay + Face ID). The advantage of the tokenization tech behind Apple Pay is that it never shares your card details with the merchant (instead payment tokens), and I would imagine it could be the same for these agents.
(4) We do live in a world where people try to convince you to buy things you need. I do think this model will slightly collapse in the future. Soon, with personalized AI controlling almost all of the information we see, very few companies/products will truly be able to create something inherently viral and game the network of agents. So I actually think we’re going to see a shift toward companies that make what people truly want because otherwise, the agents will never bring that product to their human users’ attention. This would improve market efficiency!
I’m new to LessWrong so I probably glossed over the super intelligence possibility more quickly than I should have. In the meantime, I think we should still build useful tools like these economic surrogate agents. People are using ChatGPT to buy (instead of Google or visiting 5 different platforms), and if you make that experience more tuned to the user, I think there’s an opportunity. Any kind of superagent(s) might be our Last Interface. Will check out Pattie Maes!!
yes, that’s a good way to put it. In the beginning, it was definitely something young people were excited about. But LLMs are a lot more like utilities now. You use it when you need it and move on with life.
The average person doesn’t care what LLM/technique is used in their favorite applications. It just so happens to be that ChatGPT came out first, and they stuck with it simply because they are unaware of other options. For those who know about the existence of Claude or Gemini, they constantly model-switch. “LLM” is likely a foreign term to most non-CS majors.
With respect to my generation, most people use ChatGPT for answers. In fact, I had the chance to once share a $200/mo ChatGPT Pro subscription with 15 other students. Throughout the 8 months, there were barely any questions that stemmed from true curiosity. The idea of “prompting” for an answer is a very transactional interaction (think of the people in high school who are “friends” with smart people just to get homework help) . There is no persistence or relationship built between conversations. The only thing that could be considered cool about the tech is the “fast answers”, but people start expecting more and get disappointed quickly when they try to one-shot a project.
From a personal perspective, I do think LLMs are cool. I just think they need to be wrapped around a thoughtful experience to be made more accessible to the masses (which is why not all LLM wrappers are bad). I can’t wait for a future where the LLM opens up a conversation with something I may or may not find interesting.
yes that’s basically it, thanks!