When you improve the capabilities of an LLM (or any AI agent with enough generality really) you also probably improve its ability to cooperate with copies of itself, where cooperation ability is some loose collection of abilities such as checking other copies work, planning work for other copies to carry out, course correcting other copies when they get sidetracked, integrating a copy’s work into the main project, etc.
Now that LLMs are good enough to create useful multi-agent systems such as Claude Code, Cursor, Gas Town, etc., I expect these systems to get better at a faster rate than the underlying models (and not because the people are putting work into improving the structure of these systems, although I suspect that will also help).
I’m not sure if these systems are currently more or less than the sum of their parts (if Claude is able to do a task with X difficulty, are 10 Claude agents able to do a task of >10X difficulty, 10X difficulty, or <10X difficulty?). But I expect that as cooperation gains accumulate, they will become less and less like “less than the sum of their parts” (as they stop bungling communication with each other) and more and more like “greater than the sum of their parts” (as they learn how to get gains from specialization or something?).
This has updated me significantly towards shorter timelines, especially for time to automated coders. I really have no idea what happens after that though.
Prediction: 80% chance of fully automated coding “teams” by the end of 2026.
Coming across this series years later, but loving it so far!
I’m no expert on meta-ethics either (nor neuroscience), but I disagree slightly with the meta-ethical takes at the end. In particular, I expect that moral reasoning processes converge a decent amount among most humans (leaving aside true sociopaths) under ideal conditions, although maybe not all the way. In particular, I think I could predict with high confidence the broad strokes of morality in 2100 or 2200, conditional on a few things like humanity’s survival and the arrival of ASI. If I’m correct on this, I think that’s mild evidence for convergence. And I think this causes me to put more weight on something like the importance of a Long Reflection period.
I think we have different intuitions here because I am putting somewhat more weight on a lot of (bad, in my view) moral reasoning being (1) reflectively inconsistent with itself, or (2) based on a fundamentally confused understanding of the universe.
Examples of “bad” moral reasoning I expect to go away due to (1) or (2) above:
Not caring about animal welfare
Not caring / actively wanting harm to certain groups of people
Most moral reasoning based on religious belief
Repulsion around a lot of transhumanist ideas
Thinking badly / demanding punishment of people who have caused harm/suffering (This last one is a bit more controversial, I’m sure, but I stand by it. Arguments against free will + a lack of instrumental reasons for justice in a well-designed Utopia seem enough to undo it for me.)
Now I don’t expect reasoning on everything to converge. I agree that different weights on different innate drives probably cause people to value things a bit differently, and these values may diverge further upon being able to self-modify. But maybe we just disagree on how much this ends up mattering for a Utopia? It seems to me that this would just cause people to seek slightly different flavors of fun.
One caveat: if people find ways to primarily use their upgraded intelligence to further rationalize their existing beliefs, then I agree that convergence is unlikely. I hope this doesn’t happen, and I think a Long Reflection period could be valuable especially with respect to getting people to not do this. In fact, I think a Long Reflection or something similar is probably extremely important for getting the above list to happen, such that a future with a Long Reflection period would be in expectation much “more valuable” than one without such a period.
If you’re interested, I’d be willing to formalize the above prediction a bit more and make a (probably very silly and moot) bet on it. Maybe 1/1000th of my wealth (if there is such a thing) post-singularity (if we’re alive)? Could be your chance to win many star systems ;).