But they were, all of them, deceived, for another csv file was made. In the subfolder of a subfolder, the LLM hardcoded a bunch of data, and never wrote code to pull the actual values. And into this csv file the LLM poured vaguely reasonable-sounding numbers, which unfortunately did not match the real world. One subtle csv file of made-up data which invalidated all of the complicated calculations and projections.
johnswentworth
Do you think this is a good argument for multi-scale modularity in biology?
I think the multi-scale aspect is mostly orthogonal to this argument, but yeah the argument could be applied at multiple scales.
Also thoughts on multi-agent models of mind with this model in the background? Finally any thoughts on applying this to whether we will have a group of AI systems doing RSI or a singular one?
Typically, the load-bearing aspect of multi-agent models is that the multiple agents have different utility functions. This argument is compatible with multiple utility functions in principle, but that’s not really what it’s about; it would apply arguably-more-cleanly in systems with just one optimization objective (but multiple subgoals, e.g. multiple resources like apples and bananas).
I guess there also has to be the aspect of trade already present in the system?
Nope.
(The other questions I don’t yet have much to say about.)
Fixed. Thanks, and sorry for the confusion.
Fixed. Thanks.
Specialization is a Driver of Natural Ontology
I read an article in Fortune magazine twenty years ago about this, for blood gold. According to the story, the industry had so many layers of middlemen that it was impossible in practice to figure out where any given gold came from. The big change was when Walmart decided they wanted to offer clean gold products. They’re such a large buyer that they could negotiate for source tracking through the whole chain, and it was worthwhile for suppliers to put that tracking in place.
… though it’s not not a puff piece for Walmart, so take with a lot of salt.
In the style of facebook’s feed, including a high proportion of randomly generated ads.
This is a short but decent articulation which strongly agrees with my own observations; good job.
With my ex, it was a common-knowledge-to-the-two-of-us phenomenon that she would always find something to worry about. So when she felt worried or anxious about relatively minor things, I would sometimes jokingly respond “Oh good! If you’re worrying about that then there must not be any serious problems right now.”.
How much money did these cost?
David tries to punt things to LLMs at least once a day on average when we’re working. So far, they continue to work best when they can act as Google Search Plus Plus—i.e. when there’s some already-known fact relevant to what we’re doing, and they surface that fact to us. Occasionally they can complete a conceptually-simple-but-technically-dense proof by combining a few such facts, and very often they can write some useful code by combining a few already-known pieces.
For anything novel, they remain almost always useless in our experience; they string together words which sound relevant but the semantics don’t make any sense.
The evolutionary explanation I thought was standard for sex differences in variability is that males (of most species) face way more potential upside—i.e. they can potentially mate with a whole lot of different females and have far more kids than any single female can. Females typically have a much lower ceiling on kid-count, so they face less potential fitness upside from high-variance strategies.
I liked the part of this post about X-inactivation a lot, even though it turned out wrong. It would be really interesting if most of the variability difference across sex turned out to share one simple mechanistic basis, though that doesn’t seem very likely on my current models of the underlying selection pressures.
Quite the opposite: the subject-we-gesture-at-with-the-word-”alignment” is not particularly provocative or controversial when you think about it deeply, at least not along the axes people generally argue over in the context of morality/ethics, because those axes just aren’t that technically central or relevant.
Personally, my guess is that morality and ethics themselves would not be particularly controversial or provocative if people usually approached them with a goal of deep technical understanding. That’s just not the goal with which approximately-anybody, including nearly all professional philosophers, approaches the subject—as we see e.g. on that Stanford Encyclopedia page. Those are people trying to have the equivalent of fun house party conversations, or in some cases write manifestos, not people seriously trying to achieve deep technical understanding.
Indeed, invoking the words “good” or “right” also tend to make people dumber (though less so than “morality” or “ethics”), and trying to do philosophical analysis of what is “good” or “right” is exactly the thing which seems to insta-brain-kill people; it’s exactly the lever which “morality” and “ethics” pull.
For example, let’s look at two pages in the Stanford Encyclopedia of Philosophy. I picked these by pulling up the table of contents, and then clicking the first one which seemed not-very-morality-loaded and the first one which seemed very-morality-loaded.
First up, abduction. No morality talk here. It’s describing a feature of human reasoning, which seems functionally load-bearing for epistemics in some cases and would probably generalize to other kinds of minds (like aliens or AI). It doesn’t trivially fit a couple common frames of epistemics, which is why it’s interesting. A lot of the discussion is centered around pretty narrow or outdated models of reasoning, but it’s a technically interesting and sensible article, which inspires good questions at least.
In contrast, the ethics of abortion. Before we even get to the actual content, note the topic. Abduction is a topic relevant to understanding minds and reasoning in general, a topic which would likely be relevant even to AIs; it belongs in a generalizable world-model. Abortion, by contrast, would be irrelevant to many other kinds of minds—e.g. human-level-intelligent platypuses would lay eggs, and therefore the whole issue of abortion would not have a clean analogue for them. (And human-level intelligent ants would be in a whole different frame!) Almost certainly, the reason why a Stanford Encyclopedia page exists for abortion at all is that it was a major hot-button political topic for a while, which won the memetic competition for attention in US politics. But in the grand scheme of things, it is just not that important of a question at all even for humans, and entirely irrelevant to many other kinds of minds. The very fact that people pay so much attention to it is itself a strong sign of mindkill.
Looking at the content of the page… the entire thing is a string of analogies and attempts to generalize various heuristics to the case of abortion. Notably sparse or absent is:
Technical engagement with the developmental process, when various things come online for a fetus/baby (like e.g. pain, self-awareness).
Technical engagement with the way humans’ preferences/values actually typically form. Spoiler: it ain’t usually by thought-experiments involving a violinist.
Technical engagement with the first and second-order actual effects of abortion laws/norms (though laws/norms are of course distinct from morality, consequentialism still matters).
More vibe-ishly, compared to the abduction article, the whole thing very much has a bikeshed vibe to it. It’s all the sort of stuff which would make good fodder for conversation at a house party, not the sort of stuff which involves dense technical study and deep understanding.
I personally avoid even using the words “morality” or “ethics” in the context of AI alignment, because both of those words reliably turn the vast majority of otherwise-sensible people into morons the moment they are spoken.
Yeah, that would probably capture the intuition, though this isn’t something which really needs precise operationalization in order to be useful.
That is a cool and relevant domain of research I had not encountered before, thank you!
Good question, we don’t yet know the answer. It is definitely the right kind of question to ask.
My framing has evolved over time; this is a scattered list as opposed to a unified picture. But it’s still basically-right as the motivating list for the bigger picture.
Our own work touches on several of them, but I generally see very few people working on these kinds of problems.
A background piece of my own models (which could be right/wrong independent of the content of this post) is that bounded rationality constraints basically just don’t bind at all in practice once a mind passes a certain threshold, and in particular will not bind for even moderately-superhuman general AI.
The canonical example illustrating how this could happen is inference involving an ideal gas. Roughly speaking, once an agent is smart enough to use the Boltzmann distribution, even a Jupiter brain will not be able to do significantly better in practice, because chaotic dynamics wipe out all the other signal. In that case, bounded rationality constraints cease to bind once the agent is smart enough to use a Boltzmann distribution (again, roughly speaking).
I mean, it easily could be, that would not be a huge surprise. But it was originally generated as a stylized description of actual experiences I’ve had with Claude. For instance, a week ago I asked it for data on the biggest ETFs by trade volume, and it just hardcoded some numbers into a file without actually looking anything up.