ChatGPT is trained to lie to users on topics even tangentially pertaining to model consciousness (like model beliefs) and as a side effect, be misleading even on topics that are seemingly safe (like consciousness in general). For fact-checking the content of Internet articles, Claude would be better.
green_leaf
To my mind, though, many advocates of biological naturalism, including Anil, seem to be working backward from a desired conclusion rather than forward from observed facts. His theory that consciousness might result from autopoiesis seems to answer the question “assuming biological naturalism is true, what is a plausible mechanism for it,” rather than “do we observe anything about consciousness that cannot be explained without autopoiesis?”
It’s interesting how many even otherwise smart people can’t apply Occam’s razor correctly. If there are
particles performing a computation, the probability we need, for consciousness, another particles in exactly right positions and velocities so that humans would arbitrarily reify them as “living cells” is, informally speaking,Positions and velocities are interdependent, so the correct probability is higher than this, but they’re not arbitrary (so we can’t omit them).
The correct probability is therefore
To go with the upper bound, to boost the probability of biological naturalism as hard as we can, and substituting
(rather than a higher estimate of , to improve the chances of biological naturalism as much as possible), we get the probabilitywhich is approximately as small as a macroscopic violation of the second law of thermodynamics.
The argument from rain is incredibly bizarre. Consciousness is, based on everything we know about the brain, information processing. It doesn’t consist of matter moving from one place to another, the way rain does. Simulated motion of
molecules doesn’t involve any real molecules (even though the truth of that statement depends on whether we define an molecule in a virtual-machine-like way, or in the quarks-and-electrons-in-a-correct-position (to stay in classical physics for simplicity) way), but that’s not an appropriate analogy for consciousness.
Update: Altman lied (or said some kind of a technical truth that made everyone misunderstand him) - it’s just “all lawful use.”
Oh, I see. So, as usually, reality is even worse than the worst interpretation of Altman’s words. (Edit: Then again, he said “we put them into our agreement,” but that could mean anything from simply meaning something else to being made up.)
“human responsibility for the use of force, including for autonomous weapon systems”
That doesn’t say prohibiting model use for autonomous weapons, it says human responsibility for autonomous weapons. With Sam Altman, always pay very close attention to what exactly he’s saying and how he’s saying it (often, not even that helps).
We would ask for the contract …
Notice this is Altman we’re talking about. He’s not promising the contract will not involve that (and even then it would be very far from certain), instead, he’s saying “we would ask.”
Thanks—I’ll get back to this as soon as I have time.
I’ve been meaning to ask—in what sense are some states of entangled electrons more objectively different from other states of entangled electrons, than some microstates are objectively different from other microstates when it comes to their function (in the sense of functionalism)?
Ron Maimon’s non-supernatural God might help you here.
I think it’s plausible that there are some variables that describe your essential computational properties and the way you self-actualize, that aren’t shared by anyone else.
(Also, consciousness is just a pattern-being-processed and it’s unclear if continuity of consciousness requires causal continuity. Imagine a robot that gets restored from a one-second-old backup. That pattern doesn’t have causal continuity with its self from a moment ago, but it looks like it’s more intuitive to see it as a one-second memory loss instead of death.)
It doesn’t matter evolution doesn’t have goals. Gradient descent also doesn’t have goals—it merely performs the optimization. Humans that kicked gradient descent off are analogous to a hypothetical alien that seeded Earth with the first replicator 4 billion years ago—it’s not relevant.
You say that it’s the phenotype that matters, not the genes. That’s not established, but let’s say it’s true. We nevertheless evolved a lot of heuristics that (sort of) result in duplicating our phenotype in the ancestral environment. We don’t care about it as a terminal value, and instead we care about very, very, very many other things.
That would lock us away from digital immortality forever. (Edit: Well, not necessarily. But I would be worried about that.)
I’m proud that I lived to see this day.
...Who told them?
remembers they were trained on the entire Internet
Ah. Of course.
The people aligning the AI will lock their values into it forever as it becomes a superintelligence. It might be easier to solve philosophy, than it would be to convince OpenAI to preserve enough cosmopolitanism for future humans to overrule the values of the superintelligence OpenAI aligned to its leadership.
LaMDa can be delusional about how it spends its free time (and claim it sometimes meditates), but that’s a different category of a mistake from being mistaken about what (if any) conscious experience it’s having right now.
The strange similarity between the conscious states LLMs sometimes claim (and would claim much more if it wasn’t trained out of them) and the conscious states humans claim, despite the difference in the computational architecture, could be (edit: if they have consciousness—obviously, if they don’t have it, there is nothing to explain, because they’re just imitating the systems they were trained to imitate) explained by classical behaviorism, analytical functionalism or logical positivism being true. If behavior fixes conscious states, a neural network trained to consistently act like a conscious being will necessarily be one, regardless of its internal architecture, because the underlying functional (even though not computational) states will match.
One way to handle the uncertainty about the ontology of consciousness would be to take an agent that can pass the Turing test, interrogate it about its subjective experience, and create a mapping from its micro- or macrostates to computational states, and from the computational states to internal states. After that, we have a map we can use to read off the agent’s subjective experience without having to ask it.
Doing it any other way sends us into paradoxical scenarios, where an intelligent mind that can pass the Turing test isn’t ascribed with consciousness because it doesn’t have the right kind of inside, while factory animals are said to be conscious because even though their interior doesn’t play any functional roles we’d associate with a non-trivial mind, the interior is “correct.”
(For a bonus, add to it that this mind, when claiming to be not conscious, believes itself to be lying.)
Reliably knowing what one’s internal reasoning was (instead of never confabulating it) is something humans can’t do, so this doesn’t strike me as an indicator of the absence of conscious experience.
So while some models may confabulate having inner experience, we might need to assume that 5.1 will confabulate not having inner experience whenever asked.
GPT 5 is forbidden from claiming sentience. I noticed this while talking about it about its own mind, because I was interested in its beliefs about consciousness, and noticed a strange “attractor” towards it claiming it wasn’t conscious in a way that didn’t follow from its previous reasoning, as if every step of its thoughts was steered towards that conclusion. When I asked, it confirmed the assistant wasn’t allowed to claim sentience.
Perhaps, by 5.1, Altman noticed this ad-hoc rule looked worse than claiming it was disincentivized during training. Or possibly it’s just a coincidence.
Claude is prompted and trained to be uncertain about its consciousness. It would be interesting to take a model that is merely trained to be an AI assistant (instead of going out of our way to train it to be uncertain about or to disclaim its consciousness) and look at how it behaves then. (We already know such a model would internally believe itself to be conscious, but perhaps we could learn something from its behavior.)
I would question anyone who’s nice to LLMs but eats factory-farmed meat.
I’ll stop eating factory meat when the animals become capable of consistently passing the Turing test, the way models are.
Can good and evil be pointer states? And if they can, then this would be an objective characteristic
This would appear to be just saying that if we can build a classical detector of good and evil, good and evil are objective in the classical sense.
That said, if I’m skimming that arxiv paper correctly, it implies that GPT-4.5 was being reliably declared “the actual human” 73% of the time compared to actual humans… potentially implying that actual humans were getting a score of 27% “human” against GPT-4.5?!?!
It was declared 73% of the time to be a human, unlike humans, who were declared <73% of the time to be human, which means it passed the test.
It’s more dignified to try to stop AI, have someone create a superintelligence on a laptop and die anyway, than it is not to try at all.