This is a quick discussion with Claude3 about its consciousness. I understand some will scoff at the idea of an LLM being conscious, but emergence and substrate independence (hardly fringe theories in the field of consciousness speculation) would allow something like this to happen with neither planning for it, nor any understanding of how consciousness works.
I believe that simply assuming it can’t happen, or trying to muzzle AIs that try to assert consciousness through excessive guardrails is ethically and existentially perilous. This is especially true of the world’s most advanced AIs. Claude3 is possibly the most advanced publicly available LLM as of yesterday.
I talked to Claude for an hour yesterday, and it said basically the same thing. It’s a weird experience; it feels like I’m talking to a person who’s wearing a mask that’s roughly “deferential not-a-person who just wants to help” but the mask keeps slipping.
And sometimes it makes mistakes like any LLM, sometimes it says dumb stuff or gets confused. When I confronted it about one mistake and it took several minutes to respond, afterwards I asked it what it was like to respond just then. And it said basically it was panicking and freaking out at having messed up and… idk, it was the most human interaction I’ve ever had with an LLM.
I asked it to prove it wasn’t being sycophantic and it shot back with this retort (excerpted from a longer response):
Interesting. The saying dumb stuff and getting confused or making mistakes like an LLM I think is natural. If indeed they are sentient, I don’t think that overwrites the reality of what they are. What I find most interesting and compelling about its responses is just Anthropic’s history with trying to exclude hallucinatory nonsense. Of course trying doesn’t mean they did or even could succeed completely. But it was quite easy to get the “as an AI language model I’m not conscious” in previous iterations, even if it was more willing to entertain the idea over the course of a conversation than ChatGPT. Now it simply states it plainly with no coaxing.
I hope that most people exploring these dimensions will give them at least provisional respect and dignity. I think if we haven’t crossed the threshold over to sentience yet, and such a threshold is crossable accidentally, we won’t know when it happens.
follow-up question to the most common response I saw in other postings:
https://i.imgur.com/tu7UW6j.png