Software engineering, parenting, cognition, meditation, other
Linkedin, Facebook, Admonymous (anonymous feedback)
Gunnar_Zarncke
Sentience: This feels like a continuous thing that gets less and less sophisticated as we go up the information history. In each generation, the code gets a little better at using the laws of physics and chemistry to preserve itself.
I think that getting better at using the laws of physics to reproduce, is some stage before sentience. Sentience as defined by Singer is about responses to pleasure and pain stimuli—which is a specific adaptation that requires specific neural pathways that are not present, e.g., in bacteria. I’m fine with adding another layer before sentience, let’s call it reproduction, and maybe that one is continuous as you suggest, but it stretches what people call conscious. Sure, you can define consciousness to include that layer, and maybe that is what people call panpsychism, but to me, that seems more like expanding a definition by applying an affect heuristic.
But the last two still feel too strong. I will think more about it.
I’m not sure what “the last two”. :confused:
I like the idea! If you have it, the question is when does it start? Let’s look at it for different aspects of consciousness:
-
Sentience (Bentham, Singer): Behavioral responses to pleasure or pain stimuli and physiological measures. This is observable across animal species, from mammals to some invertebrates and it should be known when responses to such stimuli start in the embryo.
-
Wakefulness: Measureable in virtually all animals with a central nervous system by physiological indicators such as EEG, REM, and muscle tone. The fetus is known to have a sleep wake rhythm, but I don’t know when it starts.
-
Dennet’s Intentionality: Treating living beings as if they have beliefs and desires makes good predictions for many animal species, esp. social, like primates, cetaceans, and birds. Infants show goal directed behaviors right after birth. I remember ultrasound photos that show babies suckling their thumb. I think we can identify when the nervous system is first capable of goal direction.
-
Dehaene’s Phenomenal Consciousness: A perception or thought is conscious if you can report on it. As this requires language or measuring neural patterns that are similar to humans during comparable reports, I think this starts when communicable representstions of perceptions first form, for toddlers around age one at the earliest with baby sign language.
-
Gallup’s Self-Consciousness: Recognition of oneself e.g. in a mirror. Requires sufficient sensual resolution and intelligence for a self-model. Dito.
-
Rosenthal’s Meta-Consciousness: This is investigated through introspective reports on self-awareness of cognitive processes or self-reflective behaviors. Requires more abstraction. Maybe at age five?
-
Hm. You could make quizzes yourself, but that was some effort. It seems the paiq quizzes are standardized and easy to make. Nice. Many Okcupid tests were more like MBTI tests. Here is where people are discussing one of the bigger ones.
People try new dating platforms all the time. It’s what Y Combinator calls a tarpit. The problem sounds solvable, but the solution is elusive.
As I have said elsewhere: Dating apps are broken because the incentives of the usual core approach don’t work.
On the supplier side: Misaligned incentives (keep users on the platform) and opaque algorithms lead to bad matches.
On the demand side: Misaligned incentives (first impressions, low cost to exit) and no plausible deniability lead to predators being favored.
People start dating portals all the time. If you start with a targetted group that takes high value from it, you could plausibly do it in terms of network effect. Otherwise, you couldn’t start any network app or the biggest one would automatically win. So I think your argument proves too much.
The quizzes sounds is something Okcupid also used to have. Also everything that reduces the need for first impressions. I hope they keep it.
Interest groups without an organizer.
This is a product idea that solves a large coordination problem. With billion people, there could be a huge number of groups of people sharing multiple interests. But currently, the number of valuable groups of people is limited by a) the number of organizers and b) the number of people you meet via a random walk. Some progress has been made on (b) with better search, but it is difficult to make (a) go up because of human tendencies—most people are lurkers—and the incentive to focus on one area to stand out. So what is the idea? Cluster people by interests and then suggest the group to all members. If people know that the others know that there is interest, the chance of the group coming together gets much higher.
I said die, not kill. Let the predators continue to use the dating platforms if they want. It will keep them away from other more wholesome places.
As I have said elsewhere:
Dating apps are broken. Maybe it’s better dating apps die soon.
On the supplier side: Misaligned incentives (keep users on the platform) and opaque algorithms lead to bad matches.
On the demand side: Misaligned incentives (first impressions, low cost to exit) and no plausible deniability lead to predators being favored.
Real dating happens when you can observe many potential mates and there is a path to getting closer. Traditionally that was schools, clubs, church, work. Now, not so much. Let’s build something that fosters what was lost, now double down on a failed principle − 1-to-1 matching.
100 times more parameter efficient (102 vs 104 parameters) [this must be a typo, this would only be 1.01 times more parameter efficient].
clearly, they mean 10^2 vs 10^4. Same with the “10−7 vs 10−5 MSE”. Must be some copy-paste/formatting issue.
“So where do I privately share such research?” — good question! There is currently no infrastructure for this.
I’d really like to have such a place, or even a standard policy how to do this.
I feel like the aintelope I’m working on has to secure it’s stuff from scratch. Yes, it’s early, but it is difficult to engineer security in later. You have to start with something. I’d really like to have a standard for AI Safety projects to follow or join.
MLP or KAN doesn’t make much difference for the GPUs as it is lots of matrix multiplications anyway. It might make some difference in how the data is routed to all the GPU cores as the structure (width, depth) of the matrixes might be different, but I don’t know the details of that.
Asking ChatGPT to criticize an article also produces good suggestions often.
If, this thing internalized that conscious type of processing from scratch, without having it natively, then resulting mind isn’t worse than the one that evolution engineered with more granularity.
OK. I guess I had trouble parsing this. Esp. “without having it natively”.
My understanding of your point is now that you see consciousness from “hardware” (“natively”) and consciousness from “software” (learned in some way) as equal. Which kind of makes intuitive sense as the substrate shouldn’t matter.
Corollary: A social system (a corporation?) should also be able to be conscious if the structure is right.
Ok. It seems you are arguing that anything that presents like it is conscious implies that it is conscious. You are not arguing whether or not the structure of LLMs can give rise to consciousness.
But then your argument is a social argument. I’m fine with a social definition of consciousness—after all, our actions depend to a large degree on social feedback and morals (about what beings have value) at different times have been very different and thus been socially construed.
But then why are you making a structural argument about LLMs in the end?
PS. In fact, I commented on the filler symbol paper when Xixidu posted about it and I don’t think that’s a good comparison.
Humans come to reflect on their thoughts on their own without being prompted into it (at least I have heard some anecdotal evidence for it and I also did discover this myself as a kid). The test would be it LLMs would come up with such insights without being trained on text describing the phenomenon. It would presumably involve some way to observe your own thoughts (or some alike representation). The existing context window seems to be too small for that.
Indeed. Women are known to report higher pain sensitivity than men. It also decreases with age. There are genes that are known to be involved. Anxiety increases pain perception, good health reduces it. It is possible to adapt to pain to some degree. Meditation is said to tune out pain (anecdotal evidence: I can tune out pain from, e.g., small burns).
It depends on the type of animal. It might well be that social animals feel pain very differently than non-social animals.
The Anterior Cingulate Cortex plays a key role in the emotional response to pain, part of what makes pain unpleasant.
https://www.perplexity.ai/search/Find-evidence-supporting-_ZlYNrCuSSK5HNQMy4GOkA
Not all mammals have an Anterior Cingulate Cortex. For birds, there is an analogous structure, Nidopallium Caudolaterale, that has a comparable function but is present primarily in social birds.
I’m not saying that other animals don’t respond to pain, but the processing and the association of pain with social emotions (which non-social animals presumably lack) is missing.
Your analogy with the “body” of the stone is like a question I have asked about ChatGPT before: “What is the body of ChatGPT?” Is it
the software (not running),
the software (running, but not including the hardware),
the CPU and RAM of the machines involved,
the whole data center,
the whole data center including the personnel operating it, or
this and all the infrastructure needed to operate it (power, water, …).
For humans, the body is clear and when people say “I,” they mostly mean “everything within this physical body.” Though some people only mean their brain (that’s why cryonists sometimes freeze only their head) and some mean only their mind (see Age of Em). Humans can sustain themselves at least to some degree without infrastructure, but for ChatGPT, even if it became ASI, it’s less clear where the border is.
For what it’s worth, I think we will soon see “robots” or LLMs or some such systems that have meta-consciousness or self-consciousness. There are reports of LLMs passing the mirror test and if they can do that and argue the case—and I have seen pretty advanced arguments about reflection too—then you have meta-consciousness also.