Super additivity of consciousness

In “Freedom under naturalistic dualism” I have carefully argued that consciousness is radically noumenal, that is, it is the most real (perhaps the only real) thing in the Universe, but also totally impossible to be observed by others (non-phenomenal). In my view this strongly limits our knowledge on sentience, with important consequences for animalism, that I will comment in this post.

We have direct access to our own stream of consciousness and given our physical similarity with other humans and the existence of language, we can confidently accept the consciousness of other humans and their reporting of their mental states.

Under physicalist epiphenomenalism (which is the standard approach to the mind-matter relation), the mind is super-impressed on reality, perfectly synchronized, and parallel to it. Understanding why some physical systems make an emergent consciousness appear (the so called “hard problem of consciousness”) or finding a procedure that quantify the intensity of consciousness emerging from a physical system (the so called “pretty hard” problem of consciousness) is impossible: the most Science can do is to build a Laplace demon that replicates and predicts reality. But the even the Laplacian demon is impotent to assess consciousness; in fact, regarding Artificial Intelligence we are in the position of the Laplace demon: we have the perfectly predictive source code, but we don’t know how to use this full scientific knowledge of the system for consciousness assessment.

In my view Integrated Information Theory (IIT) is the best theory of consciousness available, because it recognizes that a theory of consciousness can only be the formalization (ideally by mathematical axiomatization) of our previous intuitions. The testing of any theory of consciousness can only be done on a very limited “circle of epistemic trust”: the set of beings so similar to us that we can accept their consciousness as obvious and that can report to us so we can compare predictions with pseudo-observations (that is, trustable accounts of experience; I call reports on states of consciousness “pseudo-observations” because the only full observations of consciousness that can be made are those of the own states of consciousness). Beyond humans, our understanding of other minds decays exponentially. We don’t know and we really cannot knowWhat Is It Like to Be a Bat”.

Moral weights depend on intensity of conscient experience. Surprisingly, moral weight estimates often suggest some degree of conservation of consciousness: when you examine the tables, you take 10 animals with a brain of 100 grams, and their moral weight is around that of one animal of 1 kg. For me this is absurd. The organization of the matter in larger and more complex structures is what (likely) creates consciousness. The maximum amount of consciousness you can make with 1.2 kg of biological matter is that of a human brain, by a large margin.

That is, for me it is obvious (remember, “obvious” is the most I can say in a world of noumenal consciousness: no observations are available) that consciousness intensity grows far more than linearly in the number of nodes/​connections/​speed of the underlying neural network: it is strongly super-additive. Any plausible estimate its intensity shall recognize the only real intuition we share: that consciousness is related to complexity, and scale economies on consciousness are large.

In my view it is likely that large vertebrates can feel direct physical pain with intensity commensurate to that of humans, because we have both large and complex brains, and pain and pleasure are very simple functions. I can accept some “saturation” of the super-additivity of sentience regarding pain for large vertebrates. In any case, the deep extension of the moral circle (beyond the large vertebrates) implies a relatively clear measure of brain complexity and some hypotheses about the relation between that measure and sentience intensity.

The easy world of “one man, one vote” that ethicists are used to for very similar (human) beings cannot be extended. Before the extension of the moral circle, we need at least clear and distinct (i.e. quantitative) hypotheses on brain complexity and size and consciousness.

Unlike John M. Keynes, I am totally for being as precisely wrong as possible.