Some thought experiments on digital consciousness

These are some hypothetical scenarios involving emulating a (human) mind on a digital computer. They seem to present counterintuitive implications for the question of whether an emulated mind would actually be conscious or not. This relates to the question of whether consciousness is substrate independent, and whether consciousness is fundamentally computational. These scenarios are inspired by ideas in the book Permutation City, by Greg Egan.

These thought experiments challenge my intuitions about digital consciousness. Some of these challenges arise from the discrete nature of digital computation; with a discrete digital simulation you can increase the “distance” (in time or space) between timesteps, which is a bit of a mind-bending prospect. Additionally some of the confusion relates to what computation actually is, i.e. if you “play back” the entire recorded trajectory of a computational process, is this meaningfully different from “computing”?

The premise

The set-up is as follows: let’s consider an experiment to emulate a human mind on a digital computer. For argument’s sake, say this mind is being simulated as a discrete 3D cellular automata (CA) with simple rules to transition to the next state (this should be possible, since there exist configurations in very simple CAs like Conway’s Game of Life which are Turing complete). This includes simulating an environment for the mind to interact with, which I would say is necessary for a valid conscious experience. This environment is also contained within the 3D CA. Since it is a CA, the instantaneous state of the mind + environment is simply a 3D array of numbers, which can be straightforwardly represented and stored in a digital computer. Let’s also stipulate that the simulation is entirely self-contained and there are no channels for input and output.

Scenarios

Scenario 1 - straightforward simulation

The CA is stepped forward in a discrete-time manner, by computing the results of the CA transition function for each cell and updating the state. This means that the mind is being simulated and progressing forward in time, along with its simulated surroundings, both contained within the CA. The “instantaneous state” of the mind is therefore represented entirely in the dynamic memory of a digital computer. This is updated at a given wall-clock frequency, say 10,000 Hz, but let’s assume that the simulation has sufficient granularity in the time dimension to capture the exact biological and physical function of the human brain. In practice this could mean that the simulation is running slower than real time, however the simulation also includes its own environment, so from the perspective of the simulated mind this makes no difference.

If you are a materialist, and a functionalist, you think that consciousness (whether something has subjective internal experience, i.e. we can consider “what it is like” to be that thing) is substrate independent and only requires the right type of information processing. So for the scenario outlined above your conclusion should be that this mind will experience consciousness within the simulation, in the same way as if it were running in a biological body. This is assuming that information is being processed in exactly the same way. This is plausible since a very large CA simulation could capture the biological mechanics at a very high level of granularity, all the way down to simulating the laws of physics of our real world.

I suspect many people will agree that this simulated mind would be conscious. However we can now make some extensions to this scenario which test this conclusion.

Scenario 2 - record and playback

In this scenario we run the simulation in the same way as scenario 1 for a given period of time, and while we are doing this we record the entire instantaneous state at every frame. This will result in a 4D array which represents the full trajectory through time (a 3D array for each frame), let’s call this 4D array a mind-trajectory. This would take up a very large amount of storage space (particularly if the CA is also simulating a large environment for the mind to exist in), however we can assume that enough disk space is available.

We can then “play back” this trajectory, similar to how you would play a movie file or 3D motion capture, by loading every frame into the computer’s dynamic memory sequentially, one frame at a time. In some respects this is identical to scenario 1; we are iterating through frames which represent the entire state of the simulation, and the computer’s dynamic memory sees each frame in order. The only difference is that we are loading each frame from disk, rather than calculating the next frame using the CA’s transition function. For arguments sake say that these operations (loading from disk or calculating the next frame) take the same amount of time, so the computer’s dynamic memory sees exactly the same progression of states for exactly the same durations of time.

My intuition tentatively agrees that the mind contained in this trajectory will still be conscious and “alive” during this replay, in exactly the same way as scenario 1, because the computer’s memory is seeing an identical progression of states. I’m not sure why computing the transition function or not would make any difference to this fact. However this does stretch my intuition a bit, because normally I would think of a dynamic and alive simulation as being computed and actually processing information as it proceeds, not being “replayed” in this way.

Scenario 3 - static stored trajectory

We can extend this even further: if we already have the full state of the trajectory on a hard disk, then why does it matter whether or not we specifically load each frame into the computer’s dynamic memory sequentially (as we are doing in scenario 2 to replay the trajectory)? What is so special about dynamic memory compared to hard disk? It is still just a chunk of binary data, there is no élan vital possessed by RAM or CPU registers that can breathe life into a mind. Even if we don’t load the frames one-by-one into memory, they all still exist and are present on the hard disk, so can we say that the mind is alive and conscious on the hard disk? Even though it is static data, the full trajectory through time is stored, so in some sense I think you could argue the mind is “living” inside that trajectory. This is the point at which my intuition fails to consider this conscious any more, but it’s hard to put a finger on why.

Is it the sequential ordering of frames that matters? I’m struggling to explain why it would, although it does seem important. If this is important then maybe the trajectory data could be laid out contiguously in memory so that subsequent frames are next to each other.

Given that the frames are discrete in time, they are already in some sense separate from each other, there will be a finite amount of time it takes to switch to the next frame, whether computed by the transition function or loaded from memory.

If I really push my imagination then I can maybe accept that a 4D trajectory on a hard drive is alive and conscious, but this is quite a mind bending prospect, since we are not used to thinking of static data being “alive” in that sense. The implications are quite bizarre, you could have a lifeless-seeming hard drive which contains someone, a person, truly alive and experiencing consciousness, but where their time dimension is in our cartesian space on the hard drive, not the same as our time dimension.

Scenario 4 - spaced out stored trajectory

Say we take the leap and accept that a mind can be alive and conscious if a trajectory is stored statically on a hard drive. Now what if we split the trajectory data in half and store both halves on separate hard drives? What if we keep splitting it so that every frame is on a different hard drive, and then we gradually move these hard drives further and further apart in space? At what point does this break down and the mind is no longer a mind, or conscious? In general we can keep increasing the distance, until the frames or individual bits are stored in separate countries, or even galaxies. Given that the data points that make up this trajectory are already discrete and therefore there is a hard separation between frames, why does adding more distance between them make any difference?

It seems hard to believe that a mind-trajectory with data points that are physically scattered far apart in space would be alive or conscious in any meaningful sense, since the data points would in no way be related to each other any more. Could a single data point be part of more than one mind at once? Can any assortment of matter across the universe form a conscious mind?

Identifying the sources of confusion

I’ve tried to outline a few potential sources of confusion below, which may go some way towards explaining the differences in intuitions between these scenarios.

The CA transition function is important

In the functionalist view it is the actual processing of the information that matters. The difference between scenario 1 and 2 is that in scenario 1 we actually compute the CA transition function, whereas in scenario 2 we just load each 3D frame sequentially. However the CA has very simple rules which don’t necessarily have much correspondence with the actual physics /​ biology being simulated. The relevant computation in my view would be the morphological computation based on the state of the CA, not the simple transition rules, however obviously the transition rules underlie this.

A counterpoint: why could we not just store the instantaneous state of all the computer’s transistors while it is computing the transition function and store that in an (even larger) mind-trajectory? Then if we replay this on an even larger computer, which can emulate the first computer, do we not just arrive back at the same original problem?

What does “computing” actually mean?

It’s possible that some of the confusion lies in the lack of a precise definition of what we actually mean by “information processing” or “computing”. For example it might help to make a clearer distinction between the process of computing and the intermediate states of a computation as it proceeds.

Continuous vs. discrete computation

What difference does it make if the medium of computation is continuous in time and space (e.g. real world biology /​ physics) vs. discrete (e.g. a digital computer)? I’m still not sure whether this matters or not. It’s also highly plausible that real world physics and biology is actually discrete at the lowest level.

Conclusion

I remain stubbornly confused about these scenarios and can’t fully explain my intuitions of which ones would result in consciousness being present or not. I’m also not sure which of the potential sources of confusion actually matter. I think the question about the CA transition function is interesting, since on the one hand it seems like that is the “actual computation” part, but on the other hand some simple CA rules seem quite divorced from the type of information processing that is actually going on in a human mind and it seems odd that this would be the part required for consciousness.

It’s worth noting that I don’t have any background in philosophy of mind, so perhaps there are good answers to these questions that I’m simply not aware of!

Many thanks to @Guillaume Corlouer for providing useful feedback on a draft of this post.