Consciousness as recurrence, potential for enforcing alignment?
An interesting comment from Max Tegmark in Lex Friedman podcast on an idea attributed to Giulio Tononi (I haven’t found publication source for this idea); that consciousness and perhaps goal-formation is a function of recurrence. Recurrence being the feeding back of output to input in a once-through neural net such as the cortical columns found in our brains.
The idea seems (to me) to be that goal setting agency and intent come about as a function of feeding back conceptions and ideas developed from passes-through of our cortical columns—similar to a single prompt-output cycle in an LLMs, that are filtered for usefulness against our goals. But that these basic pass-throughs are innately unmotivated, and perhaps near-deterministic. The outputs from a prompt are then a point where you can safely get a snapshot of thinking process without fear of deceit.
If correct in conception this may offer a useful means for enforcement of alignment—by editorial control of that feedback-to-prompt, pruning out anything (for example) that isn’t in accordance with human-controlled preferences and tasking. Maybe we can sculpt safe mono-mania of the design-me-a-better-x, or solve-problem-y type, or more generally does this align with human preferences (Along the lines of Asimov’s 4-Laws conception), while still accessing the creative potential and greater scope of understanding of an AGI or even ASI.
If you mean the Tononi pet model of consciousness https://en.wikipedia.org/wiki/Integrated_information_theory, it has been debunked a long time ago, see https://scottaaronson.blog/?p=1799.
Oh I don’t know about debunked. certainly shown to be an insufficient picture, yes. but research continues to point towards critical brain hypothesis being very important, which relates informationally to IIT. eg a random paper on that research subgraph https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8851554/
I mean, it could be a necessary condition, but by no means sufficient.
yeah okay fair enough.
This seems like a solid attempt at understanding the concepts, but I think the true complexity of these concepts is much greater than you’re anticipating. I simply dropped your post into GPT4′s input—this isn’t that great of a summary tbh but it’s a start:
Recurrence is a requirement of Turing completeness, so if you assume that consciousness can be modeled as some kind of universal Turing machine then it’s required.