I enjoyed reading that, in the same way that I enjoyed reading Roko’s Banned Post—I don’t believe it for a moment, but it stretches the mind a little. This one is much more metaphysical, and it also has an eschatological optimism that Roko’s didn’t. I think such optimism has no rational basis whatsoever, and in any case it means little to those unfortunates stuck in Hell to be told that Heaven is coming at the end of the computation, but unfortunately it can’t come any faster because of logical incompressibility. I’m thinking of Raymond Smullyan’s dialogue, in which God says that the devil is “the unfortunate length of time” that the process of “enlightenment” inevitably takes, and I think Tipler might make similar apologies for his Omega Point on occasion. All possible universes eventually reach the Omega Point (because, according to a sophistical argument of Tipler’s, space-time itself is inconsistent otherwise, so it’s logically impossible for this not to happen), so goodness and justice will inevitably triumph in every part of the multiverse, but in some of them it will take a really long time.
So, if I approach your essay anthropologically, it’s a mix of the very new cosmology and crypto-metaphysics (of Singularities in the multiverse, of everything as computation) with a much older thought-form—and of course you know this, having mentioned Neoplatonism—but I’d go further and say that the contents of this philosophy are being partly determined by a wishful thinking, which in turn is made possible by the fundamental uncertainty about the nature of reality. In other words, all sorts of terrible things may happen and may keep happening, but if you embrace Humean skepticism about induction, you can still say, nonetheless, reality might start functioning differently at any moment, therefore I have license to hope. In that case, uncertainty about the future course of mundane events provides the epistemic license for the leap of optimism.
Here, we have the new cosmological vision, of a universe (or multiverse) dominated by the rise of superintelligence in diverse space-time locations. It hasn’t happened locally yet, but it’s supposed to lie ahead of us in time. Then, we have the extra ingredient of acausal interaction between these causally remote (or even causally disjoint) superintelligences, who know about each other through simulation, reasoning, or other logically and mathematically structured explorations of the multiverse. And here is where the unreasonable optimism enters. We don’t know what these superintelligences choose to do, once they sound out the structure of the multiverse, but it is argued that they will come to a common, logically preordained set of values, and that these values will be good. Thus, the idea of a pre-established harmony, as in Tipler (and I think in Leibniz too, and surely many others), complete with a reason why the past and present are so unharmonious (our local singularity hasn’t happened yet), and also with an extra bit of hope that’s entirely new and probably doesn’t make sense: maybe the evil things that already happened will be cancelled out by reversing the computation—as if something can both have happened and could nonetheless be made to have never happened. Still, I bet Spinoza never thought of that one; all he could come up with was that evil is always an absence, that all things which actually exist are good, and so there’s nothing that’s actually bad.
The Stoics had a tendency to equate the order of nature with a cosmic Reason that was also a cosmic Good. Possibly Bertrand Russell was the one who pointed out that this is a form of power worship: just because this is the universal order, or this is the way that things have always been, does not in itself make it good. This point can easily be carried across to the picture of superintelligences arriving at their decisional equilibrium via mutual simulation: What exactly foreordains that the resulting equilibrium deserves the name of Good? Wouldn’t the concrete outcome depend on the distribution of superintelligence value systems arising in the multiverse—something we know nothing about—and on the resources that each superintelligence brings to the table of acausal trade and negotiation? It’s intriguing that even when confronted by such a bizarrely novel world-concept, the human mind is nonetheless capable, not only of interpreting it in a way originating from cultures which didn’t even know that the sun is a star, but of finding a way to affirm the resulting cosmology as good and as predestined to be so.
I have mentioned Russell’s reason for scorning the Stoic equation of the cosmic order with the cosmic good (it’s just worship of overwhelming force), but I will admit that, from an elemental perspective which values personal survival (and perhaps the personal gains that can come from siding with power), it does make sense to ask oneself what the values of the hypothetical future super-AI might be. That is, even if one scorns the beatific cyber-vision as wishful thinking, one might agree that a future super-purge of the Earth, conducted according to the super-AI’s value system, is a possibility, and attempt to shape oneself so as to escape it. But as we know, that might require shaping oneself to be a thin loop, a few centimeters long, optimized for the purpose of holding together several sheets of paper.
in any case it means little to those unfortunates stuck in Hell to be told that Heaven is coming at the end of the computation, but unfortunately it can’t come any faster because of logical incompressibility.
I acknowledge your points about not equating Goodness with Power, which is probably the failure mode of lusting for reflective consistency. (The lines of reasoning I go through in that link are pretty often missed by people who think they understand the nature of direction of morality, I think.) Maybe I should explicitly note that I was not at all describing my own beliefs, just trying to come up with a modern rendition of old-as-dirt Platonistic religionesque ideas. (Taoism is admirable in being more ‘complete’ and human-useful than the Big Good Metaphysical Attractor memeplexes (e.g. Neoplatonism), I think, though that’s just a cached thought.) I’ll go back over your comment again soon with a finer-toothed comb.
I enjoyed reading that, in the same way that I enjoyed reading Roko’s Banned Post—I don’t believe it for a moment, but it stretches the mind a little. This one is much more metaphysical, and it also has an eschatological optimism that Roko’s didn’t. I think such optimism has no rational basis whatsoever, and in any case it means little to those unfortunates stuck in Hell to be told that Heaven is coming at the end of the computation, but unfortunately it can’t come any faster because of logical incompressibility. I’m thinking of Raymond Smullyan’s dialogue, in which God says that the devil is “the unfortunate length of time” that the process of “enlightenment” inevitably takes, and I think Tipler might make similar apologies for his Omega Point on occasion. All possible universes eventually reach the Omega Point (because, according to a sophistical argument of Tipler’s, space-time itself is inconsistent otherwise, so it’s logically impossible for this not to happen), so goodness and justice will inevitably triumph in every part of the multiverse, but in some of them it will take a really long time.
So, if I approach your essay anthropologically, it’s a mix of the very new cosmology and crypto-metaphysics (of Singularities in the multiverse, of everything as computation) with a much older thought-form—and of course you know this, having mentioned Neoplatonism—but I’d go further and say that the contents of this philosophy are being partly determined by a wishful thinking, which in turn is made possible by the fundamental uncertainty about the nature of reality. In other words, all sorts of terrible things may happen and may keep happening, but if you embrace Humean skepticism about induction, you can still say, nonetheless, reality might start functioning differently at any moment, therefore I have license to hope. In that case, uncertainty about the future course of mundane events provides the epistemic license for the leap of optimism.
Here, we have the new cosmological vision, of a universe (or multiverse) dominated by the rise of superintelligence in diverse space-time locations. It hasn’t happened locally yet, but it’s supposed to lie ahead of us in time. Then, we have the extra ingredient of acausal interaction between these causally remote (or even causally disjoint) superintelligences, who know about each other through simulation, reasoning, or other logically and mathematically structured explorations of the multiverse. And here is where the unreasonable optimism enters. We don’t know what these superintelligences choose to do, once they sound out the structure of the multiverse, but it is argued that they will come to a common, logically preordained set of values, and that these values will be good. Thus, the idea of a pre-established harmony, as in Tipler (and I think in Leibniz too, and surely many others), complete with a reason why the past and present are so unharmonious (our local singularity hasn’t happened yet), and also with an extra bit of hope that’s entirely new and probably doesn’t make sense: maybe the evil things that already happened will be cancelled out by reversing the computation—as if something can both have happened and could nonetheless be made to have never happened. Still, I bet Spinoza never thought of that one; all he could come up with was that evil is always an absence, that all things which actually exist are good, and so there’s nothing that’s actually bad.
The Stoics had a tendency to equate the order of nature with a cosmic Reason that was also a cosmic Good. Possibly Bertrand Russell was the one who pointed out that this is a form of power worship: just because this is the universal order, or this is the way that things have always been, does not in itself make it good. This point can easily be carried across to the picture of superintelligences arriving at their decisional equilibrium via mutual simulation: What exactly foreordains that the resulting equilibrium deserves the name of Good? Wouldn’t the concrete outcome depend on the distribution of superintelligence value systems arising in the multiverse—something we know nothing about—and on the resources that each superintelligence brings to the table of acausal trade and negotiation? It’s intriguing that even when confronted by such a bizarrely novel world-concept, the human mind is nonetheless capable, not only of interpreting it in a way originating from cultures which didn’t even know that the sun is a star, but of finding a way to affirm the resulting cosmology as good and as predestined to be so.
I have mentioned Russell’s reason for scorning the Stoic equation of the cosmic order with the cosmic good (it’s just worship of overwhelming force), but I will admit that, from an elemental perspective which values personal survival (and perhaps the personal gains that can come from siding with power), it does make sense to ask oneself what the values of the hypothetical future super-AI might be. That is, even if one scorns the beatific cyber-vision as wishful thinking, one might agree that a future super-purge of the Earth, conducted according to the super-AI’s value system, is a possibility, and attempt to shape oneself so as to escape it. But as we know, that might require shaping oneself to be a thin loop, a few centimeters long, optimized for the purpose of holding together several sheets of paper.
So hell is a slow internet connection?
Hmm, maybe there’s something to this after all.
I acknowledge your points about not equating Goodness with Power, which is probably the failure mode of lusting for reflective consistency. (The lines of reasoning I go through in that link are pretty often missed by people who think they understand the nature of direction of morality, I think.) Maybe I should explicitly note that I was not at all describing my own beliefs, just trying to come up with a modern rendition of old-as-dirt Platonistic religionesque ideas. (Taoism is admirable in being more ‘complete’ and human-useful than the Big Good Metaphysical Attractor memeplexes (e.g. Neoplatonism), I think, though that’s just a cached thought.) I’ll go back over your comment again soon with a finer-toothed comb.