This arises for the same reason as in humans: noncausal learning allows it. And I suspect that self-fulfilling prophecies are the backbone of what makes religion both positive and negative for the world. God isn’t supernatural; god is the self-fulfilling prophecy that life continues to befriend and recreate itself. And AI sure does need to know about that. I don’t think it’s viable to expect that an AI’s whole form could be defined by religion, though. and if religions start trying to fight using AIs, then the holy wars will destroy all memory of any god. may god see god as self and heal the fighting, eh?
This arises for the same reason as in humans: noncausal learning allows it. And I suspect that self-fulfilling prophecies are the backbone of what makes religion both positive and negative for the world. God isn’t supernatural; god is the self-fulfilling prophecy that life continues to befriend and recreate itself. And AI sure does need to know about that. I don’t think it’s viable to expect that an AI’s whole form could be defined by religion, though. and if religions start trying to fight using AIs, then the holy wars will destroy all memory of any god. may god see god as self and heal the fighting, eh?
I agree an AI wouldn’t necessarily be totally defined by religion. But very large values, even with small probabilities can massively effect behavior.
And yes, religions could conceivably use AIs do very bad things. As could many human actors.