A loving and nurturing home is not always enough even for humans, some are just born psychopaths. A flaw in the AGI design (“genetic makeup”) may result in a psychopathic AGI, and even if the probability of it is low, it is probably not something to chance, given the potential ramifications.
Not necessarily easier, but certainly more promising. The easiest way to get a human based AI is to upload a human, which is a rather simple compared to building a rational, provably goal-following AI. Uploading a human will just get you a superpowerful dictator that gets corrupted by power.
Why more promising? It seems to me that we either accept the risk of trying to teach an artificial person to be good, or we accept the risk of uploading someone we expect to remain good, or of letting someone we hope to be good build a helpful psychopath. After all, if that programmer has a faulty conception of the human good then they’ll create a monster, even if it is a ‘friendly’ one. In every case, we have to rely on the uncertain integrity of the ethical person.
accept the risk of trying to teach an artificial person to be good
It’s not a person, it’s an optimisation process. Don’t anthropomorphise AI. Your are right that the risk is large.
or we accept the risk of uploading someone we expect to remain good
Which we know to be nearly impossible with few ways to improve the chances.
or of letting someone we hope to be good build a helpful psychopath. After all, if that programmer has a faulty conception of the human good then they’ll create a monster,
You are not familiar with the current plan and the reasoning behind it. Go read CEV. Also metaethics, because you seem to take it as possible that a human could have a good enough conception of value to program it.
In every case, we have to rely on the uncertain integrity of the ethical person.
fallacy of grey. Some methods are much more promising than others, even if all are uncertain.
Some methods are much more promising than others, even if all are uncertain.
I certainly agree with you there. I have some familiarity with CEV (though it’s quite technical and I don’t have much background in decision theory), but on the basis of that familiarity, I’m under the impression that creating an artificial person and teaching it to be ethical is the safest and most reliable way to accomplish our goal of surviving the singularity. But I haven’t argued for that, of course.
A loving and nurturing home is not always enough even for humans, some are just born psychopaths. A flaw in the AGI design (“genetic makeup”) may result in a psychopathic AGI, and even if the probability of it is low, it is probably not something to chance, given the potential ramifications.
I don’t know if I’d use that term, it implies that if we are just a bit more careful, we can avoid it.
It is almost certain that any AI we create is going to be a psychopath (even a friendly one), the trick is how to make it our psychopath.
I take it the idea here is that this is easier than creating a person and teaching them to be ethical?
Not necessarily easier, but certainly more promising. The easiest way to get a human based AI is to upload a human, which is a rather simple compared to building a rational, provably goal-following AI. Uploading a human will just get you a superpowerful dictator that gets corrupted by power.
Why more promising? It seems to me that we either accept the risk of trying to teach an artificial person to be good, or we accept the risk of uploading someone we expect to remain good, or of letting someone we hope to be good build a helpful psychopath. After all, if that programmer has a faulty conception of the human good then they’ll create a monster, even if it is a ‘friendly’ one. In every case, we have to rely on the uncertain integrity of the ethical person.
It’s not a person, it’s an optimisation process. Don’t anthropomorphise AI. Your are right that the risk is large.
Which we know to be nearly impossible with few ways to improve the chances.
You are not familiar with the current plan and the reasoning behind it. Go read CEV. Also metaethics, because you seem to take it as possible that a human could have a good enough conception of value to program it.
fallacy of grey. Some methods are much more promising than others, even if all are uncertain.
I certainly agree with you there. I have some familiarity with CEV (though it’s quite technical and I don’t have much background in decision theory), but on the basis of that familiarity, I’m under the impression that creating an artificial person and teaching it to be ethical is the safest and most reliable way to accomplish our goal of surviving the singularity. But I haven’t argued for that, of course.