Verifying is hard. Specifying what a FAI is well enough that you’ve even got a chance of having your Unspecified AI developing one is a whole ’nother sort of challenge.
Are there convenient acronyms for differentiating between Uncaring AIs and AIs actively opposed to human interests?
I was assuming that xamdam’s AGI will invent an FAI if people can adequately specify it and it’s possible, or at least it won’t be looking for ways to make things break.
There’s some difference between Murphy’s law and trying to make a deal with the devil. This doesn’t mean I have any certainty that people can find out which one a given AGI has more resemblance to.
I will say that if you tell the AGI “Make me an FAI”, and it doesn’t reply “What do you mean by Friendly?”, it’s either too stupid or too Unfriendly for the job.
Verifying is hard. Specifying what a FAI is well enough that you’ve even got a chance of having your Unspecified AI developing one is a whole ’nother sort of challenge.
Are there convenient acronyms for differentiating between Uncaring AIs and AIs actively opposed to human interests?
I was assuming that xamdam’s AGI will invent an FAI if people can adequately specify it and it’s possible, or at least it won’t be looking for ways to make things break.
There’s some difference between Murphy’s law and trying to make a deal with the devil. This doesn’t mean I have any certainty that people can find out which one a given AGI has more resemblance to.
I will say that if you tell the AGI “Make me an FAI”, and it doesn’t reply “What do you mean by Friendly?”, it’s either too stupid or too Unfriendly for the job.