Yes, continued development of AI seems unstoppable. But this brings up another very good point: if humanity cannot become a Singleton in our search for good egalitarian shared values, what is the chance of creating FAI? After years of good work in that direction and perhaps even success in determining a good approximation, what prevents some powerful secret entity like the CIA from hijacking it at the last minute and simply narrowing its objectives for something it determines is a “greater” good?
Our objectives are always better than the other guy’s, and while violence is universally despicable, it is fast, cheap, easy to program and the other guy (including FAI developers) won’t be expecting it. For the guy running the controls, that’s friendly enough. :-)
Yes, continued development of AI seems unstoppable. But this brings up another very good point: if humanity cannot become a Singleton in our search for good egalitarian shared values, what is the chance of creating FAI? After years of good work in that direction and perhaps even success in determining a good approximation, what prevents some powerful secret entity like the CIA from hijacking it at the last minute and simply narrowing its objectives for something it determines is a “greater” good?
Our objectives are always better than the other guy’s, and while violence is universally despicable, it is fast, cheap, easy to program and the other guy (including FAI developers) won’t be expecting it. For the guy running the controls, that’s friendly enough. :-)