AIs aren’t Gods, with time and care and lots of preparation reading their thoughts should be doable.
Sure, possible. Just a lot harder than creating an FAI to do it for you—especially when the AI has an incentive to obfuscate.
Why are you so confident that the first version of FAI we make will be safe? Doing both is safest and seems like it would be worth the investment.
Why are you so confident that the first version of FAI we make will be safe?
I’m not. I expect it to kill us all with high probability (which is nevertheless lower than the probability of obliteration if no FAI is actively attempted.)
Sure, possible. Just a lot harder than creating an FAI to do it for you—especially when the AI has an incentive to obfuscate.
Why are you so confident that the first version of FAI we make will be safe? Doing both is safest and seems like it would be worth the investment.
I’m not. I expect it to kill us all with high probability (which is nevertheless lower than the probability of obliteration if no FAI is actively attempted.)