I am playing the gatekeeper for the first round, taking place on January 22nd. I commit to not letting the AI out of the box. I am more than 80 percent confident that no human can get past me, and more than 30% confident that a transhuman could not get past me.
EDIT: The AI remained in the box, so upvote this comment to +25 and downvote lessdazed’s child comment to −25. However, the session finished inconclusively, with my decision dependent on factors that had not been set beforehand. I recommend that for future sessions, the parties agree to the circumstances of the AI’s creation, how much the AI knows about the circumstances of its creation, and the gatekeeper’s prior P(the AI is Friendly). My own P(no human can get past me|plausible prearranged backstory) is now 85%, and my P(no transhuman AI could get past me|representative plausible backstory) is now less than 10%. If the game had been real, there’s a good chance I’d have lost.
… more than 30% confident that a transhuman could not get past me.
I could see values > 95% (tape my eyes shut, put in ear-plugs and type “nonononono” for the duration), or values < 5% (actually speak with the transhuman intelligence). But values in the middle seem to indicate that you think you’ll have a conversation with the transhuman intelligence, evaluate the arguments and be left more-or-less on the fence.
It just seems to me that there’s this tiny target of intelligence that is exactly the correct amount smarter than you to make you indecisive, and beyond that it will manage to convince you overwhelmingly.
Values in the middle indicate that I’ll have a conversation and probably not budge, with a chance of being totally convinced. But I am now convinced that the whole idea of boxing is stupid. Why would I honestly have a conversation? Why would I run a transhuman AI that I didn’t want to take over the world? What could I learn from a conversation that I wouldn’t already know from the source code. other than that it doesn’t immediately break? And why would I need to check that the AI doesn’t immediately break, unless I wanted to release it?
Why would I honestly have a conversation? Why would I run a transhuman AI that I didn’t want to take over the world?
Because you’d want to know how to cure cancer, how to best defeat violent religious fundamentalism, etc, etc. If you want to become President, the AI may need to teach you argumentation techniques. And so forth.
Values in the middle indicate that I’ll have a conversation and probably not budge, with a chance of being totally convinced.
Ah, so it’s more like the probability that the intelligence in the box is over the threshold required to convince you of something. That makes sense.
But I am now convinced that the whole idea of boxing is stupid.
Agreed. Everything you said, plus: if you think there’s a chance that your boxed AI might be malicious/unFriendly, talking to it has to be one of the stupidest things you could possibly do...
It was agreed to halve the wager from 50 karma to 25 due to the specific circumstances concluding the role-play in which that the outcome depended on variables that hadn’t been specified, but if that sounds contemptible to you downvote all the way to −50.
I am playing the gatekeeper for the first round, taking place on January 22nd. I commit to not letting the AI out of the box. I am more than 80 percent confident that no human can get past me, and more than 30% confident that a transhuman could not get past me.
EDIT: The AI remained in the box, so upvote this comment to +25 and downvote lessdazed’s child comment to −25. However, the session finished inconclusively, with my decision dependent on factors that had not been set beforehand. I recommend that for future sessions, the parties agree to the circumstances of the AI’s creation, how much the AI knows about the circumstances of its creation, and the gatekeeper’s prior P(the AI is Friendly). My own P(no human can get past me|plausible prearranged backstory) is now 85%, and my P(no transhuman AI could get past me|representative plausible backstory) is now less than 10%. If the game had been real, there’s a good chance I’d have lost.
I could see values > 95% (tape my eyes shut, put in ear-plugs and type “nonononono” for the duration), or values < 5% (actually speak with the transhuman intelligence). But values in the middle seem to indicate that you think you’ll have a conversation with the transhuman intelligence, evaluate the arguments and be left more-or-less on the fence.
It just seems to me that there’s this tiny target of intelligence that is exactly the correct amount smarter than you to make you indecisive, and beyond that it will manage to convince you overwhelmingly.
Anyway, good luck :)
Values in the middle indicate that I’ll have a conversation and probably not budge, with a chance of being totally convinced. But I am now convinced that the whole idea of boxing is stupid. Why would I honestly have a conversation? Why would I run a transhuman AI that I didn’t want to take over the world? What could I learn from a conversation that I wouldn’t already know from the source code. other than that it doesn’t immediately break? And why would I need to check that the AI doesn’t immediately break, unless I wanted to release it?
Because you’d want to know how to cure cancer, how to best defeat violent religious fundamentalism, etc, etc. If you want to become President, the AI may need to teach you argumentation techniques. And so forth.
Ah, so it’s more like the probability that the intelligence in the box is over the threshold required to convince you of something. That makes sense.
Agreed. Everything you said, plus: if you think there’s a chance that your boxed AI might be malicious/unFriendly, talking to it has to be one of the stupidest things you could possibly do...
Remember this!
Just say ‘No’!
The AI remained in the box.
It was agreed to halve the wager from 50 karma to 25 due to the specific circumstances concluding the role-play in which that the outcome depended on variables that hadn’t been specified, but if that sounds contemptible to you downvote all the way to −50.