In order to get that provably friendly thing to work
Again, I think “provably friendly thing” mischaracterizes what MIRI thinks will be possible.
I’m not sure exactly what you’re saying in the rest of your comment. Have you read the section on indirect normativity in Superintelligence? I’d start there.
Given the apparent misconceptions about MIRI’s work even among LWers, it seems like you need to write a Main post clarifying what MIRI does and does not claim, and does and does not work on.
Again, I think “provably friendly thing” mischaracterizes what MIRI thinks will be possible.
From what I can gather, there’s still supposed to be some kind of proof, even if it’s just the mathematical kind where you’re not really certain because there might be an error in it. The intent is to have some sort of program that maximizes utility function U, and then explicitly write the utility function as something along the lines of “do what I mean”.
Have you read the section on indirect normativity in Superintelligence? I’d start there.
I’m not sure what you’re referring to. Can you give me a link?
Typo?
Again, I think “provably friendly thing” mischaracterizes what MIRI thinks will be possible.
I’m not sure exactly what you’re saying in the rest of your comment. Have you read the section on indirect normativity in Superintelligence? I’d start there.
Given the apparent misconceptions about MIRI’s work even among LWers, it seems like you need to write a Main post clarifying what MIRI does and does not claim, and does and does not work on.
Fixed.
From what I can gather, there’s still supposed to be some kind of proof, even if it’s just the mathematical kind where you’re not really certain because there might be an error in it. The intent is to have some sort of program that maximizes utility function U, and then explicitly write the utility function as something along the lines of “do what I mean”.
I’m not sure what you’re referring to. Can you give me a link?
Superintelligence is a recent book by Nick Bostrom