See the first post of this series of comments. In brief I’m hacking the comment space of my own publication, as a safer place (less exposed) to discuss hot topics that can generate feedback that would make me go away. The guest is Karl and you’re welcome to join if you’re ok with the courtesy policy written in the first comment. If not please send me a pm and I’m sure we can try to agree on some policy for your own subspace here.
[The quote is from me, as the parody I tend to perceive. Yes, I fully agree an agent with conflicted preference is the opposite of a paperclip maximiser. Would we also agree that a random set of preference is more likely self-contradictory and that would have obvious impact on any ASI trying to guess my password?]
I don’t know where the quote is a quote from.
Conflicted preferences are obviously impactful on effectiveness. An agent with conflicted preferences is the opposite of a paperclipper.
See the first post of this series of comments. In brief I’m hacking the comment space of my own publication, as a safer place (less exposed) to discuss hot topics that can generate feedback that would make me go away. The guest is Karl and you’re welcome to join if you’re ok with the courtesy policy written in the first comment. If not please send me a pm and I’m sure we can try to agree on some policy for your own subspace here.
[The quote is from me, as the parody I tend to perceive. Yes, I fully agree an agent with conflicted preference is the opposite of a paperclip maximiser. Would we also agree that a random set of preference is more likely self-contradictory and that would have obvious impact on any ASI trying to guess my password?]