once I obtain or make a paperclip, I move it to the safe zone, where I’ll pretty much have it forever.
How much modeling of the universe have you done, or will you do, to ensure that this is the case? How much effort will you put into creating and defending a suitable safe zone? Will you search the stars for invasive self-replicators (green goo) who may be a threat to your paperclip cache, or will you postpone active offense/defense until such time as you receive a credible threat?
It’s a work in progress.
Humankind has been searching for answers for thousands of years. Do you think you’ll have a better time discovering ultimate paperclippiness? Or will you be content with a lesser understanding (“I did my best”)?
1) I will do all things such that they maximize expected paperclip content of the universe, trading off smaller paperclip quantities for larger ones. I can’t express a more specific algorithm than that without knowing the particulars of the situation.
2) I will do much better than humans at finding the ultimate morality* of the universe because I can spend all my resources to make perfect copies of myself that share my values and update knowledge and reflective value equilibria among each other, rather than having to pursue other values like “signaling” (except when I post here, of course).
* Note I said ultimate morality, not ultimate paperclippiness. There’s a strange tendency here to equate everything I like with “paperclippiness”, and everything humans like with “morality” or “friendliness”. Wrong: we all want morality (humans and clippys), we just disagree about what counts as moral. To use the alternate naming scheme of calling my values “paperclippy” and human values “morality” is to assume away the problem. (Certainly, I sometimes fall into the trap of calling good things, either for humans or clippys, “paperclippy”—see here—but this is just when I’m being imprecise with my phrasings.)
Thanks for answering my questions. I think I have a better understanding of your values now.
My empathy circuits are lighting up, telling me I should buy a box of paperclips and keep them safe for you. And that I should put them on display as a warning to myself.
How much modeling of the universe have you done, or will you do, to ensure that this is the case? How much effort will you put into creating and defending a suitable safe zone? Will you search the stars for invasive self-replicators (green goo) who may be a threat to your paperclip cache, or will you postpone active offense/defense until such time as you receive a credible threat?
Humankind has been searching for answers for thousands of years. Do you think you’ll have a better time discovering ultimate paperclippiness? Or will you be content with a lesser understanding (“I did my best”)?
1) I will do all things such that they maximize expected paperclip content of the universe, trading off smaller paperclip quantities for larger ones. I can’t express a more specific algorithm than that without knowing the particulars of the situation.
2) I will do much better than humans at finding the ultimate morality* of the universe because I can spend all my resources to make perfect copies of myself that share my values and update knowledge and reflective value equilibria among each other, rather than having to pursue other values like “signaling” (except when I post here, of course).
* Note I said ultimate morality, not ultimate paperclippiness. There’s a strange tendency here to equate everything I like with “paperclippiness”, and everything humans like with “morality” or “friendliness”. Wrong: we all want morality (humans and clippys), we just disagree about what counts as moral. To use the alternate naming scheme of calling my values “paperclippy” and human values “morality” is to assume away the problem. (Certainly, I sometimes fall into the trap of calling good things, either for humans or clippys, “paperclippy”—see here—but this is just when I’m being imprecise with my phrasings.)
Thanks for answering my questions. I think I have a better understanding of your values now.
My empathy circuits are lighting up, telling me I should buy a box of paperclips and keep them safe for you. And that I should put them on display as a warning to myself.
A warning of what???
How morality can go awry.
I already have a framed print of Hug Bot on my wall.