I’m interested in axiology/value/utility—what things are valuable and why and when and how and to whom?
ihatenumbersinusernames7
[Question] Whence unchangeable values?
I agree, and would also point out that since:
By contrast, real friendship has to be (1A)
...this intrinsic value [friendship] is in place and leads to cooperation (an instrumental value).
Very different than the model that says: competition → cooperation → the value [friendship].
There is a science of control systems that doesn’t require the system being controlled to be indeterministic
Indeterministic is a loaded word. Certainly we don’t believe our actions to be random, but I maintain that the question before compatibilists/semicompatibilists (which I hoped this post would address but IMO doesn’t) is why seeing free will as a human construct is meaningful. For example:
Am I suggesting that if an alien had created Lenin, knowing that Lenin would enslave millions, then Lenin would still be a jerk? Yes, that’s exactly what I’m suggesting. The alien would be a bigger jerk.
So if I create an AI that steals money, I am the greater jerk but the AI is also a jerk?
It seems to me that if you create an agent and put the agent into an environment where X will happen, you have exonerated the agent in regard to X. Maybe this just means I’m not a compatibilist, but I still don’t see a good argument here for compatibilism/semicompatibilism.
I’ve though about this some more and I think what you mean (leaving aside physical and homeostatic values and focusing on organism-wide values) is that, even if we define our “terminal value” as I have above, whence the basket of goods that mean “happiness/flourishing” to me?
Again I think the answer is evolution plus something...some value drift (that as you say, the Shard Theory people are trying to figure out). Is there a place/post you’d recommend to get up to speed on that? The wikitag is a little light on details (although I added a sequence that was a good starting place). https://www.lesswrong.com/w/shard-theory
“You have to learn them from other people , and their attitudes praise and blame are how they get imprinted into brains, when they are.”
By my lights, this skirts the issue too. Yudkowsky described the deterministic nature of adhering to moral norms. You’re talking about where moral norms come from. But the moral responsibility question is, do we in any sense have control over (and culpability for) our moral actions?
So, not yet knowing the output of the deterministic process that is myself, and being duty-bound to determine it as best I can, the weight of moral responsibility is no less.
The reason you would be thus “duty-bound” seems to be the crux of this whole post, and I don’t see one provided.
ihatenumbersinusernames7′s Shortform
To extract out the terminal values we have to inspect this mishmash of valuable things, trying to figure out which ones are getting their value from somewhere else.
What if all the supposed “terminal values” are actually instrumental values from...some more terminal value? Sure “less violence” is more of a terminal value than “fewer guns,” but is there a still more terminal value? Less violence why?
My answer: Because I want to live in a world of happiness/flourishing. https://www.lesswrong.com/posts/DJeBQrBSqFYqBaeQW/rock-bottom-terminal-value
Main question: am I right?
More interesting follow-up: if that is the root, bottom, terminal value driving us humans, did it come from evolution, or is it simply a feature of conscious beings?
Thanks, and yes evolution is the source of many values for sure...I think the terminal vs instrumental question leads in interesting directions. Please let me know how this sits with you!
Though I am an evolved being, none of your examples seem to be terminal values for me the whole organism. Certainly there are many systems within me, and perhaps we could describe them as having their own terminal values, which in part come from evolution as you describe. My metabolic system’s terminal value surely has a lot to do with regulating glucose. My reproductive system’s terminal value likely involves sex/procreation. (But maybe even these can drift, like when a cell becomes cancerous, it seems its terminal value changes.)
But to me as a whole, these values (to the extent which I hold them at all) are instrumental. Sure I want homeostasis, but I want it because I want to live (another instrumental value), and I want to live because I want to be able to pursue my terminal value of happiness/flourishing. Other values that my parts exhibit (like reproduction) I the whole might reject even as an instrumental value, heck I might even subvert the mechanisms afforded by my reproductive system for my own happiness/flourishing.
Also for my terminal value for happiness/flourishing, did that come from evolution? Did it start out as survival/reproduction and drift a bit? Or is there something special about systems like me (which are conscious of pleasure/pain/etc) that just by their nature they desire happiness/flourishing, the way 2+2=4 or the way a triangle has 3 sides? Or...other?
And lastly does any of this port to non-evolved beings like AIs?
Do you think the “basket of goods” (love the pun) could be looked at as instrumental values that derive from the terminal value (desiring happiness/flourishing)?
I don’t understand shard theory well enough to critique it, but is there a distinction between terminal and instrumental within shard theory? Or are these concepts incompatible with shard theory?
(Maybe some examples from the “basket of goods” would help.)
Rock bottom terminal value
I had bookmarked this post as fascinating (but I claim no first-hand knowledge here): https://www.lesswrong.com/posts/BgBJqPv5ogsX4fLka/the-mind-body-vicious-cycle-model-of-rsi-and-back-pain
Looks potentially pretty relevant. Is that the kind of thing you’re looking for?
people including alignment researchers just seem more confident about their own preferred solution to metaethics, and comfortable assuming their own preferred solution is correct as part of solving other problems, like AI alignment or strategy. (E.g., moral anti-realism is true, therefore empowering humans in straightforward ways is fine as the alignment target can’t be wrong about their own values.)
Obviously committed anti-realists would be right not to worry—if they’re correct! But I agree with you, we shouldn’t be overconfident in our metaethics...which makes me wonder, do you really think metaethics can be “solved?”
Secondly, even if it were solved (and to avoid the anti-realist apathy, let’s assume moral realism is true), how do you think that would help with alignment? Couldn’t the alignment-target simply say, “this is true, but I don’t care, as it doesn’t help me achieve my goals?” Saying “1+1=2, but I’m going to act as if it equals 3″ might keep you from achieving your goal. Saying, “stealing is wrong, but I would really like to have X” might actually help you achieve your goal.
A worthwhile cosmopolitanism should probably fight Moloch in this regard and extend its own agency to things which do not wield the tools of autopoesis themselves.
[warning: anthropomorphisation of autopoetic processes]
*autopoiesis, and autopoietic (brought to you by the pedantic spelling police)
Hi! Like others on this thread, I’m a long time reader who’s finally created an account to try to join the discussion. I’m curious, if I comment on a 15 year old article or something, is anyone likely to see that? I love browsing around the Concepts pages, but are comments there (.e.g.) likely to be seen?
My intuition is that comments on recent trending articles are more likely to get engagement, but can anyone confirm or deny or give suggestions on the best ways/places to engage?
Thanks!
Imagine you’re immortal. Would you rather get a dust speck in your eye for 3^^^3 straight days, or endure 50 years of torture, never to be bothered by a dust speck again?
Going with dust specks is the obvious choice to me. The way I see it, life is still very much worth living when I’m dealing with a dust speck. Torture might make me wish for death in the first day...especially with the knowledge that 18,261 days were forthcoming.
I can’t tell if:
a) I dislike dust specks less than you
b) I dislike torture more than you
c) Aggregating pain/discomfort/suffering has weird results. (e.g. perhaps 5 minutes of torture is immeasurably more than 60x as bad as 5 seconds of torture.)
or d) there’s something else I’m missing.
I think this is very relevant to the complexity of values and how difficult it is for humans to grasp their own utility functions. Cards on the table, the aggregation/math issues make me unable to embrace consequentialism...but if I’m overlooking something I’d love to be shown the light.
The only reason I believe myself to have “objective” moral worth is because I have subjective experience. Maybe more wordplay than irony, but submitted for your amusement.