Relatively new here (hi) and without adequate ability to warp spacetime so that I may peruse all that EY has written on this topic, but am still wondering—Why pursue the idea that morality is hardwired, or that there is an absolute code of what is right or wrong?
Thou shall not kill—well, except is someone is trying to kill you.
To be brief—it seems to me that 1) Morality exists in a social context. 2) Morality is fluid, and can change/has changed over time. 3) If there is a primary moral imperative that underlies everything we know about morality, it seems that that imperative is SURVIVAL, of self first, kin second, group/species third.
Empathy exists because it is a useful survival skill. Altruism is a little harder to explain.
But what justifies the assumption that there IS an absolute (or even approximate) code of morality that can be hardwired and impervious to change?
The other thing I wonder about when reading EY on morality is—would you trust your AI to LEARN morality and moral codes in the same way a human does? (See Kohlberg’s Levels of Moral Reasoning.)Or would you presume that SOMETHING must be hardwired? If so, why?
(EY—Do you summarize your views on these points somewhere? Pointers to said location very much appreciated.)
Why not divide the pie according to who will ultimately put the pie to the best use? If X and Y intend to take a nap after eating the pie, but Z is willing to plant a tree, wouldn’t the best outcome for the pie favor Z getting more?
Before you dismiss the analogy, consider this—what if the pie was $1800.00 that none of the three had earned? What if the $1800.00 had been BORROWED with a certain expectation of its utility? Should X, Y, and Z each get $600.00, even though there is no stipulation as to what each of them must DO with that money? If X intends to save his portion, and Y intends to pay down debt, but Z will spend the money though it may not be is HIS best interests to do so, should he still only get an equal portion, even though his actions with his share best accomplish the purpose of the money?
If we return to pie, you may now see that pie represents potential action (as one of the earlier commenters who mentioned carbs noted). Instead of arguing for division based on merit for PAST actions/attributes (as mentioned by another commenter), why not argue for division based on merit of INTENDED actions? Who provides the best return on the invested carbs? Why assume that ‘fair’ division should reflect mere existence? Why can’t ‘fairness’ include an evalutation of potential return?
This may simply deflect the argument of ‘fairness’ to one wherein ‘best return’ must be determined with regard to each individual and the group as a whole. If Y gets no shade from the tree Z plants, then perhaps her ‘best return’ might be a contented nap.
The ratio of productive and beneficial action, as a function of the input (pie), calculated across time (a tree has longer benefits than an immediate nap) seems to also be a ‘fair’ way to divide the pie.