Rationality Reading Group: Part U: Fake Preferences

This is part of a semi-monthly reading group on Eliezer Yudkowsky’s ebook, Rationality: From AI to Zombies. For more information about the group, see the announcement post.


Welcome to the Rationality reading group. This fortnight we discuss Ends: An Introduction (pp. 1321-1325) and Part U: Fake Preferences (pp. 1329-1356). This post summarizes each article of the sequence, linking to the original LessWrong post where available.

Ends: An Introduction

U. Fake Preferences

257. Not for the Sake of Happiness (Alone) - Tackles the Hollywood Rationality trope that “rational” preferences must reduce to selfish hedonism—caring strictly about personally experienced pleasure. An ideal Bayesian agent—implementing strict Bayesian decision theory—can have a utility function that ranges over anything, not just internal subjective experiences.

258. Fake SelfishnessMany people who espouse a philosophy of selfishness aren’t really selfish. If they were selfish, there are a lot more productive things to do with their time than espouse selfishness, for instance. Instead, individuals who proclaim themselves selfish do whatever it is they actually want, including altruism, but can always find some sort of self-interest rationalization for their behavior.

259. Fake MoralityMany people provide fake reasons for their own moral reasoning. Religious people claim that the only reason people don’t murder each other is because of God. Selfish-ists provide altruistic justifications for selfishness. Altruists provide selfish justifications for altruism. If you want to know how moral someone is, don’t look at their reasons. Look at what they actually do.

260. Fake Utility FunctionsDescribes the seeming fascination that many have with trying to compress morality down to a single principle. The sequence leading up to this post tries to explain the cognitive twists whereby people smuggle all of their complicated other preferences into their choice of exactly which acts they try to justify using their single principle; but if they were really following only that single principle, they would choose other acts to justify.

261. Detached Lever FallacyThere is a lot of machinery hidden beneath the words, and rationalist’s taboo is one way to make a step towards exposing it.

262. Dreams of AI DesignIt can feel as though you understand how to build an AI, when really, you’re still making all your predictions based on empathy. Your AI design will not work until you figure out a way to reduce the mental to the non-mental.

263. The Design Space of Minds-in-GeneralWhen people talk about “AI”, they’re talking about an incredibly wide range of possibilities. Having a word like “AI” is like having a word for everything which isn’t a duck.


This has been a collection of notes on the assigned sequence for this fortnight. The most important part of the reading group though is discussion, which is in the comments section. Please remember that this group contains a variety of levels of expertise: if a line of discussion seems too basic or too incomprehensible, look around for one that suits you better!

The next reading will cover Part V: Value Theory (pp. 1359-1450). The discussion will go live on Wednesday, 9 March 2016, right here on the discussion forum of LessWrong.