Credence polls for 26 claims from the 2019 Review

This post is a whirlwind tour of claims made in the LessWrong 2019 Review. In some cases, the claim is literally quoted from the post. In others, I have tried operationalising it into something more falsifiable. For example:

Book Review: The Secret of Our Success

10%20%30%40%50%60%70%80%90%
Cultural Intelligence Hypothesis: humans evolved big brains in order to be able to maintain [complicated, detailed, arbitrary-seeming cultural knowledge like 20+ step Inuit seal hunting techniques]. Everything that separates us from the apes is part of an evolutionary package designed to help us maintain this kind of culture, exploit this kind of culture, or adjust to the new abilities that this kind of culture gave us.
10%20%30%40%50%60%70%80%90%
Machiavellian Intelligence Hypothesis: humans evolved big brains in order to succeed at social manuevering and climbing dominance hierarchies.
10%20%30%40%50%60%70%80%90%
For most of history, a human attempting to use reasoning to do things like cooking, crafting, or planning (instead of using inherited cultural heuristics, like omens or folk wisdom), would have been substantially worse off, and faced a major increase in their risk of death (without a commensurate increase in life quality).

Overall, treat the claims in this post more like polls, and less like the full-blown forecasting questions you’d find on Metaculus or PredictIt. (The latter have extremely high bars for crisp definitions.) They point in a direction, but don’t completely pin it down.

Overall, this is an experiment. I’m trying to find interesting ways for people to relate to the Review.

Maybe speeding through these questions gets you thinking good thoughts, that you can then turn into full-blown reviews? Maybe others’ answers allow you to find a discussion partner who disagrees on a core question? Maybe the data will be useful in the voting phase?

We’ll see!

Feel free to leave a comment about how you found the experience, if you want.

If you want to discuss the questions with others over a call, you can do so during the Review forecasting sessions we’re organising this weekend (January 9-10).

If you want to hide other user’s predictions until you’ve made your own, here’s how to do that:

Click “Edit Settings”
Go to “Site customizations”
Press “Hide other users’ Elicit predictions until I have predicted myself”

Make More Land

Making more land out of the about 50mi^2 shallow water in the San Francisco Bay, South of the Dumbarton Bridge, would…

10%20%30%40%50%60%70%80%90%
...be an environmental disaster.
10%20%30%40%50%60%70%80%90%
...create buildings where the expected yearly damage from earthquakes (both in terms of reduced quality of life and property destroyed) is >1.5x that of nearby buildings on old land.
10%20%30%40%50%60%70%80%90%
...significantly worsen traffic in San Francisco.
10%20%30%40%50%60%70%80%90%
...all-things-considered, be good for the world.

Why Wasn’t Science Invented in China?

10%20%30%40%50%60%70%80%90%
The modern Scientific Revolution occurred in Europe between the 16th and 18th Centuries. Why did it not happen in China? Historian Toby Huff claims the reason is that China was unable to produce modern science primarily because a lack of the requisite intellectual freedom. Was he basically correct?

The Strategy-Stealing Assumption

10%20%30%40%50%60%70%80%90%
The strategy-stealing assumption is "a good enough approximation that we can basically act as if it’s true". That is, for any strategy an unaligned AI could use to influence the long-run future, there is an analogous strategy that a similarly-sized group of humans can use in order to capture a similar amount of flexible influence over the future. By “flexible” is meant that humans can decide later what to do with that influence  (which is important since humans don’t yet know what we want in the long run).

Becoming the Pareto-best in the World

The Hard Work of Translation

The Forces of Blandness and the Disagreeable Majority

Bioinfohazards

Two explanations for variation in human abilities

Reframing Impact

These questions are quite technical, and might be hard to answer if you’re unfamiliar with the terminology used in TurnTrout’s sequence on Impact Measures.

---

(Note that when you answer questions in this summary post, and it will automatically update the prediction questions that I have linked in comments on each individual post. The distributions will later be visible when users are voting to rank the posts.)