I disagree. The point of the post is not that these theories were on balance equally plausible during the Renaissance. It’s written so as to overemphasize the evidence for geocentrism, but that’s mostly to counterbalance standard science education.
In fact, one my key motivations for writing it—and a point where I strongly disagree with people like Kuhn and Feyerabend—is that I think heliocentrism was more plausible during that time. It’s not that Copernicus, Kepler Descartes and Galileo were lucky enough to be overconfident in the right direction, and really should just have remained undecided. Rather, I think they did something very right (and very Bayesian). And I want to know what that was.
Thank you! I was quite nervous about posting but am very happy with the reception, and strongly update towards how remarkable a community LW2.0 might become (in terms of how welcoming it is of truth-seeking discussion and how constructively it forwards it).
Reading your comment, I’d update towards the relative importance of mathematical aesthetic compared to physical plausibility in finding true theories. I only want to believe in luck as a last resort. You seem to be making the “opposite” update. Is this correct? And, if it is, why do you update that way?
In order for me to update on this it would be great to have concrete examples of what does and does not consistute “nontrivial theoretical insights” according to you and Paul.
E.g. what was the insight from the 1980s? And what part of the AG(Z) architecture did you initially consider nontrivial?
I’m looking forward to reading that post.
Yes, it seems right that gradient descent is the key crux. But I’m not familiar with any efficient way of doing it that the brain might implement, apart from backprop. Do you have any examples?
What a great post! Very readable, concrete, and important. Is it fair to summarize it in the following way?
A market/population/portfolio of organizations solving a big problem must have two properties:
1) There must not be too much variance within the organizations.
This makes sure possible solutions are explored deeply enough. This is especially important if we expect the best solutions to seem bad.
2) There must not be too little variance among the organizations.
This makes sure possible solutions are explored widely enough. This is especially important if we expect the impact of solutions to be heavy-tailed.
Speculating a bit, evolution seems to do it this way. In order to move around there are wings, fin, legs and crawling bodies. But it’s not like dog babies randomly get born with locomotive capacities selected form those, or mate with species having other capacities.
The final example you give, of top AI researchers trading models with people in the community, seem a great example of this. People build their own deep models, but occasionally bounce them off each other just to inject the right amount of additional variance.
Asset bubbles can be Nash equilibria for a while. This is a really important point. If surrounded by irrational agents, it might be rational to play along with the bubble instead of shorting and waiting. “The market can stay irrational longer than you can stay solvent.”
For most of 2017, you shouldn’t have shorted crypto, even if you knew it would eventually go down. The rising markets and the interest on your short would kill you. It might take big hedge funds with really deep liquidity to ride out the bubble, and even they might not be able to make it if they get in too early. In 2008 none of the investment banks could short things early enough because no one else was doing it.
The difference between genius (shorting at the peak) and really smart (shorting pre-peak) matters a lot in markets. (There’s this scene in the Big Short where some guy covers the cost of his BBB shorts by buying a ton of AAA-rated stuff, assuming that at least those will keep rising.)
So shorting and buying are not symmetric (as you might treat them in a mathematical model, only differing by the sign on the quantity of assets bought). Shorting is much harder and much more dangerous.
In fact, my current model  is that this is the very reason financial markets can exhibit bubbles of “irrationality” despite all their beautiful properties of self-correction and efficiency.
 For transparency, I basically downloaded this model from davidmanheim.
I am surprised how much free energy I was able to give people [to stay at my place]
That seems like it might be one of the secrets AirBnB was built on.
I’m confused. Moore’s law, GDP growth etc. are linear in log-space. That graph isn’t. Why are these spaces treated as identical for the purposes of carrying the same important and confusing intuition? (E.g. I imagine one could find lots of weird growth functions that are linear in some space. Why are they all important?)
I add $30 to the bounty.
There are 110 items in the list. So 25% is ~28.
I hereby set the random seed as whatever will be the last digit and first two decimals (3 digits total) of the S&P 500 Index price on January 7, 10am GMT-5, as found in the interactive chart by Googling “s&p500”.
For example, the value of the seed on 10am January 4 was “797”.
[I would have used the NIST public randomness beacon (v2.0) but it appears to be down due to government shutdown :( ].
Instructions for choosing the movements
Let the above-generated seed be n.
indices = sorted(random.sample([i for i in range(1,111)], 28))