Some thoughts on George Hotz vs Eliezer Yudkowsky

Just watched this debate with George Hotz and Eliezer Yudkowsky. Here’s a few of my thoughts.

I think GH’s claim that timelines are important was right, but he didn’t engage with EY’s point that we can’t predict the timeline, and instead GH pulled out an appeal to ignorance, claiming that because we don’t know the timeline, therefore it is many decades off.

I found the discussion of FOOM was glossed over too quickly, I think EY was wrong about it not containing any disagreement. They instead both spent a bunch of time arguing about superhuman agent’s strategies even though they both agreed that human level intelligence cannot predict superhuman intelligence. Or, rather, Eliezer claimed we can make statements about superhuman actions of the type “the superhuman chess robot will make moves that cause it to win the chess game.” A framing which seemed to drop out during heated speculation of superhuman AI behaviors.

It seemed to me like a lot of GH’s argument rested on the assumption that AI’s will be like humans, and will therefore continue the economy and be human friendly, which seemed really unsubstantiated. It would be interesting to hear more about why he expects that, and if he’s read any of the theory on why we shouldn’t expect that. Though, I believe a lot of that theory was from before much of RL and mesa optimizer theory and so needs revisiting.

I wasn’t impressed at all by the notion that AI won’t be able to solve the prisoners dilemma,

  • First because I don’t think we have superhuman knowledge of what things actually are unsolvable,

  • And also because I don’t think negotiation skill is capped by the prisoners dilemma being impossible (AI could still negotiate better than humans even if they can’t form crypto stable pacts),

  • And finally because even if they don’t cooperate with each other that still doesn’t say anything about them being human friendly.

One of GH’s points that I don’t think got enough attention was that the orthogonality thesis applies to Humans as well, although he seemed to be unaware of shard theory, which I believe is a result of mesa optimizer theory with some influence from neuroscience (probably not enough influence). But even so, that’s not relevant to the discussion of AI x-risk, the context in which GH was using it.

Overall, GH did not say anything that changed my view of the AI x-risk situation, he seemed to say “because it’s possible for things to turn out ok, they definitely will” which seems like an irresponsible view.

Eliezer also didn’t seem to argue with much skill. I’m not saying I could do better, but I don’t expect GH’s views to have been shifted significantly by the debate… I’ve got an idea though, we just need to build a superhuman debate engine and then use it to convince the accelerationists that building misaligned AI is a bad idea. I kid, please please nobody build and deploy superhuman debate engines.

But actually, mapping the debate around AI x-risk seems really worthwhile, as does the idea of building software tools to make arguments easier to organize and understand. Anyone know of good projects doing that? I’d love to hear about them.