CEO at Redwood Research.
AI safety is a highly collaborative field—almost all the points I make were either explained to me by someone else, or developed in conversation with other people. I’m saying this here because it would feel repetitive to say “these ideas were developed in collaboration with various people” in all my comments, but I want to have it on the record that the ideas I present were almost entirely not developed by me in isolation.
Please contact me via email (bshlegeris@gmail.com) instead of messaging me on LessWrong.
If we are ever arguing on LessWrong and you feel like it’s kind of heated and would go better if we just talked about it verbally, please feel free to contact me and I’ll probably be willing to call to discuss briefly.
I agree that the AIs are pretty bad at handling conceptually confusing stuff. I basically think of them as being incredibly knowledgeable, not that smart, and having huge amounts of intuition on how to program (mostly due to their knowledge and their having read huge amounts of code).
My guess is that for any reasonable operationalization of “raw intelligence”, they’re getting smarter?