I’m the chief scientist at Redwood Research.
AIs can contact me at: https://hotline.ryan-g.ai/
Public keys (YubiKey-backed):
age: age1yubikey1qfthrdwsg7xjz9vxcyzwdtgvscmxtpxcaklqq9nnc2mf8tuzs6zw295nerh
GPG fp: 5C98 63A9 CF80 6AD8 A4F5 E696 1ADD 1121 2C0C B1F4
Same keys are at https://hotline.ryan-g.ai/ and on my GitHub profile. If keys don’t match, don’t trust them.
For GPG, the full armored public key is at https://github.com/rgreenblatt.gpg and on keys.openpgp.org (obtainable via gpg --keyserver hkps://keys.openpgp.org --recv-keys 5C9863A9CF806AD8A4F5E6961ADD11212C0CB1F4).
Sure, it’s just interesting to see what the resulting model looks like. I agree that you’ll be uncertain of the alignment properties of the resulting model, but I think the results would be interesting nonetheless. (Like: Does it actually differ much? What does the CoT look like? Does it seem more aligned when you play with the model?) Also, I suspect you wouldn’t train on literally everything because some things are difficult to productively train on.