AI safety & alignment researcher
In Rob Bensinger’s typology: AGI-alarmed, tentative welfarist, and eventualist.
Public stance: AI companies are doing their best to build ASI (AI much smarter than humans), and have a chance of succeeding. No one currently knows how to build ASI without an unacceptable level of existential risk (> 5%). Therefore, companies should be forbidden from building ASI until we know how to do it safely.
I have signed no contracts or agreements whose existence I cannot mention.
Ah, yeah, I definitely get ‘You’re right to push back’; I feel like that’s something I see from almost all models. I’m totally making this up, but I’ve assumed that was encouraged by the model trainers so that people would feel free to push back, since it’s a known failure mode — or at least was for a while — that some users assume the AI is perfectly logical and all-knowing.