Idle musing: should we all be writing a Claude Constitution-esque set of posts about our hopes for how AIs help humans around the dangerous moments in takeoff, in hopes that this influences the models advise not only us, but people who are coming into these issues fresher than us?
(Yes, I know that from the outside I have exponentially less influence on model behavior than Anthropic, and for MIRI-like reasons maybe this doesn’t go well at all. But, you know, play to all of your outs.)
Idle musing: should we all be writing a Claude Constitution-esque set of posts about our hopes for how AIs help humans around the dangerous moments in takeoff, in hopes that this influences the models advise not only us, but people who are coming into these issues fresher than us?
(Yes, I know that from the outside I have exponentially less influence on model behavior than Anthropic, and for MIRI-like reasons maybe this doesn’t go well at all. But, you know, play to all of your outs.)