Sure, but you get feedback for whether it helps customers with their immediate problems. You don’t get feedback on whether it helps with AI safety.
It’s the direction vs speed thing again. You’ll get good at building widgets that sell. You won’t get good at AI notkilleveryoneism.
Fair enough. There are some for profits where profit and impact are more related than others.
But it’s also quite likely your evals are not actually evaluating anything to do with x-risks or s-risks, and so it just feels like it’s making progress, but isn’t.
I’m assuming here people are trying to prevent AI from killing everyone. If you have other goals, this doesn’t apply.
I’d say this is the same thing for AI for-profits from the perspective of AI notkilleveryoneism. Probably, the modal outcome is slightly increasing the odds AI kills everyone. At least the non-profits the modal outcome is not doing anything, rather than making things worse.