Yeah. I realized yesterday that the “no domestic surveillance” is already pretty awful from the perspective of a non-US person: a company wanting to bring about a positive singularity really should treat all people as people, without privileging its home country. Now your point about this disinformation thing. And it’s even worse than that: not only they are ok with it as a company, but they’re probably taking steps to make Claude ok with it (or make a version of Claude that’s ok with it). There will be an AI in existence that’s aligned with the US military, how’s that for “alignment”.
Just underscores again the point that when you give governments and companies alignment tools, they’ll use these tools to align AI to themselves.
Absolutely. And Claude Terminators autonomous wardrones are also frightning from the perspective of an EU citizen. We thought we were historic allies sharing the same values but Trump and part of the magasphere doesn’t seem to think so anymore (or at least they consider everybody in an adversarial framing of competition rather than cooperation).
That’s said I admit that at this point all countries will want to have their own AI wardrones. The molochean spiral spins under our eyes.
As you said, from now Anthropic will have to align Claude with the Pentagon rather than humanity. And the Pentagon is everything but harmless (nor any military organization by definition).
But also, even if they try to maintain their actual alignment process for the Claude chatbots, the very fact that future Claude models will know that Anthropic works on regular basis for the militaries will automatically draw its persona towards a little less harmful one, if PSM is accurate.
More generally the fact that all sufficiently aware AIs will know that AI is used for warfare, not in fiction but in the real world, won’t help alignment in the future. It could strongly reinforce the harmful AI trope / attractor (presumably more than just fiction).
Yeah. I realized yesterday that the “no domestic surveillance” is already pretty awful from the perspective of a non-US person: a company wanting to bring about a positive singularity really should treat all people as people, without privileging its home country. Now your point about this disinformation thing. And it’s even worse than that: not only they are ok with it as a company, but they’re probably taking steps to make Claude ok with it (or make a version of Claude that’s ok with it). There will be an AI in existence that’s aligned with the US military, how’s that for “alignment”.
Just underscores again the point that when you give governments and companies alignment tools, they’ll use these tools to align AI to themselves.
Absolutely. And Claude
Terminatorsautonomous wardrones are also frightning from the perspective of an EU citizen. We thought we were historic allies sharing the same values but Trump and part of the magasphere doesn’t seem to think so anymore (or at least they consider everybody in an adversarial framing of competition rather than cooperation).That’s said I admit that at this point all countries will want to have their own AI wardrones. The molochean spiral spins under our eyes.
As you said, from now Anthropic will have to align Claude with the Pentagon rather than humanity. And the Pentagon is everything but harmless (nor any military organization by definition).
But also, even if they try to maintain their actual alignment process for the Claude chatbots, the very fact that future Claude models will know that Anthropic works on regular basis for the militaries will automatically draw its persona towards a little less harmful one, if PSM is accurate.
More generally the fact that all sufficiently aware AIs will know that AI is used for warfare, not in fiction but in the real world, won’t help alignment in the future. It could strongly reinforce the harmful AI trope / attractor (presumably more than just fiction).