Destroying the robot greatly diminishes its future ability to shoot, but it would also greatly diminishes its future ability to see blue. The robot doesn’t prefer ‘shooting blue’ to ‘not shooting blue’, it prefers ‘seeing blue and shooting’ to ‘seeing blue and not shooting’.
So the original poster was right.
Edit: I’m wrong, see below
Hi LW! My name is Yaacov, I’ve been lurking here for maybe 6 months but I’ve only recently created an account. I’m interested in minimizing human existential risk, effective altruism, and rationalism. I’m just starting a computer science degree at UCLA, so I don’t know much about the topic now but I’ll learn more quickly.
Specific questions:
What can I do to reduce existential risk, especially that posed by AI? I don’t have an income as of yet. What are the best investments I can make now in my future ability to reduce existential risk?