The “Humans do X because evolution” argument does not actually explain anything about mechanisms. I keep seeing people make this argument, but it’s a non sequitur to the points I’m making in this post. You’re explaining how the behavior may have gotten there, not how the behavior is implemented. I think that “because selection pressure” is a curiosity-stopper, plain and simple.
AGI won’t be subjected to the same evolutionary pressures, so every alignment strategy relying on empathy or social reward functions, it is, in my opinion, hopelessly naive.
This argument proves too much, since it implies that planes can’t work because we didn’t subject them to evolutionary pressures for flight. It’s locally invalid.
The “Humans do X because evolution” argument does not actually explain anything about mechanisms. I keep seeing people make this argument, but it’s a non sequitur to the points I’m making in this post. You’re explaining how the behavior may have gotten there, not how the behavior is implemented. I think that “because selection pressure” is a curiosity-stopper, plain and simple.
This argument proves too much, since it implies that planes can’t work because we didn’t subject them to evolutionary pressures for flight. It’s locally invalid.