So, granting the assumption of not corrupting the humans (which is maybe what you are denying), doesn’t this imply that we can go on adding sensors after the fact until, at some point, the difference between fooling them all and being honest becomes unproblematic?
So, granting the assumption of not corrupting the humans (which is maybe what you are denying), doesn’t this imply that we can go on adding sensors after the fact until, at some point, the difference between fooling them all and being honest becomes unproblematic?