Those rather are the questions, aren’t they? My thought when the original paper showed up on HN was that we can’t do anything remotely similar to constructing adversarial examples for a human visual cortex, and we already know of a lot of visual illusions (I’m particularly thinking of the Magic Eyeautostereograms)… “Perhaps there are thoughts we cannot think”.
Hard to see how we could test it without solving AI, though.
I don’t think we’d need to solve AI to test this. If we could get a detailed enough understanding of how the optical cortex functions it might be doable. Alternatively, we could try it on a very basic uploaded mouse or similar creature. On the other hand, if we can upload mice then we’re pretty close to uploading people, and if we can upload people we’ve got AI.
Those rather are the questions, aren’t they? My thought when the original paper showed up on HN was that we can’t do anything remotely similar to constructing adversarial examples for a human visual cortex, and we already know of a lot of visual illusions (I’m particularly thinking of the Magic Eye autostereograms)… “Perhaps there are thoughts we cannot think”.
Hard to see how we could test it without solving AI, though.
I don’t think we’d need to solve AI to test this. If we could get a detailed enough understanding of how the optical cortex functions it might be doable. Alternatively, we could try it on a very basic uploaded mouse or similar creature. On the other hand, if we can upload mice then we’re pretty close to uploading people, and if we can upload people we’ve got AI.