I’m confused. I know that it is like something to be me (this is in some sense the only thing I know for sure). It seems like there rules which shape the things I experience, and some of those rules can be studied (like the laws of physics). We are good enough at understanding some of these rules to predict certain systems with a high degree of accuracy, like how an asteroid will orbit a star or how electrons will be pushed through a wire by a particular voltage in a circuit. But I have no way to know or predict if it is like something to be a fish or GPT-4. I know that physical alterations to my brain seem to affect my experience, so it seems like there is a mapping from physical matter to experiences. I do not know precisely what this mapping is, and this indeed seems like a hard problem. In what sense do you disagree with my framing here?
But I have no way to know or predict if it is like something to be a fish or GPT-4
But I can predict what you say; I can predict if you are confused by the hard problem just by looking at your neural activation; I can predict word by word the following sentence that you are uttering: “The hard problem is really hard.”
I would be curious to know what you think about the box solving the meta-problem just before the addendum. Do you think it is unlikely that AI would rediscover the hard problem in this setting?
this is in some sense the only thing I know for sure
You don’t. All your specific experiences are imprecise approximations: you can’t be sure what exact color you saw for how many nanoseconds, you can’t be sure all your brain except small part implementing only current thought haven’t evaporated microsecond ago. So you can have imprecise models of a fish brain the same way you have imprecise models of your brain—your awareness of your brain is casually connected to your brain the same way your thoughts can be casually connected to a fish brain. You just can’t be fully fish.
I’m confused. I know that it is like something to be me (this is in some sense the only thing I know for sure). It seems like there rules which shape the things I experience, and some of those rules can be studied (like the laws of physics). We are good enough at understanding some of these rules to predict certain systems with a high degree of accuracy, like how an asteroid will orbit a star or how electrons will be pushed through a wire by a particular voltage in a circuit. But I have no way to know or predict if it is like something to be a fish or GPT-4. I know that physical alterations to my brain seem to affect my experience, so it seems like there is a mapping from physical matter to experiences. I do not know precisely what this mapping is, and this indeed seems like a hard problem. In what sense do you disagree with my framing here?
But I can predict what you say; I can predict if you are confused by the hard problem just by looking at your neural activation; I can predict word by word the following sentence that you are uttering: “The hard problem is really hard.”
I would be curious to know what you think about the box solving the meta-problem just before the addendum. Do you think it is unlikely that AI would rediscover the hard problem in this setting?
You don’t. All your specific experiences are imprecise approximations: you can’t be sure what exact color you saw for how many nanoseconds, you can’t be sure all your brain except small part implementing only current thought haven’t evaporated microsecond ago. So you can have imprecise models of a fish brain the same way you have imprecise models of your brain—your awareness of your brain is casually connected to your brain the same way your thoughts can be casually connected to a fish brain. You just can’t be fully fish.