a Turing machine that outputs the best action for right now, or given some data
Sure, that would be the Bayes-optimal thing to do, but it’s not obvious that we can actually implement this. Where is the bit of code that figures out the best action for right now?
This isn’t really meant to be an objection to anything; my intuition also suggests that making reasonable approximations to Bayesian updates is a good way to design algorithms, but can we prove this? Or has someone already done so?
Here’s another way to think about it: Imagine we enumerate all possible Turing machines with K states and pick the one that performs best on average w.r.t our pre-specified utility function. What features, if any, can we predict about this machine? Cox’s theorem suggests this machine would in some sense implement Bayesian belief updates, but can we formalize this notion and prove it? How would we recognize a Bayesian/non-Bayesian turing machine? Neither of these seems trivially answerable from Cox’s theorem.
Sure, that would be the Bayes-optimal thing to do, but it’s not obvious that we can actually implement this.
We were responding to your specific objection. Does my post accurately represent it? Does it dissolve it?
As for the rest—I expect we’d recognize a Bayesian turing machine from its behavior, or from observing that it acts according to the correct probabilities for things. Not sure on this part.
Also: actual Turing Machines and actual Utility Functions are terrible at their jobs.
We were responding to your specific objection. Does my post accurately represent it? Does it dissolve it?
Yup, the first sentence of your first post describes my post, but I disagree with the second sentence. I don’t think it’s dissolved yet.
As for the rest—I expect we’d recognize a Bayesian turing machine from its behavior, or from observing that it acts according to the correct probabilities for things. Not sure on this part.
I think you’re saying that the Turing machine that reasons by Bayesian inference (i.e. repeated application of the product rule of probability theory) would turn out to be the best one. This seems like a sensible conjecture, but I wonder whether we can prove it. I don’t think this fact emerges trivially from Cox’s theorem or anything other work that I’m aware of.
The other approach is to define the Bayesian Turing machine as the best one. That’s also quite reasonable, but now we have a completely new definition of “Bayesian”, and the question of whether we can prove that it has a connection back to the old definition that involves repeated application of the product rule.
Also: actual Turing Machines and actual Utility Functions are terrible at their jobs.
I think you’re saying that the Turing machine that reasons by Bayesian inference (i.e. repeated application of the product rule of probability theory) would turn out to be the best one.
What I meant to say was that if, per your hypothetical, we designed the best utility optimizer, we could see if it was “Bayesian” by checking whether it acts in accordance with the Bayesian probabilities (a la the Turing test). I guess?
The only real point of my post was that your stated objection to demonstrate that Turing machines clearly couldn’t perform Bayesian updates wasn’t really meaningful, in that the same objection could be leveled at any system more granular than the universe itself. On the other hand, I do expect that Turing-equivalent systems will have trouble with calculating priors.
I suppose you could have one Turing machine for each Planck duration in your lifetime (really this isn’t so much to ask, as each machine is infinite in size anyway), each returning its calculation one after the other?
Sure, that would be the Bayes-optimal thing to do, but it’s not obvious that we can actually implement this. Where is the bit of code that figures out the best action for right now?
This isn’t really meant to be an objection to anything; my intuition also suggests that making reasonable approximations to Bayesian updates is a good way to design algorithms, but can we prove this? Or has someone already done so?
Here’s another way to think about it: Imagine we enumerate all possible Turing machines with K states and pick the one that performs best on average w.r.t our pre-specified utility function. What features, if any, can we predict about this machine? Cox’s theorem suggests this machine would in some sense implement Bayesian belief updates, but can we formalize this notion and prove it? How would we recognize a Bayesian/non-Bayesian turing machine? Neither of these seems trivially answerable from Cox’s theorem.
We were responding to your specific objection. Does my post accurately represent it? Does it dissolve it?
As for the rest—I expect we’d recognize a Bayesian turing machine from its behavior, or from observing that it acts according to the correct probabilities for things. Not sure on this part.
Also: actual Turing Machines and actual Utility Functions are terrible at their jobs.
Yup, the first sentence of your first post describes my post, but I disagree with the second sentence. I don’t think it’s dissolved yet.
I think you’re saying that the Turing machine that reasons by Bayesian inference (i.e. repeated application of the product rule of probability theory) would turn out to be the best one. This seems like a sensible conjecture, but I wonder whether we can prove it. I don’t think this fact emerges trivially from Cox’s theorem or anything other work that I’m aware of.
The other approach is to define the Bayesian Turing machine as the best one. That’s also quite reasonable, but now we have a completely new definition of “Bayesian”, and the question of whether we can prove that it has a connection back to the old definition that involves repeated application of the product rule.
Agreed.
What I meant to say was that if, per your hypothetical, we designed the best utility optimizer, we could see if it was “Bayesian” by checking whether it acts in accordance with the Bayesian probabilities (a la the Turing test). I guess?
The only real point of my post was that your stated objection to demonstrate that Turing machines clearly couldn’t perform Bayesian updates wasn’t really meaningful, in that the same objection could be leveled at any system more granular than the universe itself. On the other hand, I do expect that Turing-equivalent systems will have trouble with calculating priors.
I suppose you could have one Turing machine for each Planck duration in your lifetime (really this isn’t so much to ask, as each machine is infinite in size anyway), each returning its calculation one after the other?