Defecting one round earlier dominates pure tit-for-tat, but defecting five rounds earlier doesn’t dominate pure tit-for-tat. Pure tit-for-tat is better against pure tit-for-tat. So there might be a nash equilibrium containing only strategies that play tit-for-tat until the last few rounds.
Ah, I see, so this approach differs a lot from Ole Peters’
I looked at his paper on the petersburg paradox and I think he gets the correct result for the iterated game. He doesn’t do fractional betting, but he has a variable for players wealth—implicitly, price/wealth is a betting fraction (and since payoffs are fixed, price is implicitly offered odds). Also, and this is quite confusing, even though in the beginning it sounds like he wants to repeat the game with the price fixed but wealth changing over time, his actual calculation assumes the wealth (or distribution over growth rates) is the same each time. He talks about this at the bottom of page 11 and argues that its fine because of commutativity. I’m not sure if that commutativity argument works out, but it means the part before is effectively calculating the growth rate of a betting fraction. And if theres no death in the game, then the highest growth rate does indeed optimize my criterion.
Conceptually though there are differences: Peters totally rejects ensemble averaging. This works in infinite games with no chance of death, because then one player will with certainty experience events at frequencies reflecting the true odds—so it works in ordinary kelly, and it works in this petersburg-bet-in-a-kelly, but it wouldn’t work on the versions with ending chance.
(Also what I said about buying multiples in the last comment was confused—that would be different from one bigger bet.)
OTOH, if we use a cutoff of 1⁄2 rather than 0, the story might be different; there might be a finite price after which it’s not worth it. Which would be interesting. But probably not, I think.
Probably not, no. And provably not for the triple payoff version, so it wouldnt avoid the paradox anyway.
Defecting one round earlier dominates pure tit-for-tat, but defecting five rounds earlier doesn’t dominate pure tit-for-tat. Pure tit-for-tat is better against pure tit-for-tat. So there might be a nash equilibrium containing only strategies that play tit-for-tat until the last few rounds.
Defecting in the last x rounds is dominated by defecting in the last x+1, so there is no pure-strategy equilibrium which involves cooperating in any rounds. But perhaps you mean there could be a mixed strategy equilibrium which involves switching to defection some time near the end, with some randomization.
Clearly such a strategy must involve defecting in the final round, since there is no incentive to cooperate.
But then, similarly, it must involve defecting on the second-to-last round, etc.
So it should not have any probability of cooperating—at least, not in the game-states which have positive probability.
Right? I think my argument is pretty clear if we assume subgame-perfect equilibria (and so can apply backwards induction). Otherwise, it’s a bit fuzzy, but it still seems to me like the equilibrium can’t have a positive probability of cooperating on any turn, even if players would hypothetically play tit-for-tat according to their strategies.
(For example, one equilibrium is for players to play tit-for-tat, but with both players’ first moves being to defect.)
Defecting one round earlier dominates pure tit-for-tat, but defecting five rounds earlier doesn’t dominate pure tit-for-tat. Pure tit-for-tat is better against pure tit-for-tat. So there might be a nash equilibrium containing only strategies that play tit-for-tat until the last few rounds.
I looked at his paper on the petersburg paradox and I think he gets the correct result for the iterated game. He doesn’t do fractional betting, but he has a variable for players wealth—implicitly, price/wealth is a betting fraction (and since payoffs are fixed, price is implicitly offered odds). Also, and this is quite confusing, even though in the beginning it sounds like he wants to repeat the game with the price fixed but wealth changing over time, his actual calculation assumes the wealth (or distribution over growth rates) is the same each time. He talks about this at the bottom of page 11 and argues that its fine because of commutativity. I’m not sure if that commutativity argument works out, but it means the part before is effectively calculating the growth rate of a betting fraction. And if theres no death in the game, then the highest growth rate does indeed optimize my criterion.
Conceptually though there are differences: Peters totally rejects ensemble averaging. This works in infinite games with no chance of death, because then one player will with certainty experience events at frequencies reflecting the true odds—so it works in ordinary kelly, and it works in this petersburg-bet-in-a-kelly, but it wouldn’t work on the versions with ending chance.
(Also what I said about buying multiples in the last comment was confused—that would be different from one bigger bet.)
Probably not, no. And provably not for the triple payoff version, so it wouldnt avoid the paradox anyway.
Defecting in the last x rounds is dominated by defecting in the last x+1, so there is no pure-strategy equilibrium which involves cooperating in any rounds. But perhaps you mean there could be a mixed strategy equilibrium which involves switching to defection some time near the end, with some randomization.
Clearly such a strategy must involve defecting in the final round, since there is no incentive to cooperate.
But then, similarly, it must involve defecting on the second-to-last round, etc.
So it should not have any probability of cooperating—at least, not in the game-states which have positive probability.
Right? I think my argument is pretty clear if we assume subgame-perfect equilibria (and so can apply backwards induction). Otherwise, it’s a bit fuzzy, but it still seems to me like the equilibrium can’t have a positive probability of cooperating on any turn, even if players would hypothetically play tit-for-tat according to their strategies.
(For example, one equilibrium is for players to play tit-for-tat, but with both players’ first moves being to defect.)
Yeah you’re right. I just realized that what I had in mind originally already implicitly had superationality.