The encoder and decoder both spontaneously settle on this as a schelling point for encoding the message.
LLMs do seem to be pretty good at picking self-consistent Schelling points, at least in simple cases—I’ve got a writeup here of some casual experiments I did with GPT-4 last January on picking various Schelling points, eg a date, a number, a word (also some discussion of that in the MATS slack).
this still seems like a pretty incredible claim
I think it seems somewhat less surprising to me (maybe because of the Schelling point experiments), but I certainly wouldn’t have been confident that it would do this well.
PS—I’m loving the frequent shortform posts, I hope you continue! I try to do something somewhat similar with my research diary, but usually no one reads that and it’s certainly not daily. I’m tempted to try doing the same thing :)
LLMs do seem to be pretty good at picking self-consistent Schelling points, at least in simple cases—I’ve got a writeup here of some casual experiments I did with GPT-4 last January on picking various Schelling points, eg a date, a number, a word (also some discussion of that in the MATS slack).
I think it seems somewhat less surprising to me (maybe because of the Schelling point experiments), but I certainly wouldn’t have been confident that it would do this well.
PS—I’m loving the frequent shortform posts, I hope you continue! I try to do something somewhat similar with my research diary, but usually no one reads that and it’s certainly not daily. I’m tempted to try doing the same thing :)