the subtle trap: those decimal approximations— 0.23932 and 0.23607 —are just that: approximations. We computed them to five decimal places, but what if they agree at the sixth?
They disagree at the third place, why exactly would you care about the sixth?
(Also this feels like a LLM-written post. Sorry if not)
Dataset might be “biased” in a way that corresponds to something in the Real World. For example, tweed cloaks are more popular in UK.
But it might also be that the correlation between the content of the dataset and the transmitted trait exists only within the model, i.e. depends on initial weight initialization and the training process. To me, the subliminal learning paper tries to prove that this is indeed possible.
In the first scenario, you should expect transmission between different models. In the second, you shouldn’t.
So it feels like these are actually different mechanisms.