Could this be because the typos increase the length of the input when serialized to tokens, which gives the same effect as the “repeat the question 3 times” trick by letting models think longer during prompt processing?
Could this be because the typos increase the length of the input when serialized to tokens, which gives the same effect as the “repeat the question 3 times” trick by letting models think longer during prompt processing?