The models know the difference between the system prompt and the user prompt, so I’m not sure how much of a difference these things would make.
From the COTs that I’ve looked at it is clear that the models understand the task. I think the reason you get some weird responses sometimes is because the models do not see an obvious shelling point and then they try their best to come up with one anyway.
The models know the difference between the system prompt and the user prompt, so I’m not sure how much of a difference these things would make.
From the COTs that I’ve looked at it is clear that the models understand the task. I think the reason you get some weird responses sometimes is because the models do not see an obvious shelling point and then they try their best to come up with one anyway.