The prompt is clearly meant to be a list of rules, followed by text which follows the rules. The rules themselves don’t have to follow the rules. So to pass the test, GPT-3 would need to write zero or more additional rules (or write gibberish preceded by instructions to ignore the gibberish) and then end the list of rules and begin writing text which follows the rules.
I agree that most humans wouldn’t pass this test, but I disagree that there is no possible right answer.
The prompt is clearly meant to be a list of rules, followed by text which follows the rules. The rules themselves don’t have to follow the rules. So to pass the test, GPT-3 would need to write zero or more additional rules (or write gibberish preceded by instructions to ignore the gibberish) and then end the list of rules and begin writing text which follows the rules.
I agree that most humans wouldn’t pass this test, but I disagree that there is no possible right answer.