So this would forego our ability to assess how well they do autonomously and make the scenario more similar to having custom or dynamic prompting per task.
I considered this, but, in the wild, we’d expect to see LLMs using a baseline level of generic prompt engineering for the interfaces they have access to. I wouldn’t suggest per-task custom prompts, but looking at the SOTA for general scaffolding might get more true-to-life results.
I don’t know that it necessarily advocates cooperation so much as coexistence, or tolerance. The message isn’t “we all want the same things, and we should work together to get them”, it’s “once you decide that it’s evil to allow the other half to continue their way of life—regardless of whether they want the same things as you do—you’ve guaranteed a war that will end your own way of life just as thoroughly”.
I think the ending makes that clear. You don’t get to just wipe out all of your enemies for free, they’ll fight back. The choice you get to make is whether the ‘good’ of your own survival outweighs the ‘bad’ of their survival, or whether you’d rather the two cancel each other out.