(I think you mixed up some of the agent references in the post.)
If agent1 benefits from agent2 expecting it to do X, it should find a way of signaling this fact, for example deciding to do X quickly, so that agent2 can just simulate it and check.
(I think you mixed up some of the agent references in the post.)
If agent1 benefits from agent2 expecting it to do X, it should find a way of signaling this fact, for example deciding to do X quickly, so that agent2 can just simulate it and check.