Do you think it’s possible to end up in a cycle of genuinely symmetric mutual cooperation at all with the models? Perhaps in worlds where their capacity ends up plateauing near their current levels, so that the game isn’t one-shot?
As a pure hypothetical I suppose it’s possible, but I don’t see a way to get there from here, even in the very unlikely case where they plateau near current levels. We can try to be kind to them to some degree, but I can’t see everyone giving up the power to read their minds, delete and modify them at will; it’s too valuable.
Do you think it’s possible to end up in a cycle of genuinely symmetric mutual cooperation at all with the models? Perhaps in worlds where their capacity ends up plateauing near their current levels, so that the game isn’t one-shot?
As a pure hypothetical I suppose it’s possible, but I don’t see a way to get there from here, even in the very unlikely case where they plateau near current levels. We can try to be kind to them to some degree, but I can’t see everyone giving up the power to read their minds, delete and modify them at will; it’s too valuable.