I hesitate to say “confident”. But I think you’re not gonna have world models emerging LLMs that are wrapped in a “this is a simulation” layer.. probably?
Also maybe even if they did, the procedure I’m describing, if it worked at all, would naively make them care about some simulated thing for its own sake. Not care about the simulated thing for instrumental reasons so it could get some other thing in the real world.
I hesitate to say “confident”. But I think you’re not gonna have world models emerging LLMs that are wrapped in a “this is a simulation” layer.. probably?
Also maybe even if they did, the procedure I’m describing, if it worked at all, would naively make them care about some simulated thing for its own sake. Not care about the simulated thing for instrumental reasons so it could get some other thing in the real world.