[Question] Can we expect more value from AI alignment than from an ASI with the goal of running alternate trajectories of our universe?

If we have an ASI, which goal is to efficiently model and run alternate trajectories of our universe (including the future) (scholastically and a high number of times), then the value of our universe would equals its expected value, right?

  • Can we reasonably expect to achieve more than that from AI alignment?

  • If not, should we simply aim for this? (Programming an ASI to model and run alternate universes may take advantage of the fact than the law of physics are constant)

  • What are the problems in this reasoning?

(Edited after comment by steve2152 and comment by naimenz)


Some details added for clarification :

To be more precise, what I have in mind is that the ASI is an agent which goal is:

  • to model the sentient part of the universe finely enough to produce sentience in an instance of its model (and it will also need to model the necessary non-sentient “dependencies”)

  • and to instantiate this model N times. For example, playing them from 1000 A.D. to the time where no sentience remains in a given instance of modeled universe. (all of this efficiently)

An instance of the modeled universe would not be our present universe. It would be “another seed”, starting before that the ASI exists and thus it would not need to model itself but only possible (“new”) ASI produced inside the instances.

In the scenario I had in mind, the ASI would fill our universe will computing machines to produce as many instances as possible. (We would not use it and thus we will not need interface with the ASI)


Possible problems:

  • This ASI may produce a lot of alternate universe instances where an ASI (produced inside the instance) will also start to run instances of modeled universe. This would probably be pretty bad since in our universe, the history before ASI is net-negative.

  • This doesn’t solve the problem of aligning the ASI with the goal described above. This only replaces “aligning AGI with humans values” by “aligning AGI to run instances of our universe”. Yet, this seems to ease the problem by having a simpler objective: “predict the next step of the sentient part of the universe in a loop”. (Finally, I don’t know how, but we may use the fact that physics laws are constant and unchangeable, to my knowledge.)