posts I may write soonish, lmk which ones sound most promising:
Alignment in a Detailed Universe: basically compressibility-of-reality and how it should shape alignment plans.
lit review of specifically the “paradigm-building” angle of conceptual/theoretical alignment.
[below ideas partly because I’m working with Orxl]
Coming up with really rudimentary desiderata/properties of the A action output of AI_0 in QACI.
Drawing links between “things mentioned in QACI” and “corresponding things mentioned elsewhere, e.g. in Superintelligence PDS”. (E.g. specific problem --> how a thing in QACI solves it or not.)
Finish the Pearl Causality math basics first, before doing any of the above. Then I can do the above but with causal networks!
posts I may write soonish, lmk which ones sound most promising:
Alignment in a Detailed Universe: basically compressibility-of-reality and how it should shape alignment plans.
lit review of specifically the “paradigm-building” angle of conceptual/theoretical alignment.
[below ideas partly because I’m working with Orxl]
Coming up with really rudimentary desiderata/properties of the A action output of AI_0 in QACI.
Drawing links between “things mentioned in QACI” and “corresponding things mentioned elsewhere, e.g. in Superintelligence PDS”. (E.g. specific problem --> how a thing in QACI solves it or not.)
Finish the Pearl Causality math basics first, before doing any of the above. Then I can do the above but with causal networks!