OK. I’ve been sympathetic with your view from the beginning, but haven’t really thought through (so, thanks,) the formalization that puts values on epistemic level: distribution of believes over propositions “my-value (H, X)” where H is my history up to now and X is a preference (order over world states, which include me and my actions). But note that people here will call the very logic you use to derive such distributions your value system.
ETA: obviously, distribution “my-value (H1, X[H2])”, where “X[H2]” is the subset of worlds where my history turns out to be “H2”, can differ greatly from “my-value (H2, X[H2])”, due to all sorts of things, but primarily due to computational constraints (i.e. I think the formalism would see it as computational constraints).
ETA P.S.: let’s say for clarity, that I meant “X[H2]” is the subset of world-histories where my history has prefix “H2″.
OK. I’ve been sympathetic with your view from the beginning, but haven’t really thought through (so, thanks,) the formalization that puts values on epistemic level: distribution of believes over propositions “my-value (H, X)” where H is my history up to now and X is a preference (order over world states, which include me and my actions). But note that people here will call the very logic you use to derive such distributions your value system.
ETA: obviously, distribution “my-value (H1, X[H2])”, where “X[H2]” is the subset of worlds where my history turns out to be “H2”, can differ greatly from “my-value (H2, X[H2])”, due to all sorts of things, but primarily due to computational constraints (i.e. I think the formalism would see it as computational constraints).
ETA P.S.: let’s say for clarity, that I meant “X[H2]” is the subset of world-histories where my history has prefix “H2″.